Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
2,029
2,603,975,206
IssuesEvent
2015-02-24 19:01:18
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳什么是疱疹
auto-migrated Priority-Medium Type-Defect
``` 沈阳什么是疱疹〓沈陽軍區政治部醫院性病〓TEL:024-31023308�� �成立于1946年,68年專注于性傳播疾病的研究和治療。位于沈� ��市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史� ��久、設備精良、技術權威、專家云集,是預防、保健、醫療 、科研康復為一體的綜合性醫院。是國家首批公立甲等部隊�� �院、全國首批醫療規范定點單位,是第四軍醫大學、東南大� ��等知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤 部衛生部評為衛生工作先進單位,先后兩次榮立集體二等功�� � ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:15
1.0
沈阳什么是疱疹 - ``` 沈阳什么是疱疹〓沈陽軍區政治部醫院性病〓TEL:024-31023308�� �成立于1946年,68年專注于性傳播疾病的研究和治療。位于沈� ��市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史� ��久、設備精良、技術權威、專家云集,是預防、保健、醫療 、科研康復為一體的綜合性醫院。是國家首批公立甲等部隊�� �院、全國首批醫療規范定點單位,是第四軍醫大學、東南大� ��等知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤 部衛生部評為衛生工作先進單位,先后兩次榮立集體二等功�� � ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:15
defect
沈阳什么是疱疹 沈阳什么是疱疹〓沈陽軍區政治部醫院性病〓tel: �� � , 。位于沈� �� 。是一所與新中國同建立共輝煌的歷史� ��久、設備精良、技術權威、專家云集,是預防、保健、醫療 、科研康復為一體的綜合性醫院。是國家首批公立甲等部隊�� �院、全國首批醫療規范定點單位,是第四軍醫大學、東南大� ��等知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤 部衛生部評為衛生工作先進單位,先后兩次榮立集體二等功�� � original issue reported on code google com by gmail com on jun at
1
189,793
6,801,799,963
IssuesEvent
2017-11-02 17:57:57
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
closed
build drake mex binaries into binary dir (instead of source dir)
configuration: matlab priority: medium team: software core type: cleanup
now is our chance to fix this workflow. if we install call_matlab_client's build output to the binary dir, and add the binary dir properly in addpath_drake, then I think we might be able to finally clean this up. note that I still don't want to be copying the .m files into an install directory, so our source dirs still stay in the path. but this would at least be progress.
1.0
build drake mex binaries into binary dir (instead of source dir) - now is our chance to fix this workflow. if we install call_matlab_client's build output to the binary dir, and add the binary dir properly in addpath_drake, then I think we might be able to finally clean this up. note that I still don't want to be copying the .m files into an install directory, so our source dirs still stay in the path. but this would at least be progress.
non_defect
build drake mex binaries into binary dir instead of source dir now is our chance to fix this workflow if we install call matlab client s build output to the binary dir and add the binary dir properly in addpath drake then i think we might be able to finally clean this up note that i still don t want to be copying the m files into an install directory so our source dirs still stay in the path but this would at least be progress
0
75,846
26,097,337,805
IssuesEvent
2022-12-26 22:50:03
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
Another encryption bug: "unencrypted block in encrypted object set"
Type: Defect
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | NixOS Distribution Version | 22.11 Kernel Version |5.15.81 Architecture |amd64 OpenZFS Version |zfs-2.1.7-1/zfs-kmod-2.1.7-1 ### Describe the problem you're observing Attempting to send a snapshot of an encrypted dataset results in a panic. ### Describe how to reproduce the problem How to get to the on-disk state that leads to this panic is unclear to me. However, on the affected machine, trying to sync a particular dataset (still trying to find a small `zfs send command` that triggers it quickly) reliably triggers it. ### Include any warning/errors/backtraces from the system logs ``` [210065.805807] PANIC: unencrypted block in encrypted object set 244177 [210065.805813] Showing stack for process 93744 [210065.805815] CPU: 0 PID: 93744 Comm: send_traverse Tainted: P IO 5.15.81 #1-NixOS [210065.805817] Hardware name: FUJITSU D3402-B1/D3402-B1, BIOS V5.0.0.11 R1.13.0 for D3402-B1x 02/19/2016 [210065.805818] Call Trace: [210065.805820] <TASK> [210065.805823] dump_stack_lvl+0x46/0x5e [210065.805830] vcmn_err.cold+0x50/0x68 [spl] [210065.805839] ? __cond_resched+0x16/0x50 [210065.805842] ? __kmalloc_node+0x14c/0x490 [210065.805844] ? mutex_lock+0xe/0x30 [210065.805846] ? spl_kmem_zalloc+0xa7/0xf0 [spl] [210065.805852] ? spa_log_error+0xcb/0x150 [zfs] [210065.805938] zfs_panic_recover+0x6d/0x90 [zfs] [210065.806013] send_cb+0x4ec/0x5e0 [zfs] [210065.806068] traverse_visitbp+0x371/0xb00 [zfs] [210065.806123] ? __cond_resched+0x16/0x50 [210065.806126] ? mutex_lock+0xe/0x30 [210065.806128] ? bqueue_enqueue_impl+0xf9/0x140 [zfs] [210065.806176] traverse_dnode+0xb9/0x1f0 [zfs] [210065.806230] traverse_visitbp+0x78e/0xb00 [zfs] [210065.806282] traverse_visitbp+0x4ca/0xb00 [zfs] [210065.806333] traverse_visitbp+0x4ca/0xb00 [zfs] [210065.806385] traverse_visitbp+0x4ca/0xb00 [zfs] [210065.806437] traverse_visitbp+0x4ca/0xb00 [zfs] [210065.806489] traverse_visitbp+0x4ca/0xb00 [zfs] [210065.806542] traverse_dnode+0xb9/0x1f0 [zfs] [210065.806594] traverse_visitbp+0x9a1/0xb00 [zfs] [210065.806644] ? zap_lookup+0xd6/0x100 [zfs] [210065.806725] traverse_impl+0x1eb/0x4c0 [zfs] [210065.806778] ? send_merge_thread+0x370/0x370 [zfs] [210065.806830] ? set_next_task_fair+0x2d/0xd0 [210065.806834] traverse_dataset_resume+0x48/0x60 [zfs] [210065.806885] ? send_merge_thread+0x370/0x370 [zfs] [210065.806935] send_traverse_thread+0x50/0xa0 [zfs] [210065.806986] ? enqueue_range+0x170/0x170 [zfs] [210065.807036] thread_generic_wrapper+0x59/0x70 [spl] [210065.807044] ? __thread_exit+0x20/0x20 [spl] [210065.807050] kthread+0x127/0x150 [210065.807054] ? set_kthread_struct+0x50/0x50 [210065.807057] ret_from_fork+0x22/0x30 [210065.807061] </TASK> ```
1.0
Another encryption bug: "unencrypted block in encrypted object set" - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | NixOS Distribution Version | 22.11 Kernel Version |5.15.81 Architecture |amd64 OpenZFS Version |zfs-2.1.7-1/zfs-kmod-2.1.7-1 ### Describe the problem you're observing Attempting to send a snapshot of an encrypted dataset results in a panic. ### Describe how to reproduce the problem How to get to the on-disk state that leads to this panic is unclear to me. However, on the affected machine, trying to sync a particular dataset (still trying to find a small `zfs send command` that triggers it quickly) reliably triggers it. ### Include any warning/errors/backtraces from the system logs ``` [210065.805807] PANIC: unencrypted block in encrypted object set 244177 [210065.805813] Showing stack for process 93744 [210065.805815] CPU: 0 PID: 93744 Comm: send_traverse Tainted: P IO 5.15.81 #1-NixOS [210065.805817] Hardware name: FUJITSU D3402-B1/D3402-B1, BIOS V5.0.0.11 R1.13.0 for D3402-B1x 02/19/2016 [210065.805818] Call Trace: [210065.805820] <TASK> [210065.805823] dump_stack_lvl+0x46/0x5e [210065.805830] vcmn_err.cold+0x50/0x68 [spl] [210065.805839] ? __cond_resched+0x16/0x50 [210065.805842] ? __kmalloc_node+0x14c/0x490 [210065.805844] ? mutex_lock+0xe/0x30 [210065.805846] ? spl_kmem_zalloc+0xa7/0xf0 [spl] [210065.805852] ? spa_log_error+0xcb/0x150 [zfs] [210065.805938] zfs_panic_recover+0x6d/0x90 [zfs] [210065.806013] send_cb+0x4ec/0x5e0 [zfs] [210065.806068] traverse_visitbp+0x371/0xb00 [zfs] [210065.806123] ? __cond_resched+0x16/0x50 [210065.806126] ? mutex_lock+0xe/0x30 [210065.806128] ? bqueue_enqueue_impl+0xf9/0x140 [zfs] [210065.806176] traverse_dnode+0xb9/0x1f0 [zfs] [210065.806230] traverse_visitbp+0x78e/0xb00 [zfs] [210065.806282] traverse_visitbp+0x4ca/0xb00 [zfs] [210065.806333] traverse_visitbp+0x4ca/0xb00 [zfs] [210065.806385] traverse_visitbp+0x4ca/0xb00 [zfs] [210065.806437] traverse_visitbp+0x4ca/0xb00 [zfs] [210065.806489] traverse_visitbp+0x4ca/0xb00 [zfs] [210065.806542] traverse_dnode+0xb9/0x1f0 [zfs] [210065.806594] traverse_visitbp+0x9a1/0xb00 [zfs] [210065.806644] ? zap_lookup+0xd6/0x100 [zfs] [210065.806725] traverse_impl+0x1eb/0x4c0 [zfs] [210065.806778] ? send_merge_thread+0x370/0x370 [zfs] [210065.806830] ? set_next_task_fair+0x2d/0xd0 [210065.806834] traverse_dataset_resume+0x48/0x60 [zfs] [210065.806885] ? send_merge_thread+0x370/0x370 [zfs] [210065.806935] send_traverse_thread+0x50/0xa0 [zfs] [210065.806986] ? enqueue_range+0x170/0x170 [zfs] [210065.807036] thread_generic_wrapper+0x59/0x70 [spl] [210065.807044] ? __thread_exit+0x20/0x20 [spl] [210065.807050] kthread+0x127/0x150 [210065.807054] ? set_kthread_struct+0x50/0x50 [210065.807057] ret_from_fork+0x22/0x30 [210065.807061] </TASK> ```
defect
another encryption bug unencrypted block in encrypted object set thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name nixos distribution version kernel version architecture openzfs version zfs zfs kmod describe the problem you re observing attempting to send a snapshot of an encrypted dataset results in a panic describe how to reproduce the problem how to get to the on disk state that leads to this panic is unclear to me however on the affected machine trying to sync a particular dataset still trying to find a small zfs send command that triggers it quickly reliably triggers it include any warning errors backtraces from the system logs panic unencrypted block in encrypted object set showing stack for process cpu pid comm send traverse tainted p io nixos hardware name fujitsu bios for call trace dump stack lvl vcmn err cold cond resched kmalloc node mutex lock spl kmem zalloc spa log error zfs panic recover send cb traverse visitbp cond resched mutex lock bqueue enqueue impl traverse dnode traverse visitbp traverse visitbp traverse visitbp traverse visitbp traverse visitbp traverse visitbp traverse dnode traverse visitbp zap lookup traverse impl send merge thread set next task fair traverse dataset resume send merge thread send traverse thread enqueue range thread generic wrapper thread exit kthread set kthread struct ret from fork
1
636,257
20,596,156,774
IssuesEvent
2022-03-05 14:32:13
BadRobots1014/BadRobot2022
https://api.github.com/repos/BadRobots1014/BadRobot2022
closed
Use lead-follow mode for drivetrain
Priority: high
Drivetrain motors are currently using a "group" instead of a lead-follow mode.
1.0
Use lead-follow mode for drivetrain - Drivetrain motors are currently using a "group" instead of a lead-follow mode.
non_defect
use lead follow mode for drivetrain drivetrain motors are currently using a group instead of a lead follow mode
0
46,944
2,967,524,772
IssuesEvent
2015-07-13 00:16:53
modxcms/revolution
https://api.github.com/repos/modxcms/revolution
closed
[2.3] Extras Compatibility
area-core bug priority-1-urgent
I think it would be good to collect working and more important not-yet-working extras in one place (maybe for the release note?) as more and more are reported. As far as I can tell, most not-working extras are affected by the connector nuking that happened in 2.3, so they don't find the appropriate processors anymore and thus do not work anymore. Another quite commonly observed issue is the use of the depreciated $modx->getMicrotime() method in extras. Just replacing it with PHP's native function microtime(true) fixes this problem. I suggest commenting (including version info) about successfully and not successfully tested extras in this issue and a I will update the following list. **Extras working and compatible with 2.3** - [x] Ace 1.5.0-pl (@pixelchutes) - [x] ActivationEmail 1.1.2-pl (@BobRay) - [x] AjaxForm-1.0.2-pl (@absent42) - [x] AjaxManager 1.1.8-pl (@whitebyte) - [x] AjaxUpload 1.1.0-pl (@absent42) - [x] Archivist 1.2.4-pl (@absent42) - [x] Babel 2.2.5-pl (@DESIGNfromWITHIN, @goldsky 3.0.0-beta is available for testing) - [x] BannerY 1.3.0-rc1 (@bezumkin) - [x] Batcher 1.1.1-pl (@sottwell) - [x] BigBrother 1.1.0-rc2 (@beau-gosse, doesn't work with PHP 5.4+ though, @exside) - [x] BotBlockX 1.0.2-pl (@BobRay) - [x] BreadCrumb 1.4.2-pl (@rdegler) - [x] Breadcrumbs 1.1.0-pl (@sepiariver) - [x] CacheClear 1.0.3-pl (@exside, @BobRay) - [x] CacheMaster 1.2.1-pl (@BobRay) - [x] Canonical 1.0-Beta1 (@BobRay) - [x] Captcha 3.3.2-pl (@BobRay) - [x] CaseInsensitiveURLs 1.2.0-pl1 (@BobRay) - [x] CheckboxSortable 0.3.0-beta3 (@sottwell) - [x] ClassExtender 2.0.0-pl (@sottwell, @BobRay) - [x] Cliche 1.1.0-rc1 (@bentejuy, has some visual glitches) - [x] ClientConfig 1.3.1-pl (@sottwell, @amdbuilder) - [x] Collections 2.0.0-pl (@theboxer) - [x] ColorPicker 1.0.2-rc2 (@atripolskyy, @pyrographics, @exside) - [x] Console 2.0.1-beta (@exside) - [x] ConvertDatabaseCharset (@BobRay, not a package, see http://bobsguides.com/convert-db-utf8.html) - [x] cookieJar 1.0.0-pl (@davidpede) - [x] Copyright 1.0-pl (@amdbuilder) - [x] CSS Star Rating 2.0-rc1 (@pixelchutes) - [x] cssSweet 1.1.0-rc1 (@sottwell) - [x] CustomUrls 1.0.0-rc3 (@Boddlnagg) - [x] DBAPI 1.0.2-pl, pending release. (@pixelchutes) - [x] DefaultResourceGroup 1.1.0-pl (@BobRay) - [x] DefaultUserGroup 1.1.0-pl (@BobRay) - [x] Dirwalker 1.0.0-pl (@BobRay) - [x] EmailResource 1.1.3 (@BobRay) - [x] EmptyAlias 1.0.0-beta1 (@BobRay) - [x] EZfaq 3.3.0-pl (@BobRay) - [x] fastField 1.4.0-pl (@mintnl) - [x] FileDownload R 1.1.7-pl (@absent42) - [x] FileLister 1.1.1-pl (@mintnl, however keep in mind to use a relative path, so paths starting with /) - [x] FileUpload 1.0.4-pl (@BobRay) - [x] firstPageLink 1.0.1-pl (@Boddlnagg) - [x] FixedPre 1.2.2-pl (@BobRay) - [x] FormIt 2.2.0-pl (@pixelchutes) - [x] Gallery 1.7.0-pl (@pixelchutes) - [x] Gateway Manager 1.1.2-pl (@amdbuilder) - [x] getCache 1.1.0-pl (@sepiariver) - [x] getDate 1.0.0-pl (@davidpede) - [x] getPage 1.2.4-pl (@sepiariver) - [x] getRelated 1.2.0-pl (@sepiariver) - [x] getResourceField 1.0.3-pl1 (@mcnickel) - [x] getResources 1.6.1-pl (@sepiariver) - [x] getUrlParam 1.0-beta1 (@absent42) - [x] getVimeo 1.1.0-pl (@davidpede) - [x] getYoutube 1.0.0-pl (@davidpede) - [x] GoogleAuthenticatorX 1.0.0-rc2 (@minagerges) - [x] GoogleSiteMap 1.3.1-pl (@mcnickel) - [x] Grid Class Key 1.0.1-beta1 (@goldsky, could break with compress_js on @labr1005) - [x] HandyMan 1.0.0-pl (@absent42) - [x] If 1.1.1-pl (@funkyduke) - [x] JSONDerulo 2.4.0-pl (@pdincubus) - [x] LikeDislike 1.0.3-rc1 (@amdbuilder, Works with Display issues) - [x] Login 1.8.1-pl (@pixelchutes) - [x] LogLogins 1.1.0-pl (@BobRay) - [x] LogPageNotFound 1.0.3-pl (@BobRay) - [x] MandrillX 1.0.3-pl (@pyrographics, @BobRay) - [x] MetaX 2.0.2-pl (@amdbuilder) - [x] mhpaypal 1.1.1-rc1 (@absent42) - [x] MIGX 2.8.1-pl, has a lot of styling / UI issues though (@exside), MIGX AJAX Upload (http://rtfm.modx.com/extras/revo/migx/migx.tutorials/migx.use-resource-specific-mediasource-and-multifile-uploader) seems not to work (@skazhikadyadya) - [x] MinifyX 1.3.1-pl (@amdbuilder) - [x] miniShop2 2.1.8-rc2 (@bezumkin) - [x] modAvatar 1.1.0-beta (@absent42) - [x] modMailchimp 1.0.6-pl (@absent42) - [x] MODX Revolution Boilerplate 1.0.1-pl (@pdincubus) - [x] MultiUploadDialog 1.0.2-pl (@absent42) - [x] MyComponent 3.2.2-pl (@pixelchutes, @BobRay) - [x] NewsPublisher 1.4.2-pl (@BobRay, issues with TinyMCE though, @sottwell) - [x] Notify 1.2.3-pl (@BobRay) - [x] ObfuscateEmail-Revo 1.0.0-pl (@absent42) - [x] ObjectExplorer 1.1.2-pl (@BobRay) - [x] Orphans 1.1.0-pl (@BobRay, @sottwell) - [x] PageLocker 1.1.0-beta1 (@gleighfield) - [x] paramStash 1.3.0-pl (@electrickite) - [x] pdoTools 1.9.2-pl1 (@sottwell) - [x] Personalize 3.7.0-pl (@wshawn, @BobRay) - [x] PiwikVisitsSummary 1.0.4-pl (@lektriq) - [x] Polls 1.3.2-pl (@exside, CMP works, Frontend untested) - [x] Profiler 1.1.0-pl (@TheBoxer) - [x] Provisioner 1.1.0-pl, confirmed for importing Evo sites *(did not test Revo import < 2.1.3)* (@pixelchutes) - [x] QuickEmail 1.1.0-pl (@BobRay) - [x] Quickstart Buttons 1.1.1-pl (@Mark-H, needs modmore account/package provider for download, see https://www.modmore.com/extras/quickstartbuttons/, css fix for 2.3.1 see here http://forums.modx.com/thread/92587/quickstart-buttons-addon---css-for-modx-2-3-1) - [x] Quip 2.3.3-pl (@amdbuilder, @mintnl) - [x] Redirector 2.0.5-pl (@beau-gosse, there may well be issues with this extra not related to 2.3 @juiceCake) - [x] ReflectBlock 1.1.0-pl (@BobRay) - [x] RefreshCache 1.1.0-pl (@BobRay) - [x] ResourceExplorer 1.0.0-pl2 (@xgenvn, minor display glitches, fix coming soon) - [x] SE Manager for MODX 0.2.0-pl (@Alroniks) - [x] selfLink 1.0.0-pl (@mcnickel) - [x] SEO Pro 1.0.3-pl (@christianseel) - [x] SEO Tab 1.0.10-pl (@christianseel) - [x] Shibboleth 0.4.0-pl (@electrickite) - [x] Shopkeeper 2.3.7-pl (@zfjoury) - [x] ShowTV 1.0.0-beta1 (@BobRay) - [x] siblingNav 1.0.1-pl (@sottwell) - [x] SimpleSearch 1.9.0-pl (@sepiariver) - [x] SiteAtoZ 1.0.3-beta1 (@BobRay) - [x] SiteCheck 1.0.2-pl (@BobRay, premium Extra https://gumroad.com/l/DLChJ) - [x] sitemapFriend 1.0.2-pl1 (@mcnickel) - [x] SPForm 3.2.2-pl (@BobRay) - [x] StageCoach 1.4.0-pl (@BobRay) - [x] Subscribe 1.2.1-pl (@BobRay) - [x] SyntaxHighlighter 1.0.0-beta1 (@BobRay) - [x] TableofContentsX 1.0.1-pl1 (@MacConin) - [x] Tagger 1.5.0-pl (@TheBoxer) - [x] tagManager2 2.0.0-rc4 (@zfjoury) - [x] ThemePackagerComponent 1.3.0-dev1 (@sottwell confirmed working but non-2.3 related issues, @jeremysanchez reported problems in 2.3.1) - [x] ThermX 3.1.0-pl1 (@BobRay) - [x] TimerangeTV 1.1.0-pl (@exside, @leqtrik) - [x] TrackDuck 1.1.1-pl1 (@shauchenka) - [x] translit 1.0.0-beta (@zfjoury) - [x] TwitterX 1.3.4-pl (@absent42) - [x] UltimateParent 2.0.0-pl (@mcnickel) - [x] UserUrls 0.1.3-beta (@wshawn) - [x] VersionX 2.1.0-pl (@amdbuilder, minor issues https://github.com/Mark-H/VersionX2/issues/67) - [x] Wayfinder 2.3.3-pl (@pixelchutes, @sepiariver) - [x] xBug 0.7.5-pl (@the-dunnock) - [x] XRouting 1.2.0-pl (@labr1005) **Extras working in 2.3 but not showing up in package manager** - [x] AutoTemplate 1.0.0-rc1 (@oblongd) - [x] MarkitUp 0.0.5-pl (@darkarchon) - [x] modxTalks 1.0.3-beta (@sottwell) - [x] sanitizeupload 1.0.0-rc (@absent42) - [x] Theme.Bootstrap 2.1.0-pl (@opus4711) - [x] visualSitemap 1.0.1-beta2 (@wshawn) - [x] xFPC 2.1.0-pl1 (@mcnickel, xFPCAjax not tested, according to @MacConin the plugin gives warnings but seems to work; @funkyduke; @Gallenkamp also reports massive error logging issues with a 2.3.2. installation; @absent42 reports these logging issues for 2.2.x installations too; @sottwell points out that the issues are most possibly coming from setting headers too late) **Extras partially working in 2.3** - [x] Articles 1.7.9-pl (@jpdevries, @Mark-H, @amdbuilder, @hitodev found that it's not possible to remove tags due to the missing "delete"/cross button, see http://forums.modx.com/thread/92526/articles-extra-tags-missing-delete-icon + probable issues together with Tagger 1.5.0-pl, View and close buttons do not seem to work properly) - [x] CKEditor 1.1.1-pl https://github.com/danyaPostfactum/modx-ckeditor/issues/17 (@exside) - [x] CodeMirror 2.2.1-pl (@pdincubus confirmed working in 2.3.1, @sottwell reported issue with longer files, see comment from @charlie17) - [x] ContextRouter 1.1.0-pl2 (@wshawn, 2.3.1 should be fine, in 2.3.0 you need to add a return statement to the plugin) - [x] CronManager 1.1.0-pl (@pyrographics) - [x] FormSave 1.1.0-pl1 (@pixelchutes, @bradbertz When viewing formsave form submission list, scroll down the list seems not to work (see https://github.com/b03tz/FormSave/issues/27). Package not showing up in packager browser - [x] Google Analytics Dashboard Widget 1.0.1-pl (@absent42, @leqtrik, layout is broken and package not showing in package browser, for fix see http://forums.modx.com/thread/92380/google-analytics-dashboard-widget-1-0-1-pl-layout-fix-for-modx-2-3-x or http://forums.modx.com/thread/92525/analytics-dashboard-widget---display-glitch-in-revo-2-3) - [x] Image+ 2.2.0-pl (has quite some UI issues, not really production ready, @exside) - [x] phpThumbOf 1.4.0-pl (affected by #11700, @le-genda) - [x] phpThumbsOn 1.2.1-pl (affected by #11700, @le-genda) - [x] pThumb 2.3.3-pl (@exside, affected by #11700 @le-genda, issue as an output filter, scaling params are ignored when using Resizer @juiceCake) - [x] SuperBoxSelect 1.0.1-pl2 (@enigmatic-user, @exside, UI issues due to additional styles for the already existing superboxselect styles in the core + missing classes for the triggers and tags, see fix in comments below, also not showing in package browser) - [x] TinyMCE 4.3.3-pl, working but some paths need updating, seems non-breaking atm though, maybe some issues with fullscreen-mode (@sepiariver, @sottwell, @linlex, @rainbowtiger) **Extras not working with 2.3** - [x] AdvSearch 1.0.1-pl (@sottwell, @whitebyte error because of depreciated/removed getMicroTime() in 2.3, for a fix see #11749) - [x] ElementHelper 1.3.3-pl (https://github.com/roryg/ElementHelper/issues/26, main issue also the use of deprecated $modx->getMicrotime(), but it's also not showing up in package manager) - [x] Formz 1.1.0-pl (@markwillis82 reported it's working but according to http://forums.modx.com/thread/92765/formz-error-on-revolution-2-3-1#dis-post-506876 saving new forms doesn't work, which renders the extra useless) - [x] Group Edit 1.2.0-pl (@Systerr; Autor recommends using Collections as repalcement) - [x] ImportX 1.1.1-pl (@jmbuytaert reports that it only imports 2 entries) - [x] Janitor 1.6-beta Search tab (port of Evo DocFinder) does not work: `Could not find action file at: controllers/default/index.php`, referencing legacy 2.2 action IDs for updating resources / elements, etc. e.g. `/manager/index.php?a=30&id=16` - [x] mxCalendar 1.1.10-pl (@amdbuilder) - [x] mxFormBuilder 1.0.0-pl (not showing up in package browser and CMP not working according to http://forums.modx.com/thread/92724/mxformbuilder-window-opens-briefly-then-blank#dis-post-506653) - [x] MODXRated 2.1.0-pl (@hugopeek) - [x] PackMan 1.2.0-pl (@exside, CMP works but has some minor UI issues, functionality has issues according to @pdincubus, @minagerges) - [x] revoSSL 1.0.3-pl (@pyrographics) - [x] Xodus 1.4.0-beta (@andreborud, It seems to generate the csv or xls but upon download it just jumps back to the dashboard, package not showing in package browser) **Extras that need testing / confirmation (list incomplete)** - [ ] charcounterTV 1.1.0-pl (not showing up in package browser) - [ ] ColorPicker 1.0-beta1 (not showing in package browser) - [ ] Databackup 1.1.9-pl (not showing in package browser) - [ ] DirectResize 1.3.1-rc1 - [ ] Discuss 1.2-pl - [ ] GoogleMapsTV 1.0-rc3 - [ ] mChimpX 1.0.0-rc1 - [ ] modActiveDirectory 1.0.1-pl - [ ] modxRepository 1.2.3-beta
1.0
[2.3] Extras Compatibility - I think it would be good to collect working and more important not-yet-working extras in one place (maybe for the release note?) as more and more are reported. As far as I can tell, most not-working extras are affected by the connector nuking that happened in 2.3, so they don't find the appropriate processors anymore and thus do not work anymore. Another quite commonly observed issue is the use of the depreciated $modx->getMicrotime() method in extras. Just replacing it with PHP's native function microtime(true) fixes this problem. I suggest commenting (including version info) about successfully and not successfully tested extras in this issue and a I will update the following list. **Extras working and compatible with 2.3** - [x] Ace 1.5.0-pl (@pixelchutes) - [x] ActivationEmail 1.1.2-pl (@BobRay) - [x] AjaxForm-1.0.2-pl (@absent42) - [x] AjaxManager 1.1.8-pl (@whitebyte) - [x] AjaxUpload 1.1.0-pl (@absent42) - [x] Archivist 1.2.4-pl (@absent42) - [x] Babel 2.2.5-pl (@DESIGNfromWITHIN, @goldsky 3.0.0-beta is available for testing) - [x] BannerY 1.3.0-rc1 (@bezumkin) - [x] Batcher 1.1.1-pl (@sottwell) - [x] BigBrother 1.1.0-rc2 (@beau-gosse, doesn't work with PHP 5.4+ though, @exside) - [x] BotBlockX 1.0.2-pl (@BobRay) - [x] BreadCrumb 1.4.2-pl (@rdegler) - [x] Breadcrumbs 1.1.0-pl (@sepiariver) - [x] CacheClear 1.0.3-pl (@exside, @BobRay) - [x] CacheMaster 1.2.1-pl (@BobRay) - [x] Canonical 1.0-Beta1 (@BobRay) - [x] Captcha 3.3.2-pl (@BobRay) - [x] CaseInsensitiveURLs 1.2.0-pl1 (@BobRay) - [x] CheckboxSortable 0.3.0-beta3 (@sottwell) - [x] ClassExtender 2.0.0-pl (@sottwell, @BobRay) - [x] Cliche 1.1.0-rc1 (@bentejuy, has some visual glitches) - [x] ClientConfig 1.3.1-pl (@sottwell, @amdbuilder) - [x] Collections 2.0.0-pl (@theboxer) - [x] ColorPicker 1.0.2-rc2 (@atripolskyy, @pyrographics, @exside) - [x] Console 2.0.1-beta (@exside) - [x] ConvertDatabaseCharset (@BobRay, not a package, see http://bobsguides.com/convert-db-utf8.html) - [x] cookieJar 1.0.0-pl (@davidpede) - [x] Copyright 1.0-pl (@amdbuilder) - [x] CSS Star Rating 2.0-rc1 (@pixelchutes) - [x] cssSweet 1.1.0-rc1 (@sottwell) - [x] CustomUrls 1.0.0-rc3 (@Boddlnagg) - [x] DBAPI 1.0.2-pl, pending release. (@pixelchutes) - [x] DefaultResourceGroup 1.1.0-pl (@BobRay) - [x] DefaultUserGroup 1.1.0-pl (@BobRay) - [x] Dirwalker 1.0.0-pl (@BobRay) - [x] EmailResource 1.1.3 (@BobRay) - [x] EmptyAlias 1.0.0-beta1 (@BobRay) - [x] EZfaq 3.3.0-pl (@BobRay) - [x] fastField 1.4.0-pl (@mintnl) - [x] FileDownload R 1.1.7-pl (@absent42) - [x] FileLister 1.1.1-pl (@mintnl, however keep in mind to use a relative path, so paths starting with /) - [x] FileUpload 1.0.4-pl (@BobRay) - [x] firstPageLink 1.0.1-pl (@Boddlnagg) - [x] FixedPre 1.2.2-pl (@BobRay) - [x] FormIt 2.2.0-pl (@pixelchutes) - [x] Gallery 1.7.0-pl (@pixelchutes) - [x] Gateway Manager 1.1.2-pl (@amdbuilder) - [x] getCache 1.1.0-pl (@sepiariver) - [x] getDate 1.0.0-pl (@davidpede) - [x] getPage 1.2.4-pl (@sepiariver) - [x] getRelated 1.2.0-pl (@sepiariver) - [x] getResourceField 1.0.3-pl1 (@mcnickel) - [x] getResources 1.6.1-pl (@sepiariver) - [x] getUrlParam 1.0-beta1 (@absent42) - [x] getVimeo 1.1.0-pl (@davidpede) - [x] getYoutube 1.0.0-pl (@davidpede) - [x] GoogleAuthenticatorX 1.0.0-rc2 (@minagerges) - [x] GoogleSiteMap 1.3.1-pl (@mcnickel) - [x] Grid Class Key 1.0.1-beta1 (@goldsky, could break with compress_js on @labr1005) - [x] HandyMan 1.0.0-pl (@absent42) - [x] If 1.1.1-pl (@funkyduke) - [x] JSONDerulo 2.4.0-pl (@pdincubus) - [x] LikeDislike 1.0.3-rc1 (@amdbuilder, Works with Display issues) - [x] Login 1.8.1-pl (@pixelchutes) - [x] LogLogins 1.1.0-pl (@BobRay) - [x] LogPageNotFound 1.0.3-pl (@BobRay) - [x] MandrillX 1.0.3-pl (@pyrographics, @BobRay) - [x] MetaX 2.0.2-pl (@amdbuilder) - [x] mhpaypal 1.1.1-rc1 (@absent42) - [x] MIGX 2.8.1-pl, has a lot of styling / UI issues though (@exside), MIGX AJAX Upload (http://rtfm.modx.com/extras/revo/migx/migx.tutorials/migx.use-resource-specific-mediasource-and-multifile-uploader) seems not to work (@skazhikadyadya) - [x] MinifyX 1.3.1-pl (@amdbuilder) - [x] miniShop2 2.1.8-rc2 (@bezumkin) - [x] modAvatar 1.1.0-beta (@absent42) - [x] modMailchimp 1.0.6-pl (@absent42) - [x] MODX Revolution Boilerplate 1.0.1-pl (@pdincubus) - [x] MultiUploadDialog 1.0.2-pl (@absent42) - [x] MyComponent 3.2.2-pl (@pixelchutes, @BobRay) - [x] NewsPublisher 1.4.2-pl (@BobRay, issues with TinyMCE though, @sottwell) - [x] Notify 1.2.3-pl (@BobRay) - [x] ObfuscateEmail-Revo 1.0.0-pl (@absent42) - [x] ObjectExplorer 1.1.2-pl (@BobRay) - [x] Orphans 1.1.0-pl (@BobRay, @sottwell) - [x] PageLocker 1.1.0-beta1 (@gleighfield) - [x] paramStash 1.3.0-pl (@electrickite) - [x] pdoTools 1.9.2-pl1 (@sottwell) - [x] Personalize 3.7.0-pl (@wshawn, @BobRay) - [x] PiwikVisitsSummary 1.0.4-pl (@lektriq) - [x] Polls 1.3.2-pl (@exside, CMP works, Frontend untested) - [x] Profiler 1.1.0-pl (@TheBoxer) - [x] Provisioner 1.1.0-pl, confirmed for importing Evo sites *(did not test Revo import < 2.1.3)* (@pixelchutes) - [x] QuickEmail 1.1.0-pl (@BobRay) - [x] Quickstart Buttons 1.1.1-pl (@Mark-H, needs modmore account/package provider for download, see https://www.modmore.com/extras/quickstartbuttons/, css fix for 2.3.1 see here http://forums.modx.com/thread/92587/quickstart-buttons-addon---css-for-modx-2-3-1) - [x] Quip 2.3.3-pl (@amdbuilder, @mintnl) - [x] Redirector 2.0.5-pl (@beau-gosse, there may well be issues with this extra not related to 2.3 @juiceCake) - [x] ReflectBlock 1.1.0-pl (@BobRay) - [x] RefreshCache 1.1.0-pl (@BobRay) - [x] ResourceExplorer 1.0.0-pl2 (@xgenvn, minor display glitches, fix coming soon) - [x] SE Manager for MODX 0.2.0-pl (@Alroniks) - [x] selfLink 1.0.0-pl (@mcnickel) - [x] SEO Pro 1.0.3-pl (@christianseel) - [x] SEO Tab 1.0.10-pl (@christianseel) - [x] Shibboleth 0.4.0-pl (@electrickite) - [x] Shopkeeper 2.3.7-pl (@zfjoury) - [x] ShowTV 1.0.0-beta1 (@BobRay) - [x] siblingNav 1.0.1-pl (@sottwell) - [x] SimpleSearch 1.9.0-pl (@sepiariver) - [x] SiteAtoZ 1.0.3-beta1 (@BobRay) - [x] SiteCheck 1.0.2-pl (@BobRay, premium Extra https://gumroad.com/l/DLChJ) - [x] sitemapFriend 1.0.2-pl1 (@mcnickel) - [x] SPForm 3.2.2-pl (@BobRay) - [x] StageCoach 1.4.0-pl (@BobRay) - [x] Subscribe 1.2.1-pl (@BobRay) - [x] SyntaxHighlighter 1.0.0-beta1 (@BobRay) - [x] TableofContentsX 1.0.1-pl1 (@MacConin) - [x] Tagger 1.5.0-pl (@TheBoxer) - [x] tagManager2 2.0.0-rc4 (@zfjoury) - [x] ThemePackagerComponent 1.3.0-dev1 (@sottwell confirmed working but non-2.3 related issues, @jeremysanchez reported problems in 2.3.1) - [x] ThermX 3.1.0-pl1 (@BobRay) - [x] TimerangeTV 1.1.0-pl (@exside, @leqtrik) - [x] TrackDuck 1.1.1-pl1 (@shauchenka) - [x] translit 1.0.0-beta (@zfjoury) - [x] TwitterX 1.3.4-pl (@absent42) - [x] UltimateParent 2.0.0-pl (@mcnickel) - [x] UserUrls 0.1.3-beta (@wshawn) - [x] VersionX 2.1.0-pl (@amdbuilder, minor issues https://github.com/Mark-H/VersionX2/issues/67) - [x] Wayfinder 2.3.3-pl (@pixelchutes, @sepiariver) - [x] xBug 0.7.5-pl (@the-dunnock) - [x] XRouting 1.2.0-pl (@labr1005) **Extras working in 2.3 but not showing up in package manager** - [x] AutoTemplate 1.0.0-rc1 (@oblongd) - [x] MarkitUp 0.0.5-pl (@darkarchon) - [x] modxTalks 1.0.3-beta (@sottwell) - [x] sanitizeupload 1.0.0-rc (@absent42) - [x] Theme.Bootstrap 2.1.0-pl (@opus4711) - [x] visualSitemap 1.0.1-beta2 (@wshawn) - [x] xFPC 2.1.0-pl1 (@mcnickel, xFPCAjax not tested, according to @MacConin the plugin gives warnings but seems to work; @funkyduke; @Gallenkamp also reports massive error logging issues with a 2.3.2. installation; @absent42 reports these logging issues for 2.2.x installations too; @sottwell points out that the issues are most possibly coming from setting headers too late) **Extras partially working in 2.3** - [x] Articles 1.7.9-pl (@jpdevries, @Mark-H, @amdbuilder, @hitodev found that it's not possible to remove tags due to the missing "delete"/cross button, see http://forums.modx.com/thread/92526/articles-extra-tags-missing-delete-icon + probable issues together with Tagger 1.5.0-pl, View and close buttons do not seem to work properly) - [x] CKEditor 1.1.1-pl https://github.com/danyaPostfactum/modx-ckeditor/issues/17 (@exside) - [x] CodeMirror 2.2.1-pl (@pdincubus confirmed working in 2.3.1, @sottwell reported issue with longer files, see comment from @charlie17) - [x] ContextRouter 1.1.0-pl2 (@wshawn, 2.3.1 should be fine, in 2.3.0 you need to add a return statement to the plugin) - [x] CronManager 1.1.0-pl (@pyrographics) - [x] FormSave 1.1.0-pl1 (@pixelchutes, @bradbertz When viewing formsave form submission list, scroll down the list seems not to work (see https://github.com/b03tz/FormSave/issues/27). Package not showing up in packager browser - [x] Google Analytics Dashboard Widget 1.0.1-pl (@absent42, @leqtrik, layout is broken and package not showing in package browser, for fix see http://forums.modx.com/thread/92380/google-analytics-dashboard-widget-1-0-1-pl-layout-fix-for-modx-2-3-x or http://forums.modx.com/thread/92525/analytics-dashboard-widget---display-glitch-in-revo-2-3) - [x] Image+ 2.2.0-pl (has quite some UI issues, not really production ready, @exside) - [x] phpThumbOf 1.4.0-pl (affected by #11700, @le-genda) - [x] phpThumbsOn 1.2.1-pl (affected by #11700, @le-genda) - [x] pThumb 2.3.3-pl (@exside, affected by #11700 @le-genda, issue as an output filter, scaling params are ignored when using Resizer @juiceCake) - [x] SuperBoxSelect 1.0.1-pl2 (@enigmatic-user, @exside, UI issues due to additional styles for the already existing superboxselect styles in the core + missing classes for the triggers and tags, see fix in comments below, also not showing in package browser) - [x] TinyMCE 4.3.3-pl, working but some paths need updating, seems non-breaking atm though, maybe some issues with fullscreen-mode (@sepiariver, @sottwell, @linlex, @rainbowtiger) **Extras not working with 2.3** - [x] AdvSearch 1.0.1-pl (@sottwell, @whitebyte error because of depreciated/removed getMicroTime() in 2.3, for a fix see #11749) - [x] ElementHelper 1.3.3-pl (https://github.com/roryg/ElementHelper/issues/26, main issue also the use of deprecated $modx->getMicrotime(), but it's also not showing up in package manager) - [x] Formz 1.1.0-pl (@markwillis82 reported it's working but according to http://forums.modx.com/thread/92765/formz-error-on-revolution-2-3-1#dis-post-506876 saving new forms doesn't work, which renders the extra useless) - [x] Group Edit 1.2.0-pl (@Systerr; Autor recommends using Collections as repalcement) - [x] ImportX 1.1.1-pl (@jmbuytaert reports that it only imports 2 entries) - [x] Janitor 1.6-beta Search tab (port of Evo DocFinder) does not work: `Could not find action file at: controllers/default/index.php`, referencing legacy 2.2 action IDs for updating resources / elements, etc. e.g. `/manager/index.php?a=30&id=16` - [x] mxCalendar 1.1.10-pl (@amdbuilder) - [x] mxFormBuilder 1.0.0-pl (not showing up in package browser and CMP not working according to http://forums.modx.com/thread/92724/mxformbuilder-window-opens-briefly-then-blank#dis-post-506653) - [x] MODXRated 2.1.0-pl (@hugopeek) - [x] PackMan 1.2.0-pl (@exside, CMP works but has some minor UI issues, functionality has issues according to @pdincubus, @minagerges) - [x] revoSSL 1.0.3-pl (@pyrographics) - [x] Xodus 1.4.0-beta (@andreborud, It seems to generate the csv or xls but upon download it just jumps back to the dashboard, package not showing in package browser) **Extras that need testing / confirmation (list incomplete)** - [ ] charcounterTV 1.1.0-pl (not showing up in package browser) - [ ] ColorPicker 1.0-beta1 (not showing in package browser) - [ ] Databackup 1.1.9-pl (not showing in package browser) - [ ] DirectResize 1.3.1-rc1 - [ ] Discuss 1.2-pl - [ ] GoogleMapsTV 1.0-rc3 - [ ] mChimpX 1.0.0-rc1 - [ ] modActiveDirectory 1.0.1-pl - [ ] modxRepository 1.2.3-beta
non_defect
extras compatibility i think it would be good to collect working and more important not yet working extras in one place maybe for the release note as more and more are reported as far as i can tell most not working extras are affected by the connector nuking that happened in so they don t find the appropriate processors anymore and thus do not work anymore another quite commonly observed issue is the use of the depreciated modx getmicrotime method in extras just replacing it with php s native function microtime true fixes this problem i suggest commenting including version info about successfully and not successfully tested extras in this issue and a i will update the following list extras working and compatible with ace pl pixelchutes activationemail pl bobray ajaxform pl ajaxmanager pl whitebyte ajaxupload pl archivist pl babel pl designfromwithin goldsky beta is available for testing bannery bezumkin batcher pl sottwell bigbrother beau gosse doesn t work with php though exside botblockx pl bobray breadcrumb pl rdegler breadcrumbs pl sepiariver cacheclear pl exside bobray cachemaster pl bobray canonical bobray captcha pl bobray caseinsensitiveurls bobray checkboxsortable sottwell classextender pl sottwell bobray cliche bentejuy has some visual glitches clientconfig pl sottwell amdbuilder collections pl theboxer colorpicker atripolskyy pyrographics exside console beta exside convertdatabasecharset bobray not a package see cookiejar pl davidpede copyright pl amdbuilder css star rating pixelchutes csssweet sottwell customurls boddlnagg dbapi pl pending release pixelchutes defaultresourcegroup pl bobray defaultusergroup pl bobray dirwalker pl bobray emailresource bobray emptyalias bobray ezfaq pl bobray fastfield pl mintnl filedownload r pl filelister pl mintnl however keep in mind to use a relative path so paths starting with fileupload pl bobray firstpagelink pl boddlnagg fixedpre pl bobray formit pl pixelchutes gallery pl pixelchutes gateway manager pl amdbuilder getcache pl sepiariver getdate pl davidpede getpage pl sepiariver getrelated pl sepiariver getresourcefield mcnickel getresources pl sepiariver geturlparam getvimeo pl davidpede getyoutube pl davidpede googleauthenticatorx minagerges googlesitemap pl mcnickel grid class key goldsky could break with compress js on handyman pl if pl funkyduke jsonderulo pl pdincubus likedislike amdbuilder works with display issues login pl pixelchutes loglogins pl bobray logpagenotfound pl bobray mandrillx pl pyrographics bobray metax pl amdbuilder mhpaypal migx pl has a lot of styling ui issues though exside migx ajax upload seems not to work skazhikadyadya minifyx pl amdbuilder bezumkin modavatar beta modmailchimp pl modx revolution boilerplate pl pdincubus multiuploaddialog pl mycomponent pl pixelchutes bobray newspublisher pl bobray issues with tinymce though sottwell notify pl bobray obfuscateemail revo pl objectexplorer pl bobray orphans pl bobray sottwell pagelocker gleighfield paramstash pl electrickite pdotools sottwell personalize pl wshawn bobray piwikvisitssummary pl lektriq polls pl exside cmp works frontend untested profiler pl theboxer provisioner pl confirmed for importing evo sites did not test revo import pixelchutes quickemail pl bobray quickstart buttons pl mark h needs modmore account package provider for download see css fix for see here quip pl amdbuilder mintnl redirector pl beau gosse there may well be issues with this extra not related to juicecake reflectblock pl bobray refreshcache pl bobray resourceexplorer xgenvn minor display glitches fix coming soon se manager for modx pl alroniks selflink pl mcnickel seo pro pl christianseel seo tab pl christianseel shibboleth pl electrickite shopkeeper pl zfjoury showtv bobray siblingnav pl sottwell simplesearch pl sepiariver siteatoz bobray sitecheck pl bobray premium extra sitemapfriend mcnickel spform pl bobray stagecoach pl bobray subscribe pl bobray syntaxhighlighter bobray tableofcontentsx macconin tagger pl theboxer zfjoury themepackagercomponent sottwell confirmed working but non related issues jeremysanchez reported problems in thermx bobray timerangetv pl exside leqtrik trackduck shauchenka translit beta zfjoury twitterx pl ultimateparent pl mcnickel userurls beta wshawn versionx pl amdbuilder minor issues wayfinder pl pixelchutes sepiariver xbug pl the dunnock xrouting pl extras working in but not showing up in package manager autotemplate oblongd markitup pl darkarchon modxtalks beta sottwell sanitizeupload rc theme bootstrap pl visualsitemap wshawn xfpc mcnickel xfpcajax not tested according to macconin the plugin gives warnings but seems to work funkyduke gallenkamp also reports massive error logging issues with a installation reports these logging issues for x installations too sottwell points out that the issues are most possibly coming from setting headers too late extras partially working in articles pl jpdevries mark h amdbuilder hitodev found that it s not possible to remove tags due to the missing delete cross button see probable issues together with tagger pl view and close buttons do not seem to work properly ckeditor pl exside codemirror pl pdincubus confirmed working in sottwell reported issue with longer files see comment from contextrouter wshawn should be fine in you need to add a return statement to the plugin cronmanager pl pyrographics formsave pixelchutes bradbertz when viewing formsave form submission list scroll down the list seems not to work see package not showing up in packager browser google analytics dashboard widget pl leqtrik layout is broken and package not showing in package browser for fix see or image pl has quite some ui issues not really production ready exside phpthumbof pl affected by le genda phpthumbson pl affected by le genda pthumb pl exside affected by le genda issue as an output filter scaling params are ignored when using resizer juicecake superboxselect enigmatic user exside ui issues due to additional styles for the already existing superboxselect styles in the core missing classes for the triggers and tags see fix in comments below also not showing in package browser tinymce pl working but some paths need updating seems non breaking atm though maybe some issues with fullscreen mode sepiariver sottwell linlex rainbowtiger extras not working with advsearch pl sottwell whitebyte error because of depreciated removed getmicrotime in for a fix see elementhelper pl main issue also the use of deprecated modx getmicrotime but it s also not showing up in package manager formz pl reported it s working but according to saving new forms doesn t work which renders the extra useless group edit pl systerr autor recommends using collections as repalcement importx pl jmbuytaert reports that it only imports entries janitor beta search tab port of evo docfinder does not work could not find action file at controllers default index php referencing legacy action ids for updating resources elements etc e g manager index php a id mxcalendar pl amdbuilder mxformbuilder pl not showing up in package browser and cmp not working according to modxrated pl hugopeek packman pl exside cmp works but has some minor ui issues functionality has issues according to pdincubus minagerges revossl pl pyrographics xodus beta andreborud it seems to generate the csv or xls but upon download it just jumps back to the dashboard package not showing in package browser extras that need testing confirmation list incomplete charcountertv pl not showing up in package browser colorpicker not showing in package browser databackup pl not showing in package browser directresize discuss pl googlemapstv mchimpx modactivedirectory pl modxrepository beta
0
273,607
20,799,298,643
IssuesEvent
2022-03-17 12:27:14
0chain/0chain
https://api.github.com/repos/0chain/0chain
closed
Buckets concept proposal for optimizing the performance issue of large `list` in MPT
enhancement discussion documentation
We will have performance issue if the nodes number is large. Currently we save all the miners node info saved in a simple MPT node on path `allMinersList`, so each time when we need to update a node's info, we will have to decode the full nodes list, then update the node info, encode the full node list and save it back to MPT. The process would be a disaster when the nodes number is huge. So what we can do to optimize this is, instead of saving all nodes in a single MPT node, we can split the them into several `buckets` that indexed by the miner index id(or something that can identify the miner exclusively) . And each bucket has a fixed capacity, for example: `50`. So the first 50 miners nodes will be saved into `bucket_1` , and nodes from 50 to 100 saved to the second `bucket_2` and so on. We will be able to get the bucket address base on the miner index. In this way, each time when we need to update a miner's info, we will fist get the bucket address, then get the bucket data from the MPT, and decode it into nodes list, then do what we want to do to the node. The benefit is obvious, which is we don't have to encode/decode the full list of nodes every time when we need to update a simple node, and the worst case would be we decode a full bucket, which is obviously much faster than decoding the full nodes list. This bucket concept could be applied to all the `list` like data struct in MPT that could grow large over time. Anyway, we can do this later and fill the details about the `bucket` concept when we finished the Mainnet issues and loadtest issues.
1.0
Buckets concept proposal for optimizing the performance issue of large `list` in MPT - We will have performance issue if the nodes number is large. Currently we save all the miners node info saved in a simple MPT node on path `allMinersList`, so each time when we need to update a node's info, we will have to decode the full nodes list, then update the node info, encode the full node list and save it back to MPT. The process would be a disaster when the nodes number is huge. So what we can do to optimize this is, instead of saving all nodes in a single MPT node, we can split the them into several `buckets` that indexed by the miner index id(or something that can identify the miner exclusively) . And each bucket has a fixed capacity, for example: `50`. So the first 50 miners nodes will be saved into `bucket_1` , and nodes from 50 to 100 saved to the second `bucket_2` and so on. We will be able to get the bucket address base on the miner index. In this way, each time when we need to update a miner's info, we will fist get the bucket address, then get the bucket data from the MPT, and decode it into nodes list, then do what we want to do to the node. The benefit is obvious, which is we don't have to encode/decode the full list of nodes every time when we need to update a simple node, and the worst case would be we decode a full bucket, which is obviously much faster than decoding the full nodes list. This bucket concept could be applied to all the `list` like data struct in MPT that could grow large over time. Anyway, we can do this later and fill the details about the `bucket` concept when we finished the Mainnet issues and loadtest issues.
non_defect
buckets concept proposal for optimizing the performance issue of large list in mpt we will have performance issue if the nodes number is large currently we save all the miners node info saved in a simple mpt node on path allminerslist so each time when we need to update a node s info we will have to decode the full nodes list then update the node info encode the full node list and save it back to mpt the process would be a disaster when the nodes number is huge so what we can do to optimize this is instead of saving all nodes in a single mpt node we can split the them into several buckets that indexed by the miner index id or something that can identify the miner exclusively and each bucket has a fixed capacity for example so the first miners nodes will be saved into bucket and nodes from to saved to the second bucket and so on we will be able to get the bucket address base on the miner index in this way each time when we need to update a miner s info we will fist get the bucket address then get the bucket data from the mpt and decode it into nodes list then do what we want to do to the node the benefit is obvious which is we don t have to encode decode the full list of nodes every time when we need to update a simple node and the worst case would be we decode a full bucket which is obviously much faster than decoding the full nodes list this bucket concept could be applied to all the list like data struct in mpt that could grow large over time anyway we can do this later and fill the details about the bucket concept when we finished the mainnet issues and loadtest issues
0
51,264
21,640,294,241
IssuesEvent
2022-05-05 18:03:00
emergenzeHack/ukrainehelp.emergenzehack.info_segnalazioni
https://api.github.com/repos/emergenzeHack/ukrainehelp.emergenzehack.info_segnalazioni
opened
Centro estivo per bambini Ucraini
Services Children
<pre><yamldata> servicetypes: materialGoods: false hospitality: false transport: false healthcare: false Legal: false translation: false job: false psychologicalSupport: false Children: true disability: false women: false education: false offerFromWho: Impresa Atlantide - Ambasciata Ucraina title: Centro estivo per bambini Ucraini recipients: '' description: L'azienda italiana Atlantide in collaborazione con l'Ambasciata Ucraina in Italia, il Municipio di Roma, il Dipartimento di Difesa Civile d'Italia, nonché le associazioni benefiche "Caritas" e "Sant'Egidio", questa estate presso villa Fassini a Roma ha iniziato l'apertura di un campo estivo per i bambini ucraini url: https://atlantia4ukraine.atlantia.com/it?fbclid=IwAR0WkVHFfg-8Wqf5r831-v-i4mSQGJZES0xE_BTBPH059K9gNQbsR48vLj4 address: mode: autocomplete address: place_id: 106584746 licence: Data © OpenStreetMap contributors, ODbL 1.0. https://osm.org/copyright osm_type: way osm_id: 23936609 boundingbox: - '41.9064355' - '41.9093479' - '12.5467987' - '12.5533818' lat: '41.9073836' lon: '12.5501801' display_name: Via Giuseppe Donati, Casal Bruciato, Collatino, Roma, Roma Capitale, Lazio, 00100, Italia class: highway type: residential importance: 0.5 address: road: Via Giuseppe Donati neighbourhood: Casal Bruciato suburb: Collatino city: Roma county: Roma Capitale state: Lazio postcode: '00100' country: Italia country_code: it iConfirmToHaveReadAndAcceptedInformativeToThreatPersonalData: true label: services submit: true </yamldata></pre>
1.0
Centro estivo per bambini Ucraini - <pre><yamldata> servicetypes: materialGoods: false hospitality: false transport: false healthcare: false Legal: false translation: false job: false psychologicalSupport: false Children: true disability: false women: false education: false offerFromWho: Impresa Atlantide - Ambasciata Ucraina title: Centro estivo per bambini Ucraini recipients: '' description: L'azienda italiana Atlantide in collaborazione con l'Ambasciata Ucraina in Italia, il Municipio di Roma, il Dipartimento di Difesa Civile d'Italia, nonché le associazioni benefiche "Caritas" e "Sant'Egidio", questa estate presso villa Fassini a Roma ha iniziato l'apertura di un campo estivo per i bambini ucraini url: https://atlantia4ukraine.atlantia.com/it?fbclid=IwAR0WkVHFfg-8Wqf5r831-v-i4mSQGJZES0xE_BTBPH059K9gNQbsR48vLj4 address: mode: autocomplete address: place_id: 106584746 licence: Data © OpenStreetMap contributors, ODbL 1.0. https://osm.org/copyright osm_type: way osm_id: 23936609 boundingbox: - '41.9064355' - '41.9093479' - '12.5467987' - '12.5533818' lat: '41.9073836' lon: '12.5501801' display_name: Via Giuseppe Donati, Casal Bruciato, Collatino, Roma, Roma Capitale, Lazio, 00100, Italia class: highway type: residential importance: 0.5 address: road: Via Giuseppe Donati neighbourhood: Casal Bruciato suburb: Collatino city: Roma county: Roma Capitale state: Lazio postcode: '00100' country: Italia country_code: it iConfirmToHaveReadAndAcceptedInformativeToThreatPersonalData: true label: services submit: true </yamldata></pre>
non_defect
centro estivo per bambini ucraini servicetypes materialgoods false hospitality false transport false healthcare false legal false translation false job false psychologicalsupport false children true disability false women false education false offerfromwho impresa atlantide ambasciata ucraina title centro estivo per bambini ucraini recipients description l azienda italiana atlantide in collaborazione con l ambasciata ucraina in italia il municipio di roma il dipartimento di difesa civile d italia nonché le associazioni benefiche caritas e sant egidio questa estate presso villa fassini a roma ha iniziato l apertura di un campo estivo per i bambini ucraini url address mode autocomplete address place id licence data © openstreetmap contributors odbl osm type way osm id boundingbox lat lon display name via giuseppe donati casal bruciato collatino roma roma capitale lazio italia class highway type residential importance address road via giuseppe donati neighbourhood casal bruciato suburb collatino city roma county roma capitale state lazio postcode country italia country code it iconfirmtohavereadandacceptedinformativetothreatpersonaldata true label services submit true
0
9,300
2,615,143,434
IssuesEvent
2015-03-01 06:18:45
chrsmith/html5rocks
https://api.github.com/repos/chrsmith/html5rocks
closed
html5rocks is not HTML5
auto-migrated Priority-Medium Type-Defect
``` Please describe the issue: There are several places in your presentation that is not HTML5 or CSS3 or part of any standardized specification. Look for places containing -webkit* or webkit* function/css rule names. Please consider being honest, not following the Apple's standard of false messages of being standard based. ``` Original issue reported on code.google.com by `jad...@gmail.com` on 23 Jun 2010 at 3:01
1.0
html5rocks is not HTML5 - ``` Please describe the issue: There are several places in your presentation that is not HTML5 or CSS3 or part of any standardized specification. Look for places containing -webkit* or webkit* function/css rule names. Please consider being honest, not following the Apple's standard of false messages of being standard based. ``` Original issue reported on code.google.com by `jad...@gmail.com` on 23 Jun 2010 at 3:01
defect
is not please describe the issue there are several places in your presentation that is not or or part of any standardized specification look for places containing webkit or webkit function css rule names please consider being honest not following the apple s standard of false messages of being standard based original issue reported on code google com by jad gmail com on jun at
1
308,324
9,437,976,044
IssuesEvent
2019-04-13 19:23:26
akalenda/CS249_GraphAlgVisualizer
https://api.github.com/repos/akalenda/CS249_GraphAlgVisualizer
opened
Create unit and acceptance tests
enhancement high priority
By further abstracting construction of the graph and processes into an API and decoupling it from the user interface, we can create unit tests that test the functionality and achieve near-100% coverage. We would then not rely on visual tests of the app to make sure that the underlying functionality is working. Furthermore, an end-to-end test built upon those unit tests would provide an excellent demo of the program in action, as it steps through everything a user might do.
1.0
Create unit and acceptance tests - By further abstracting construction of the graph and processes into an API and decoupling it from the user interface, we can create unit tests that test the functionality and achieve near-100% coverage. We would then not rely on visual tests of the app to make sure that the underlying functionality is working. Furthermore, an end-to-end test built upon those unit tests would provide an excellent demo of the program in action, as it steps through everything a user might do.
non_defect
create unit and acceptance tests by further abstracting construction of the graph and processes into an api and decoupling it from the user interface we can create unit tests that test the functionality and achieve near coverage we would then not rely on visual tests of the app to make sure that the underlying functionality is working furthermore an end to end test built upon those unit tests would provide an excellent demo of the program in action as it steps through everything a user might do
0
61,810
3,154,763,174
IssuesEvent
2015-09-17 02:30:46
xcat2/xcat-core
https://api.github.com/repos/xcat2/xcat-core
reopened
[FVT]2.11 duplicate kit components are added to osimage
priority:normal type:bug
xCAT 2.11 9/15 build on rhels7.2 pp64le 1) download xlc-13.1.2-0-ppc64le.NEED_PRODUCT_PKGS.tar.bz2 2) download other xlf packages which are needed by complete kit [root@c910f03c05k08 build_input]# ls /compilers/ libxlc-13.1.2.0-150526a.ppc64le.rpm libxlc-devel.13.1.2-13.1.2.0-150526a.ppc64le.rpm libxlf-15.1.2.0-150526a.ppc64le.rpm libxlf-devel.15.1.2-15.1.2.0-150526a.ppc64le.rpm libxlmass-devel.8.1.2-8.1.2.0-150526.ppc64le.rpm libxlsmp-4.1.2.0-150526.ppc64le.rpm libxlsmp-devel.4.1.2-4.1.2.0-150526.ppc64le.rpm xlc.13.1.2-13.1.2.0-150526a.ppc64le.rpm xlc-license.13.1.2-13.1.2.0-150526a.ppc64le.rpm xlf.15.1.2-15.1.2.0-150526a.ppc64le.rpm xlf-license.15.1.2-15.1.2.0-150526a.ppc64le.rpm 3) create complete kit using partial kit buildkit addpkgs xlc-13.1.2-0-ppc64le.NEED_PRODUCT_PKGS.tar.bz2 --pkgdir /compilers/ 4) addkit xlc-13.1.2-0-ppc64le.tar.bz2 5) addkitcomp -a -i "rhels7.2-ppc64le-install-compute" xlc.license-compute,xlc.compiler-compute, xlc.rte-compute 6) check osimage rhels7.2-ppc64le-install-compute [root@c910f03c05k08 build_input]# lsdef -t osimage -o rhels7.2-ppc64le-install-compute Object name: rhels7.2-ppc64le-install-compute imagetype=linux kitcomponents=xlc.license-compute-13.1.2-0-rhels-7.2-ppc64le,xlc.rte-compute-13.1.2-0-rhels-7.2-ppc64le,xlc.rte-compute-13.1.2-0-rhels-7.2-ppc64le,xlc.compiler-compute-13.1.2-0-rhels-7.2-ppc64le,xlf.license-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.rte-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.compiler-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.rte-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.license-compute-15.1.2-0-rhels-7.2-ppc64le osarch=ppc64le osdistroname=rhels7.2-ppc64le osname=Linux osvers=rhels7.2 otherpkgdir=/install/post/otherpkgs/rhels7.2/ppc64le otherpkglist=/install/osimages/rhels7.2-ppc64le-install-compute/kits/KIT_DEPLOY_PARAMS.otherpkgs.pkglist,/install/osimages/rhels7.2-ppc64le-install-compute/kits/KIT_COMPONENTS.otherpkgs.pkglist pkgdir=/install/rhels7.2/ppc64le pkglist=/opt/xcat/share/xcat/install/rh/compute.rhels7.pkglist profile=compute provmethod=install template=/opt/xcat/share/xcat/install/rh/compute.rhels7.tmpl
1.0
[FVT]2.11 duplicate kit components are added to osimage - xCAT 2.11 9/15 build on rhels7.2 pp64le 1) download xlc-13.1.2-0-ppc64le.NEED_PRODUCT_PKGS.tar.bz2 2) download other xlf packages which are needed by complete kit [root@c910f03c05k08 build_input]# ls /compilers/ libxlc-13.1.2.0-150526a.ppc64le.rpm libxlc-devel.13.1.2-13.1.2.0-150526a.ppc64le.rpm libxlf-15.1.2.0-150526a.ppc64le.rpm libxlf-devel.15.1.2-15.1.2.0-150526a.ppc64le.rpm libxlmass-devel.8.1.2-8.1.2.0-150526.ppc64le.rpm libxlsmp-4.1.2.0-150526.ppc64le.rpm libxlsmp-devel.4.1.2-4.1.2.0-150526.ppc64le.rpm xlc.13.1.2-13.1.2.0-150526a.ppc64le.rpm xlc-license.13.1.2-13.1.2.0-150526a.ppc64le.rpm xlf.15.1.2-15.1.2.0-150526a.ppc64le.rpm xlf-license.15.1.2-15.1.2.0-150526a.ppc64le.rpm 3) create complete kit using partial kit buildkit addpkgs xlc-13.1.2-0-ppc64le.NEED_PRODUCT_PKGS.tar.bz2 --pkgdir /compilers/ 4) addkit xlc-13.1.2-0-ppc64le.tar.bz2 5) addkitcomp -a -i "rhels7.2-ppc64le-install-compute" xlc.license-compute,xlc.compiler-compute, xlc.rte-compute 6) check osimage rhels7.2-ppc64le-install-compute [root@c910f03c05k08 build_input]# lsdef -t osimage -o rhels7.2-ppc64le-install-compute Object name: rhels7.2-ppc64le-install-compute imagetype=linux kitcomponents=xlc.license-compute-13.1.2-0-rhels-7.2-ppc64le,xlc.rte-compute-13.1.2-0-rhels-7.2-ppc64le,xlc.rte-compute-13.1.2-0-rhels-7.2-ppc64le,xlc.compiler-compute-13.1.2-0-rhels-7.2-ppc64le,xlf.license-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.rte-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.compiler-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.rte-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.license-compute-15.1.2-0-rhels-7.2-ppc64le osarch=ppc64le osdistroname=rhels7.2-ppc64le osname=Linux osvers=rhels7.2 otherpkgdir=/install/post/otherpkgs/rhels7.2/ppc64le otherpkglist=/install/osimages/rhels7.2-ppc64le-install-compute/kits/KIT_DEPLOY_PARAMS.otherpkgs.pkglist,/install/osimages/rhels7.2-ppc64le-install-compute/kits/KIT_COMPONENTS.otherpkgs.pkglist pkgdir=/install/rhels7.2/ppc64le pkglist=/opt/xcat/share/xcat/install/rh/compute.rhels7.pkglist profile=compute provmethod=install template=/opt/xcat/share/xcat/install/rh/compute.rhels7.tmpl
non_defect
duplicate kit components are added to osimage xcat build on download xlc need product pkgs tar download other xlf packages which are needed by complete kit ls compilers libxlc rpm libxlc devel rpm libxlf rpm libxlf devel rpm libxlmass devel rpm libxlsmp rpm libxlsmp devel rpm xlc rpm xlc license rpm xlf rpm xlf license rpm create complete kit using partial kit buildkit addpkgs xlc need product pkgs tar pkgdir compilers addkit xlc tar addkitcomp a i install compute xlc license compute xlc compiler compute xlc rte compute check osimage install compute lsdef t osimage o install compute object name install compute imagetype linux kitcomponents xlc license compute rhels xlc rte compute rhels xlc rte compute rhels xlc compiler compute rhels xlf license compute rhels xlf rte compute rhels xlf compiler compute rhels xlf rte compute rhels xlf license compute rhels osarch osdistroname osname linux osvers otherpkgdir install post otherpkgs otherpkglist install osimages install compute kits kit deploy params otherpkgs pkglist install osimages install compute kits kit components otherpkgs pkglist pkgdir install pkglist opt xcat share xcat install rh compute pkglist profile compute provmethod install template opt xcat share xcat install rh compute tmpl
0
31,856
12,033,132,579
IssuesEvent
2020-04-13 13:39:50
crystal-lang/crystal
https://api.github.com/repos/crystal-lang/crystal
closed
SECURITY: context_spec.cr specifies "only TLS 1.0" by calling tlsv1_method. OpenSSL says avoid tlsv1_method. TLS 1.0 and 1.1 are deprecated.
community:help-wanted security topic:stdlib:crypto
A. OpenSSL docs tell us to avoid calling `tlsv1_method` but crystal is calling it. B. Calling `tlsv1_method` means a TLS/SSL connection will only understand TLS 1.0. C. Current industry recommendation is to use least TLS 1.2: * __June 30, 2018 deadline to drop TLS 1.0__ - The PCI Council (Payment Card Industry) suggested that organizations migrate from TLS 1.0 to TLS 1.1 or higher before June 30, 2018. * __March 2020 deadline to deprecate TLS 1.1__ - In October 2018, Apple, Google, Microsoft, and Mozilla jointly announced they would deprecate TLS 1.0 and 1.1 in March 2020. D. See Background (provided below) for timeline and some attacks on TLS for more context. <hr> ### OpenSSL docs __OpenSSL 1.1.0 says use TLS_method__ and avoid TLSv1_method, TLSv1_1_method, etc. __OpenSSL 1.0.2 says use SSLv23_method__ and avoid TLSv1_method, TLSv1_1_method, etc. <details> <summary>(click to expand) Quote from docs and links</summary><p> >TLS_method(), TLS_server_method(), TLS_client_method() > >These are the general-purpose version-flexible SSL/TLS methods. The actual protocol version used will be negotiated to the highest version mutually supported by the client and the server. The supported protocols are SSLv3, TLSv1, TLSv1.1 and TLSv1.2. Applications should use these methods, and avoid the version-specific methods described below. > ... > TLSv1_2_method(), TLSv1_2_server_method(), TLSv1_2_client_method() > ... > TLSv1_1_method(), TLSv1_1_server_method(), TLSv1_1_client_method() > ... > TLSv1_method(), TLSv1_server_method(), TLSv1_client_method() > A TLS/SSL connection established with these methods will only understand the TLSv1 protocol. * https://www.openssl.org/docs/man1.1.0/man3/TLSv1_method.html * https://www.openssl.org/docs/man1.0.2/man3/TLSv1_method.html </details> ### Relevant code in crystal https://github.com/crystal-lang/crystal/blob/fd0780c3a4ef972a6090e09abc9b09a0e39345ff/spec/std/openssl/ssl/context_spec.cr#L31 https://github.com/crystal-lang/crystal/blob/fd0780c3a4ef972a6090e09abc9b09a0e39345ff/spec/std/openssl/ssl/context_spec.cr#L48 Additionally, TLS 1.0 and TLS 1.1 should be disabled from being chosen during auto-negotiation. See Background (provided below). ## Background __1999__. TLS 1.0 was first defined in RFC 2246 in January 1999. __2006__. TLS 1.1 was defined in RFC 4346 in April 2006. __2008__. TLS 1.2 was defined in RFC 5246 in August 2008. __2014__. TLS 1.0 allows downgrading the connection to SSL 3.0, thus weakening security (POODLE SSL Variant). Source: https://en.wikipedia.org/wiki/Transport_Layer_Security#TLS_1.0 __2014__. TLS 1.0, TLS 1.1, and TLS 1.2 (if not implemented properly) are vulnerable to POODLE TLS Variant even if SSLv3 is disabled. Source: https://en.wikipedia.org/wiki/POODLE __2017__. Google Chrome set TLS 1.3 as the default version for a short time in 2017. It then removed it as the default, due to incompatible middleboxes such as Blue Coat web proxies. Source: https://en.wikipedia.org/wiki/Transport_Layer_Security#TLS_1.3 __2018__. TLS 1.3 was defined in RFC 8446 in August 2018. It mandates use of AEAD ciphers, key exchanges that offer perfect forward secrecy, integrates session hash, and drops support for many insecure or obsolete features. Source: https://en.wikipedia.org/wiki/Transport_Layer_Security#TLS_1.3 #### TLS Interception Appliances POODLE from 2014 is not the only security issue to consider, see this paper: > To analyze TLS-encrypted data, network appliances implement a Man-in-the-Middle TLS proxy, by acting as the intended web server to a requesting client (e.g., a browser), and acting as the client to the actual/outside web server. Source: [The Sorry State of TLS Security in Enterprise - Interception Appliances (PDF, arxiv.org)]( https://arxiv.org/pdf/1809.08729.pdf)
True
SECURITY: context_spec.cr specifies "only TLS 1.0" by calling tlsv1_method. OpenSSL says avoid tlsv1_method. TLS 1.0 and 1.1 are deprecated. - A. OpenSSL docs tell us to avoid calling `tlsv1_method` but crystal is calling it. B. Calling `tlsv1_method` means a TLS/SSL connection will only understand TLS 1.0. C. Current industry recommendation is to use least TLS 1.2: * __June 30, 2018 deadline to drop TLS 1.0__ - The PCI Council (Payment Card Industry) suggested that organizations migrate from TLS 1.0 to TLS 1.1 or higher before June 30, 2018. * __March 2020 deadline to deprecate TLS 1.1__ - In October 2018, Apple, Google, Microsoft, and Mozilla jointly announced they would deprecate TLS 1.0 and 1.1 in March 2020. D. See Background (provided below) for timeline and some attacks on TLS for more context. <hr> ### OpenSSL docs __OpenSSL 1.1.0 says use TLS_method__ and avoid TLSv1_method, TLSv1_1_method, etc. __OpenSSL 1.0.2 says use SSLv23_method__ and avoid TLSv1_method, TLSv1_1_method, etc. <details> <summary>(click to expand) Quote from docs and links</summary><p> >TLS_method(), TLS_server_method(), TLS_client_method() > >These are the general-purpose version-flexible SSL/TLS methods. The actual protocol version used will be negotiated to the highest version mutually supported by the client and the server. The supported protocols are SSLv3, TLSv1, TLSv1.1 and TLSv1.2. Applications should use these methods, and avoid the version-specific methods described below. > ... > TLSv1_2_method(), TLSv1_2_server_method(), TLSv1_2_client_method() > ... > TLSv1_1_method(), TLSv1_1_server_method(), TLSv1_1_client_method() > ... > TLSv1_method(), TLSv1_server_method(), TLSv1_client_method() > A TLS/SSL connection established with these methods will only understand the TLSv1 protocol. * https://www.openssl.org/docs/man1.1.0/man3/TLSv1_method.html * https://www.openssl.org/docs/man1.0.2/man3/TLSv1_method.html </details> ### Relevant code in crystal https://github.com/crystal-lang/crystal/blob/fd0780c3a4ef972a6090e09abc9b09a0e39345ff/spec/std/openssl/ssl/context_spec.cr#L31 https://github.com/crystal-lang/crystal/blob/fd0780c3a4ef972a6090e09abc9b09a0e39345ff/spec/std/openssl/ssl/context_spec.cr#L48 Additionally, TLS 1.0 and TLS 1.1 should be disabled from being chosen during auto-negotiation. See Background (provided below). ## Background __1999__. TLS 1.0 was first defined in RFC 2246 in January 1999. __2006__. TLS 1.1 was defined in RFC 4346 in April 2006. __2008__. TLS 1.2 was defined in RFC 5246 in August 2008. __2014__. TLS 1.0 allows downgrading the connection to SSL 3.0, thus weakening security (POODLE SSL Variant). Source: https://en.wikipedia.org/wiki/Transport_Layer_Security#TLS_1.0 __2014__. TLS 1.0, TLS 1.1, and TLS 1.2 (if not implemented properly) are vulnerable to POODLE TLS Variant even if SSLv3 is disabled. Source: https://en.wikipedia.org/wiki/POODLE __2017__. Google Chrome set TLS 1.3 as the default version for a short time in 2017. It then removed it as the default, due to incompatible middleboxes such as Blue Coat web proxies. Source: https://en.wikipedia.org/wiki/Transport_Layer_Security#TLS_1.3 __2018__. TLS 1.3 was defined in RFC 8446 in August 2018. It mandates use of AEAD ciphers, key exchanges that offer perfect forward secrecy, integrates session hash, and drops support for many insecure or obsolete features. Source: https://en.wikipedia.org/wiki/Transport_Layer_Security#TLS_1.3 #### TLS Interception Appliances POODLE from 2014 is not the only security issue to consider, see this paper: > To analyze TLS-encrypted data, network appliances implement a Man-in-the-Middle TLS proxy, by acting as the intended web server to a requesting client (e.g., a browser), and acting as the client to the actual/outside web server. Source: [The Sorry State of TLS Security in Enterprise - Interception Appliances (PDF, arxiv.org)]( https://arxiv.org/pdf/1809.08729.pdf)
non_defect
security context spec cr specifies only tls by calling method openssl says avoid method tls and are deprecated a openssl docs tell us to avoid calling method but crystal is calling it b calling method means a tls ssl connection will only understand tls c current industry recommendation is to use least tls june deadline to drop tls the pci council payment card industry suggested that organizations migrate from tls to tls or higher before june march deadline to deprecate tls in october apple google microsoft and mozilla jointly announced they would deprecate tls and in march d see background provided below for timeline and some attacks on tls for more context openssl docs openssl says use tls method and avoid method method etc openssl says use method and avoid method method etc click to expand quote from docs and links tls method tls server method tls client method these are the general purpose version flexible ssl tls methods the actual protocol version used will be negotiated to the highest version mutually supported by the client and the server the supported protocols are and applications should use these methods and avoid the version specific methods described below method server method client method method server method client method method server method client method a tls ssl connection established with these methods will only understand the protocol relevant code in crystal additionally tls and tls should be disabled from being chosen during auto negotiation see background provided below background tls was first defined in rfc in january tls was defined in rfc in april tls was defined in rfc in august tls allows downgrading the connection to ssl thus weakening security poodle ssl variant source tls tls and tls if not implemented properly are vulnerable to poodle tls variant even if is disabled source google chrome set tls as the default version for a short time in it then removed it as the default due to incompatible middleboxes such as blue coat web proxies source tls was defined in rfc in august it mandates use of aead ciphers key exchanges that offer perfect forward secrecy integrates session hash and drops support for many insecure or obsolete features source tls interception appliances poodle from is not the only security issue to consider see this paper to analyze tls encrypted data network appliances implement a man in the middle tls proxy by acting as the intended web server to a requesting client e g a browser and acting as the client to the actual outside web server source
0
663,664
22,201,173,844
IssuesEvent
2022-06-07 11:20:29
woocommerce/woocommerce
https://api.github.com/repos/woocommerce/woocommerce
closed
WP CLI db update bug?
type: bug priority: low focus: wc-cli type: task
It seems like [this part of the code](https://github.com/woocommerce/woocommerce/blame/trunk/plugins/woocommerce/includes/cli/class-wc-cli-update-command.php#L64-L72) either uses some magic, or won't run as expected. The while loop probably never runs and result from the callback is ignored. We should check if this actually works.
1.0
WP CLI db update bug? - It seems like [this part of the code](https://github.com/woocommerce/woocommerce/blame/trunk/plugins/woocommerce/includes/cli/class-wc-cli-update-command.php#L64-L72) either uses some magic, or won't run as expected. The while loop probably never runs and result from the callback is ignored. We should check if this actually works.
non_defect
wp cli db update bug it seems like either uses some magic or won t run as expected the while loop probably never runs and result from the callback is ignored we should check if this actually works
0
53,296
28,068,125,528
IssuesEvent
2023-03-29 16:56:41
jupyterlite/jupyterlite
https://api.github.com/repos/jupyterlite/jupyterlite
closed
Be more intentional about some of the imports hacks
performance kernel:pyodide
Could we, to the extent possible, be more intentional about some of the imports hacks so that these things run under `nbconvert --execute` and `nbval`? e.g. ```py try: import pyolite # do hacky things except ImportError: pass ``` It's also worth adding a `<details><summary>Limitations</summary>...</details>` markdown cell above any cells like these to all of our examples. Totally aside, but dang i sure wish there was a good markdown flavor that supported `details`... _Originally posted by @bollwyvl in https://github.com/jtpio/jupyterlite/issues/219#issuecomment-874920456_
True
Be more intentional about some of the imports hacks - Could we, to the extent possible, be more intentional about some of the imports hacks so that these things run under `nbconvert --execute` and `nbval`? e.g. ```py try: import pyolite # do hacky things except ImportError: pass ``` It's also worth adding a `<details><summary>Limitations</summary>...</details>` markdown cell above any cells like these to all of our examples. Totally aside, but dang i sure wish there was a good markdown flavor that supported `details`... _Originally posted by @bollwyvl in https://github.com/jtpio/jupyterlite/issues/219#issuecomment-874920456_
non_defect
be more intentional about some of the imports hacks could we to the extent possible be more intentional about some of the imports hacks so that these things run under nbconvert execute and nbval e g py try import pyolite do hacky things except importerror pass it s also worth adding a limitations markdown cell above any cells like these to all of our examples totally aside but dang i sure wish there was a good markdown flavor that supported details originally posted by bollwyvl in
0
16,692
21,791,963,265
IssuesEvent
2022-05-15 03:02:55
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Batch processing gui for "Export layers to DXF" is missing functionality for loading layers into separate rows in the table
Feedback stale Processing Bug
### What is the bug or the crash? BACKGROUND COMMENTS The "Export layers to DXF" algorithm is relatively unusual (but not unique) because it has a single "input layers" input where more than one input layer can be selected. People are probably relatively unlikely to need to "run as batch process". ISSUE If people do need to "run as batch process", for some reason this algorithm is missing this functionality that is provided with other algorithms: ``` Add Files by Pattern... Select Files... Add All Files from a Directory... Select from Open Layers... ```` Maybe someone thought this functionality isn't necessary for this algorithm? ![image](https://user-images.githubusercontent.com/714600/153734363-f663b01b-db11-4c1d-b1c3-fe55bf9e1186.png) Compare with the vector translate algorithm: ![image](https://user-images.githubusercontent.com/714600/153734477-7d01d59f-f1b4-49cd-be70-818f9798201a.png) And with r.patch, which also has a single "input layers" input where more than one input layer can be selected: ![image](https://user-images.githubusercontent.com/714600/153734449-71d437e9-54ee-489e-81f3-e430fc6b1db9.png) ### Steps to reproduce the issue - click to open "Export layers to DXF" from the processing toolbox - click "run as batch process" in the bottom left - click "Autofill" button in the first column The options are missing... ### Versions QGIS version 3.22.3-Białowieża QGIS code revision 1628765ec7 Qt version 5.15.2 Python version 3.9.5 GDAL/OGR version 3.4.1 PROJ version 8.2.1 EPSG Registry database version v10.041 (2021-12-03) GEOS version 3.10.0-CAPI-1.16.0 SQLite version 3.35.2 PDAL version 2.3.0 PostgreSQL client version 13.0 SpatiaLite version 5.0.1 QWT version 6.1.3 QScintilla2 version 2.11.5 OS version Windows 10 Version 2009 Active Python plugins AnotherDXF2Shape 1.2.3 AutoLayoutTool 1.1 autoSaver 2.6 batchvectorlayersaver 0.9 BulkVectorExport 1.1 CalculateGeometry 0.6.4 changeDataSource 3.1 DataPlotly 3.8.1 deactivate_active_labels 0.5 flowTrace 1.1.1 Generalizer3 1.0 GeoCoding 2.18 geo_sim_processing 1.2.0 getthemfiltered 0.1.3 gridSplitter 0.4.0 GroupStats 2.2.5 HideDocks 0.6.1 ImageServerConnector 2.1.1 ImportPhotos 3.0.3 joinmultiplelines Version 0.4.1 karika 1.5 LayerBoard 1.0.1 linz-data-importer 2.2.3 loadthemall 3.3.0 MagicWand-master 1.3.1 mask 1.10.1 MemoryLayerSaver 4.0.4 mmqgis 2021.9.10 nominatim_locator_filter 0.2.4 numerator 0.2 numericalDigitize 0.4.6 pathfinder version 0.4.1 plaingeometryeditor 3.0.0 plugin_reloader 0.9.1 powerpan 2.0 processing_saga 0.5.0 processing_taudem 3.0.0 processing_whitebox 0.14.0 profiletool 4.2.1 qchainage 3.0.1 QCopycanvas 0.5 qgis-plugin-findreplace-main 1 Qgis2threejs 2.6 QGIS3-getWKT 1.4 QuickMultiAttributeEdit3 version 3.0.3 QuickPrint 3.6.1 quicksaveqml 0.1.5 quick_map_services 0.19.27 rasmover-master version 0.2 redLayer 2.2 selectThemes 3.0.1 simple_tools 0.4.1 splitmultipart 1.0.0 SplitPolygonShowingAreas 0.13 statist 3.2 themeselector 3.2.2 valuetool 3.0.15 ViewshedAnalysis 1.7 volume_calculation_tool 0.4 WaterNetAnalyzer-master 1.7 grassprovider 2.12.99 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
1.0
Batch processing gui for "Export layers to DXF" is missing functionality for loading layers into separate rows in the table - ### What is the bug or the crash? BACKGROUND COMMENTS The "Export layers to DXF" algorithm is relatively unusual (but not unique) because it has a single "input layers" input where more than one input layer can be selected. People are probably relatively unlikely to need to "run as batch process". ISSUE If people do need to "run as batch process", for some reason this algorithm is missing this functionality that is provided with other algorithms: ``` Add Files by Pattern... Select Files... Add All Files from a Directory... Select from Open Layers... ```` Maybe someone thought this functionality isn't necessary for this algorithm? ![image](https://user-images.githubusercontent.com/714600/153734363-f663b01b-db11-4c1d-b1c3-fe55bf9e1186.png) Compare with the vector translate algorithm: ![image](https://user-images.githubusercontent.com/714600/153734477-7d01d59f-f1b4-49cd-be70-818f9798201a.png) And with r.patch, which also has a single "input layers" input where more than one input layer can be selected: ![image](https://user-images.githubusercontent.com/714600/153734449-71d437e9-54ee-489e-81f3-e430fc6b1db9.png) ### Steps to reproduce the issue - click to open "Export layers to DXF" from the processing toolbox - click "run as batch process" in the bottom left - click "Autofill" button in the first column The options are missing... ### Versions QGIS version 3.22.3-Białowieża QGIS code revision 1628765ec7 Qt version 5.15.2 Python version 3.9.5 GDAL/OGR version 3.4.1 PROJ version 8.2.1 EPSG Registry database version v10.041 (2021-12-03) GEOS version 3.10.0-CAPI-1.16.0 SQLite version 3.35.2 PDAL version 2.3.0 PostgreSQL client version 13.0 SpatiaLite version 5.0.1 QWT version 6.1.3 QScintilla2 version 2.11.5 OS version Windows 10 Version 2009 Active Python plugins AnotherDXF2Shape 1.2.3 AutoLayoutTool 1.1 autoSaver 2.6 batchvectorlayersaver 0.9 BulkVectorExport 1.1 CalculateGeometry 0.6.4 changeDataSource 3.1 DataPlotly 3.8.1 deactivate_active_labels 0.5 flowTrace 1.1.1 Generalizer3 1.0 GeoCoding 2.18 geo_sim_processing 1.2.0 getthemfiltered 0.1.3 gridSplitter 0.4.0 GroupStats 2.2.5 HideDocks 0.6.1 ImageServerConnector 2.1.1 ImportPhotos 3.0.3 joinmultiplelines Version 0.4.1 karika 1.5 LayerBoard 1.0.1 linz-data-importer 2.2.3 loadthemall 3.3.0 MagicWand-master 1.3.1 mask 1.10.1 MemoryLayerSaver 4.0.4 mmqgis 2021.9.10 nominatim_locator_filter 0.2.4 numerator 0.2 numericalDigitize 0.4.6 pathfinder version 0.4.1 plaingeometryeditor 3.0.0 plugin_reloader 0.9.1 powerpan 2.0 processing_saga 0.5.0 processing_taudem 3.0.0 processing_whitebox 0.14.0 profiletool 4.2.1 qchainage 3.0.1 QCopycanvas 0.5 qgis-plugin-findreplace-main 1 Qgis2threejs 2.6 QGIS3-getWKT 1.4 QuickMultiAttributeEdit3 version 3.0.3 QuickPrint 3.6.1 quicksaveqml 0.1.5 quick_map_services 0.19.27 rasmover-master version 0.2 redLayer 2.2 selectThemes 3.0.1 simple_tools 0.4.1 splitmultipart 1.0.0 SplitPolygonShowingAreas 0.13 statist 3.2 themeselector 3.2.2 valuetool 3.0.15 ViewshedAnalysis 1.7 volume_calculation_tool 0.4 WaterNetAnalyzer-master 1.7 grassprovider 2.12.99 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
non_defect
batch processing gui for export layers to dxf is missing functionality for loading layers into separate rows in the table what is the bug or the crash background comments the export layers to dxf algorithm is relatively unusual but not unique because it has a single input layers input where more than one input layer can be selected people are probably relatively unlikely to need to run as batch process issue if people do need to run as batch process for some reason this algorithm is missing this functionality that is provided with other algorithms add files by pattern select files add all files from a directory select from open layers maybe someone thought this functionality isn t necessary for this algorithm compare with the vector translate algorithm and with r patch which also has a single input layers input where more than one input layer can be selected steps to reproduce the issue click to open export layers to dxf from the processing toolbox click run as batch process in the bottom left click autofill button in the first column the options are missing versions qgis version białowieża qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version active python plugins autolayouttool autosaver batchvectorlayersaver bulkvectorexport calculategeometry changedatasource dataplotly deactivate active labels flowtrace geocoding geo sim processing getthemfiltered gridsplitter groupstats hidedocks imageserverconnector importphotos joinmultiplelines version karika layerboard linz data importer loadthemall magicwand master mask memorylayersaver mmqgis nominatim locator filter numerator numericaldigitize pathfinder version plaingeometryeditor plugin reloader powerpan processing saga processing taudem processing whitebox profiletool qchainage qcopycanvas qgis plugin findreplace main getwkt version quickprint quicksaveqml quick map services rasmover master version redlayer selectthemes simple tools splitmultipart splitpolygonshowingareas statist themeselector valuetool viewshedanalysis volume calculation tool waternetanalyzer master grassprovider processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
0
47,199
13,056,052,690
IssuesEvent
2020-07-30 03:30:40
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
how not to clobber LD_LIBRARY_PATH in your env-shell.sh docs (Trac #124)
Migrated from Trac defect documentation
done in new .rst docs Migrated from https://code.icecube.wisc.edu/ticket/124 ```json { "status": "closed", "changetime": "2014-11-23T03:37:56", "description": "done in new .rst docs\n\n", "reporter": "troy", "cc": "", "resolution": "fixed", "_ts": "1416713876900096", "component": "documentation", "summary": "how not to clobber LD_LIBRARY_PATH in your env-shell.sh docs", "priority": "normal", "keywords": "", "time": "2008-09-06T18:48:11", "milestone": "", "owner": "troy", "type": "defect" } ```
1.0
how not to clobber LD_LIBRARY_PATH in your env-shell.sh docs (Trac #124) - done in new .rst docs Migrated from https://code.icecube.wisc.edu/ticket/124 ```json { "status": "closed", "changetime": "2014-11-23T03:37:56", "description": "done in new .rst docs\n\n", "reporter": "troy", "cc": "", "resolution": "fixed", "_ts": "1416713876900096", "component": "documentation", "summary": "how not to clobber LD_LIBRARY_PATH in your env-shell.sh docs", "priority": "normal", "keywords": "", "time": "2008-09-06T18:48:11", "milestone": "", "owner": "troy", "type": "defect" } ```
defect
how not to clobber ld library path in your env shell sh docs trac done in new rst docs migrated from json status closed changetime description done in new rst docs n n reporter troy cc resolution fixed ts component documentation summary how not to clobber ld library path in your env shell sh docs priority normal keywords time milestone owner troy type defect
1
70,609
23,262,903,751
IssuesEvent
2022-08-04 14:50:45
NREL/EnergyPlus
https://api.github.com/repos/NREL/EnergyPlus
closed
FuelFactors: Units of Measures and Energy per Unit Factor are never used
Defect
Issue overview -------------- IDD: ``` FuelFactors, [...] A2, \field Units of Measure N1, \field Energy per Unit Factor ``` https://github.com/NREL/EnergyPlus/blob/f39484f14fe4df6874c0334c90d79916767cf6ad/src/EnergyPlus/PollutionModule.cc#L260 As you can see, `cAlphaArgs(2)` is never used (this is the Units of Measure), and neither is `rNumericArgs(1)` (Energy per Unit Factor) Either use them or remove them from the IDD **Additional issue:** Wrong index due to copy paste here: https://github.com/NREL/EnergyPlus/blob/f39484f14fe4df6874c0334c90d79916767cf6ad/src/EnergyPlus/PollutionModule.cc#L248-L251 Should be `3` in the `if` ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version): all - Version of EnergyPlus (if using an intermediate build, include SHA): - Unmethours link or helpdesk ticket number ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [x] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
1.0
FuelFactors: Units of Measures and Energy per Unit Factor are never used - Issue overview -------------- IDD: ``` FuelFactors, [...] A2, \field Units of Measure N1, \field Energy per Unit Factor ``` https://github.com/NREL/EnergyPlus/blob/f39484f14fe4df6874c0334c90d79916767cf6ad/src/EnergyPlus/PollutionModule.cc#L260 As you can see, `cAlphaArgs(2)` is never used (this is the Units of Measure), and neither is `rNumericArgs(1)` (Energy per Unit Factor) Either use them or remove them from the IDD **Additional issue:** Wrong index due to copy paste here: https://github.com/NREL/EnergyPlus/blob/f39484f14fe4df6874c0334c90d79916767cf6ad/src/EnergyPlus/PollutionModule.cc#L248-L251 Should be `3` in the `if` ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version): all - Version of EnergyPlus (if using an intermediate build, include SHA): - Unmethours link or helpdesk ticket number ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [x] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
defect
fuelfactors units of measures and energy per unit factor are never used issue overview idd fuelfactors field units of measure field energy per unit factor as you can see calphaargs is never used this is the units of measure and neither is rnumericargs energy per unit factor either use them or remove them from the idd additional issue wrong index due to copy paste here should be in the if details some additional details for this issue if relevant platform operating system version all version of energyplus if using an intermediate build include sha unmethours link or helpdesk ticket number checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
1
8,230
2,611,472,162
IssuesEvent
2015-02-27 05:16:36
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
turns should not be possible while using blowtorch
auto-migrated Priority-Low Type-Defect
``` What steps will reproduce the problem? * see the video at http://www.youtube.com/watch?v=a9-3_8NQpUk What is the expected output? What do you see instead? * it should not be possible for user to turn 180 degrees once blowtorch is activated What version of the product are you using? On what operating system? * 0.9.15 Ubuntu 10.04 ``` Original issue reported on code.google.com by `jose1...@gmail.com` on 19 Jul 2011 at 11:58
1.0
turns should not be possible while using blowtorch - ``` What steps will reproduce the problem? * see the video at http://www.youtube.com/watch?v=a9-3_8NQpUk What is the expected output? What do you see instead? * it should not be possible for user to turn 180 degrees once blowtorch is activated What version of the product are you using? On what operating system? * 0.9.15 Ubuntu 10.04 ``` Original issue reported on code.google.com by `jose1...@gmail.com` on 19 Jul 2011 at 11:58
defect
turns should not be possible while using blowtorch what steps will reproduce the problem see the video at what is the expected output what do you see instead it should not be possible for user to turn degrees once blowtorch is activated what version of the product are you using on what operating system ubuntu original issue reported on code google com by gmail com on jul at
1
28,060
5,170,553,100
IssuesEvent
2017-01-18 06:50:02
TNGSB/eWallet
https://api.github.com/repos/TNGSB/eWallet
closed
e-Wallet_Mobile App (My Inbox - Content) 17012017 #21
Defect - Medium (Sev-3)
All of the inbox content displayed "Merchant Reference No." as "Null". Attached screenshot for POT. ![screenshot_20170117-155831](https://cloud.githubusercontent.com/assets/23113211/22012277/187a7ffa-dcce-11e6-82d2-7796fa535edf.png) ![screenshot_20170117-155945](https://cloud.githubusercontent.com/assets/23113211/22012276/187a5930-dcce-11e6-9ab8-b86892649806.png) ![screenshot_20170117-155951](https://cloud.githubusercontent.com/assets/23113211/22012279/187cdbd8-dcce-11e6-95dc-52ba352389ee.png) ![screenshot_20170117-155956](https://cloud.githubusercontent.com/assets/23113211/22012278/187cd2f0-dcce-11e6-8dd0-1cc6647b9a80.png)
1.0
e-Wallet_Mobile App (My Inbox - Content) 17012017 #21 - All of the inbox content displayed "Merchant Reference No." as "Null". Attached screenshot for POT. ![screenshot_20170117-155831](https://cloud.githubusercontent.com/assets/23113211/22012277/187a7ffa-dcce-11e6-82d2-7796fa535edf.png) ![screenshot_20170117-155945](https://cloud.githubusercontent.com/assets/23113211/22012276/187a5930-dcce-11e6-9ab8-b86892649806.png) ![screenshot_20170117-155951](https://cloud.githubusercontent.com/assets/23113211/22012279/187cdbd8-dcce-11e6-95dc-52ba352389ee.png) ![screenshot_20170117-155956](https://cloud.githubusercontent.com/assets/23113211/22012278/187cd2f0-dcce-11e6-8dd0-1cc6647b9a80.png)
defect
e wallet mobile app my inbox content all of the inbox content displayed merchant reference no as null attached screenshot for pot
1
393,476
26,993,605,194
IssuesEvent
2023-02-09 22:11:19
kitsudaiki/Hanami-AI
https://api.github.com/repos/kitsudaiki/Hanami-AI
opened
Use cert-manager
Documentation Improvement Security Torii Deploy-Process
## Feature-request ### Description Use the cert-manager in kubernetes-setup for creating certificates for the https-connection. At the moment these are still hard-coded, which has to be fixed.
1.0
Use cert-manager - ## Feature-request ### Description Use the cert-manager in kubernetes-setup for creating certificates for the https-connection. At the moment these are still hard-coded, which has to be fixed.
non_defect
use cert manager feature request description use the cert manager in kubernetes setup for creating certificates for the https connection at the moment these are still hard coded which has to be fixed
0
10,277
2,622,140,467
IssuesEvent
2015-03-04 00:01:48
byzhang/spserver
https://api.github.com/repos/byzhang/spserver
closed
SP_HttpEchoHandler如何停掉?
auto-migrated Priority-Medium Type-Defect
``` 下载了你的SPSERVER测试,发现使用很方便,但有一个问题我比较� ��惑,在testiocphttp 里,怎么断开一个客户端连接呢.在例子里好象是keep-alive,这样� ��些恶意的连接不断开 怎么办?我还不是很清楚这个过程,望不吝赐教.3X ``` Original issue reported on code.google.com by `VxS...@gmail.com` on 20 Apr 2009 at 9:15
1.0
SP_HttpEchoHandler如何停掉? - ``` 下载了你的SPSERVER测试,发现使用很方便,但有一个问题我比较� ��惑,在testiocphttp 里,怎么断开一个客户端连接呢.在例子里好象是keep-alive,这样� ��些恶意的连接不断开 怎么办?我还不是很清楚这个过程,望不吝赐教.3X ``` Original issue reported on code.google.com by `VxS...@gmail.com` on 20 Apr 2009 at 9:15
defect
sp httpechohandler如何停掉 下载了你的spserver测试 发现使用很方便 但有一个问题我比较� ��惑 在testiocphttp 里 怎么断开一个客户端连接呢 在例子里好象是keep alive 这样� ��些恶意的连接不断开 怎么办 我还不是很清楚这个过程 望不吝赐教 original issue reported on code google com by vxs gmail com on apr at
1
113,435
9,646,262,967
IssuesEvent
2019-05-17 10:46:20
appium/appium
https://api.github.com/repos/appium/appium
closed
Result of iOS page source becomes unexpectedly large and time-out happens
NeedsInfo ThirdParty XCUITest
## The problem When I tries to retrieve the XML tree of my customer's iOS app by `driver.source` command, unexpectedly only the root element is returned. After I changed [AX_TIMEOUT](https://github.com/appium/WebDriverAgent/blob/master/WebDriverAgentLib/Categories/XCUIElement%2BFBUtilities.m#L77) to `300.` and rebuilt WebDriverAgent by myself, I could retrieve [very large tree](https://gist.githubusercontent.com/NozomiIto/cfd35c6a6281724e731cb1ef036fe20f/raw/ea5e7311037d1ec827a4fda56be401f8020880b1/Large%2520XML%2520tree). It contains about 300 `XCUIElementTypeCell` (most of them are `visible=false`) and I guess it makes the XML tree retrieval slow. But what is strange is, the actual app does not have such so many TableViewCell elements. When the customer checked the [view hierarchy by Xcode](https://developer.apple.com/library/archive/documentation/DeveloperTools/Conceptual/debugging_with_xcode/chapters/special_debugging_workflows.html#//apple_ref/doc/uid/TP40015022-CH9-SW2), TableViewCells are only about 8. Why source command returns so many elements..? ## Environment * Appium 1.13.0. As I described about, I changed AX_TIMEOUT and rebuilt WebDriverAgent. * Desktop OS/version used to run Appium: MacOS Mojave 10.14.4 * Node.js version: v10.11.0 * Npm or Yarn package manager: 6.4.1 * Mobile platform/version under test: iOS real device. iPad mini2. iOS12.2. (But this is also the same on iPhone8/iOS12.1/simulator) * Appium CLI ## Details - The back ground Cell elements continue to move upward slowly, since they contain the background image. This move may be related to the problem. - I don't have the access right to the customer's app's source code. But I think they can offer necessary app's information to resolve the problem. - To be exact, the user uses custom component whose name is ***.CollectionTableViewCell - I attach the image of view hierarchy by Xcode. This is for a little different page than the page I retrieved the XML, but most of TableViewCell's structure and large XML problem is the same. ![MaskedImage](https://user-images.githubusercontent.com/4276719/57908106-2f0ad500-78ba-11e9-821b-fc3f6cf73c00.png) ## Link to Appium logs https://gist.github.com/NozomiIto/0b529fe6b9ae50ea44d45f02a209023b
1.0
Result of iOS page source becomes unexpectedly large and time-out happens - ## The problem When I tries to retrieve the XML tree of my customer's iOS app by `driver.source` command, unexpectedly only the root element is returned. After I changed [AX_TIMEOUT](https://github.com/appium/WebDriverAgent/blob/master/WebDriverAgentLib/Categories/XCUIElement%2BFBUtilities.m#L77) to `300.` and rebuilt WebDriverAgent by myself, I could retrieve [very large tree](https://gist.githubusercontent.com/NozomiIto/cfd35c6a6281724e731cb1ef036fe20f/raw/ea5e7311037d1ec827a4fda56be401f8020880b1/Large%2520XML%2520tree). It contains about 300 `XCUIElementTypeCell` (most of them are `visible=false`) and I guess it makes the XML tree retrieval slow. But what is strange is, the actual app does not have such so many TableViewCell elements. When the customer checked the [view hierarchy by Xcode](https://developer.apple.com/library/archive/documentation/DeveloperTools/Conceptual/debugging_with_xcode/chapters/special_debugging_workflows.html#//apple_ref/doc/uid/TP40015022-CH9-SW2), TableViewCells are only about 8. Why source command returns so many elements..? ## Environment * Appium 1.13.0. As I described about, I changed AX_TIMEOUT and rebuilt WebDriverAgent. * Desktop OS/version used to run Appium: MacOS Mojave 10.14.4 * Node.js version: v10.11.0 * Npm or Yarn package manager: 6.4.1 * Mobile platform/version under test: iOS real device. iPad mini2. iOS12.2. (But this is also the same on iPhone8/iOS12.1/simulator) * Appium CLI ## Details - The back ground Cell elements continue to move upward slowly, since they contain the background image. This move may be related to the problem. - I don't have the access right to the customer's app's source code. But I think they can offer necessary app's information to resolve the problem. - To be exact, the user uses custom component whose name is ***.CollectionTableViewCell - I attach the image of view hierarchy by Xcode. This is for a little different page than the page I retrieved the XML, but most of TableViewCell's structure and large XML problem is the same. ![MaskedImage](https://user-images.githubusercontent.com/4276719/57908106-2f0ad500-78ba-11e9-821b-fc3f6cf73c00.png) ## Link to Appium logs https://gist.github.com/NozomiIto/0b529fe6b9ae50ea44d45f02a209023b
non_defect
result of ios page source becomes unexpectedly large and time out happens the problem when i tries to retrieve the xml tree of my customer s ios app by driver source command unexpectedly only the root element is returned after i changed to and rebuilt webdriveragent by myself i could retrieve it contains about xcuielementtypecell most of them are visible false and i guess it makes the xml tree retrieval slow but what is strange is the actual app does not have such so many tableviewcell elements when the customer checked the tableviewcells are only about why source command returns so many elements environment appium as i described about i changed ax timeout and rebuilt webdriveragent desktop os version used to run appium macos mojave node js version npm or yarn package manager mobile platform version under test ios real device ipad but this is also the same on simulator appium cli details the back ground cell elements continue to move upward slowly since they contain the background image this move may be related to the problem i don t have the access right to the customer s app s source code but i think they can offer necessary app s information to resolve the problem to be exact the user uses custom component whose name is collectiontableviewcell i attach the image of view hierarchy by xcode this is for a little different page than the page i retrieved the xml but most of tableviewcell s structure and large xml problem is the same link to appium logs
0
28,842
5,390,067,332
IssuesEvent
2017-02-25 10:02:39
patmun/pynetdicom
https://api.github.com/repos/patmun/pynetdicom
closed
Migration to GitHub (or another place)
auto-migrated Priority-Medium Type-Defect
``` Google Code will become read only in less than a month, and the project should be moved somewhere else. Are you going to mantain it? Are you going to move it somewhere else? Should we just fork it? Thanks! ``` Original issue reported on code.google.com by `gafo...@gmail.com` on 4 Aug 2015 at 7:46
1.0
Migration to GitHub (or another place) - ``` Google Code will become read only in less than a month, and the project should be moved somewhere else. Are you going to mantain it? Are you going to move it somewhere else? Should we just fork it? Thanks! ``` Original issue reported on code.google.com by `gafo...@gmail.com` on 4 Aug 2015 at 7:46
defect
migration to github or another place google code will become read only in less than a month and the project should be moved somewhere else are you going to mantain it are you going to move it somewhere else should we just fork it thanks original issue reported on code google com by gafo gmail com on aug at
1
99,620
16,449,115,415
IssuesEvent
2021-05-21 01:06:47
AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches
https://api.github.com/repos/AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches
opened
CVE-2021-29579 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl
security vulnerability
## CVE-2021-29579 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /FinalProject/requirements.txt</p> <p>Path to vulnerable library: teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. The implementation of `tf.raw_ops.MaxPoolGrad` is vulnerable to a heap buffer overflow. The implementation(https://github.com/tensorflow/tensorflow/blob/ab1e644b48c82cb71493f4362b4dd38f4577a1cf/tensorflow/core/kernels/maxpooling_op.cc#L194-L203) fails to validate that indices used to access elements of input/output arrays are valid. Whereas accesses to `input_backprop_flat` are guarded by `FastBoundsCheck`, the indexing in `out_backprop_flat` can result in OOB access. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range. <p>Publish Date: 2021-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29579>CVE-2021-29579</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79fv-9865-4qcv">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79fv-9865-4qcv</a></p> <p>Release Date: 2021-05-14</p> <p>Fix Resolution: tensorflow - 2.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-29579 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2021-29579 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /FinalProject/requirements.txt</p> <p>Path to vulnerable library: teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. The implementation of `tf.raw_ops.MaxPoolGrad` is vulnerable to a heap buffer overflow. The implementation(https://github.com/tensorflow/tensorflow/blob/ab1e644b48c82cb71493f4362b4dd38f4577a1cf/tensorflow/core/kernels/maxpooling_op.cc#L194-L203) fails to validate that indices used to access elements of input/output arrays are valid. Whereas accesses to `input_backprop_flat` are guarded by `FastBoundsCheck`, the indexing in `out_backprop_flat` can result in OOB access. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range. <p>Publish Date: 2021-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29579>CVE-2021-29579</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79fv-9865-4qcv">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79fv-9865-4qcv</a></p> <p>Release Date: 2021-05-14</p> <p>Fix Resolution: tensorflow - 2.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file finalproject requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an end to end open source platform for machine learning the implementation of tf raw ops maxpoolgrad is vulnerable to a heap buffer overflow the implementation fails to validate that indices used to access elements of input output arrays are valid whereas accesses to input backprop flat are guarded by fastboundscheck the indexing in out backprop flat can result in oob access the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow step up your open source security game with whitesource
0
41,887
10,694,431,226
IssuesEvent
2019-10-23 10:50:42
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
SIGSEGV com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization
Priority: High Team: Core Type: Critical Type: Defect Type: Test-Failure
2 different test are crashing the member JVM http://jenkins.hazelcast.com/view/wan/job/wan-sync-oome/28/console http://jenkins.hazelcast.com/view/split/job/split-quorum/35/console ``` # # A fatal error has been detected by the Java Runtime Environment: # # SIGSEGV (0xb) at pc=0x00007f6841db8a69, pid=2476, tid=0x00007f68393a5700 # # JRE version: Java(TM) SE Runtime Environment (8.0_201-b09) (build 1.8.0_201-b09) # Java VM: Java HotSpot(TM) 64-Bit Server VM (25.201-b09 mixed mode linux-amd64 compressed oops) # Problematic frame: # J 4782 C2 com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization /Data;)V (128 bytes) @ 0x00007f6841db8a69 [0x00007f6841db8960+0x109] # # Core dump written. Default location: /home/ec2-user/hz-root/HzMember2B/core or core.2476 # # If you would like to submit a bug report, please visit: # http://bugreport.java.com/bugreport/crash.jsp # ``` ``` Native frames: (J=compiled Java code, j=interpreted, Vv=VM code, C=native code) J 4782 C2 com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization/D ata;)V (128 bytes) @ 0x00007f6841db8a69 [0x00007f6841db8960+0x109] J 5103 C2 com.hazelcast.map.impl.operation.PutAllBackupOperation.writeInternal(Lcom/hazelcast/nio/ObjectDataOutput;)V (92 bytes) @ 0x00007f684190c6bc [0x00007f684190c460+0x25c] J 4162 C2 com.hazelcast.spi.impl.operationservice.Operation.writeData(Lcom/hazelcast/nio/ObjectDataOutput;)V (230 bytes) @ 0x00007f6841b4d770 [0x00007f6841b4d200+0x570] J 4235 C2 com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.write(Lcom/hazelcast/nio/ObjectDataOutput;Ljava/lang/Object;)V (10 bytes) @ 0x00007f6841b604 74 [0x00007f6841b60320+0x154] J 4187 C2 com.hazelcast.spi.impl.operationservice.impl.operations.Backup.writeInternal(Lcom/hazelcast/nio/ObjectDataOutput;)V (164 bytes) @ 0x00007f6841b5b92c [0x00007f6841b5b700+0x 22c] J 4162 C2 com.hazelcast.spi.impl.operationservice.Operation.writeData(Lcom/hazelcast/nio/ObjectDataOutput;)V (230 bytes) @ 0x00007f6841b4d770 [0x00007f6841b4d200+0x570] J 4235 C2 com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.write(Lcom/hazelcast/nio/ObjectDataOutput;Ljava/lang/Object;)V (10 bytes) @ 0x00007f6841b604 74 [0x00007f6841b60320+0x154] J 5407 C2 com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(Ljava/lang/Object;IZLcom/hazelcast/partition/PartitioningStrategy;Ljava/nio/ByteOrder;)[B (1 39 bytes) @ 0x00007f68416232a4 [0x00007f6841622ea0+0x404] J 5142 C2 com.hazelcast.spi.impl.operationservice.impl.OutboundOperationHandler.send(Lcom/hazelcast/spi/impl/operationservice/Operation;Lcom/hazelcast/cluster/Address;)Z (83 bytes) @ 0x00007f6841758ea0 [0x00007f6841758b80+0x320] J 5149 C2 com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.handleResponse(Lcom/hazelcast/spi/impl/operationservice/Operation;)V (52 bytes) @ 0x00007f6841b280a4 [0x00 007f6841b27100+0xfa4] J 5365 C2 com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(Lcom/hazelcast/spi/impl/operationservice/Operation;)V (139 bytes) @ 0x00007f6841185614 [0x00007f68411 85560+0xb4] J 5054 C2 com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(Ljava/lang/Object;)V (206 bytes) @ 0x00007f6841e53d44 [0x00007f6841e539e0+0x364] J 5070% C2 com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.run()V (101 bytes) @ 0x00007f6841169dc8 [0x00007f6841169d20+0xa8] v ~StubRoutines::call_stub ``` Danny Conlon 1:45 PM http://54.147.27.51/~jenkins/workspace/split-quorum/4.0-SNAPSHOT/2019_10_23-07_52_53/quorum/output/HZ/HzMember3HZBB/hs_err_pid2321.log core dump in this run http://jenkins.hazelcast.com/view/split/job/split-quorum/35/console ./output/HZ/HzMember3HZBB/hs_err_pid2302.log ./output/HZ/HzMember1HZBB/hs_err_pid2301.log ./output/HZ/HzMember2HZBB/hs_err_pid2318.log ./output/HZ/HzMember3HZBB/core.2302 ./output/HZ/HzMember1HZBB/core.2301 ./output/HZ/HzMember2HZBB/core.2318 I also a different test which also crashes the JVM http://jenkins.hazelcast.com/view/wan/job/wan-sync-oome/28/console ``` # Problematic frame: # J 4782 C2 com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization /Data;)V (128 bytes) @ 0x00007f6841db8a69 [0x00007f6841db8960+0x109] ``` probably has same root cause http://54.147.27.51/~jenkins/workspace/wan-sync-oome/4.0-SNAPSHOT/2019_10_23-10_10_48/sync-oome/ output/B/HzMember2B/core.2476
1.0
SIGSEGV com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization - 2 different test are crashing the member JVM http://jenkins.hazelcast.com/view/wan/job/wan-sync-oome/28/console http://jenkins.hazelcast.com/view/split/job/split-quorum/35/console ``` # # A fatal error has been detected by the Java Runtime Environment: # # SIGSEGV (0xb) at pc=0x00007f6841db8a69, pid=2476, tid=0x00007f68393a5700 # # JRE version: Java(TM) SE Runtime Environment (8.0_201-b09) (build 1.8.0_201-b09) # Java VM: Java HotSpot(TM) 64-Bit Server VM (25.201-b09 mixed mode linux-amd64 compressed oops) # Problematic frame: # J 4782 C2 com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization /Data;)V (128 bytes) @ 0x00007f6841db8a69 [0x00007f6841db8960+0x109] # # Core dump written. Default location: /home/ec2-user/hz-root/HzMember2B/core or core.2476 # # If you would like to submit a bug report, please visit: # http://bugreport.java.com/bugreport/crash.jsp # ``` ``` Native frames: (J=compiled Java code, j=interpreted, Vv=VM code, C=native code) J 4782 C2 com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization/D ata;)V (128 bytes) @ 0x00007f6841db8a69 [0x00007f6841db8960+0x109] J 5103 C2 com.hazelcast.map.impl.operation.PutAllBackupOperation.writeInternal(Lcom/hazelcast/nio/ObjectDataOutput;)V (92 bytes) @ 0x00007f684190c6bc [0x00007f684190c460+0x25c] J 4162 C2 com.hazelcast.spi.impl.operationservice.Operation.writeData(Lcom/hazelcast/nio/ObjectDataOutput;)V (230 bytes) @ 0x00007f6841b4d770 [0x00007f6841b4d200+0x570] J 4235 C2 com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.write(Lcom/hazelcast/nio/ObjectDataOutput;Ljava/lang/Object;)V (10 bytes) @ 0x00007f6841b604 74 [0x00007f6841b60320+0x154] J 4187 C2 com.hazelcast.spi.impl.operationservice.impl.operations.Backup.writeInternal(Lcom/hazelcast/nio/ObjectDataOutput;)V (164 bytes) @ 0x00007f6841b5b92c [0x00007f6841b5b700+0x 22c] J 4162 C2 com.hazelcast.spi.impl.operationservice.Operation.writeData(Lcom/hazelcast/nio/ObjectDataOutput;)V (230 bytes) @ 0x00007f6841b4d770 [0x00007f6841b4d200+0x570] J 4235 C2 com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.write(Lcom/hazelcast/nio/ObjectDataOutput;Ljava/lang/Object;)V (10 bytes) @ 0x00007f6841b604 74 [0x00007f6841b60320+0x154] J 5407 C2 com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(Ljava/lang/Object;IZLcom/hazelcast/partition/PartitioningStrategy;Ljava/nio/ByteOrder;)[B (1 39 bytes) @ 0x00007f68416232a4 [0x00007f6841622ea0+0x404] J 5142 C2 com.hazelcast.spi.impl.operationservice.impl.OutboundOperationHandler.send(Lcom/hazelcast/spi/impl/operationservice/Operation;Lcom/hazelcast/cluster/Address;)Z (83 bytes) @ 0x00007f6841758ea0 [0x00007f6841758b80+0x320] J 5149 C2 com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.handleResponse(Lcom/hazelcast/spi/impl/operationservice/Operation;)V (52 bytes) @ 0x00007f6841b280a4 [0x00 007f6841b27100+0xfa4] J 5365 C2 com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(Lcom/hazelcast/spi/impl/operationservice/Operation;)V (139 bytes) @ 0x00007f6841185614 [0x00007f68411 85560+0xb4] J 5054 C2 com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(Ljava/lang/Object;)V (206 bytes) @ 0x00007f6841e53d44 [0x00007f6841e539e0+0x364] J 5070% C2 com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.run()V (101 bytes) @ 0x00007f6841169dc8 [0x00007f6841169d20+0xa8] v ~StubRoutines::call_stub ``` Danny Conlon 1:45 PM http://54.147.27.51/~jenkins/workspace/split-quorum/4.0-SNAPSHOT/2019_10_23-07_52_53/quorum/output/HZ/HzMember3HZBB/hs_err_pid2321.log core dump in this run http://jenkins.hazelcast.com/view/split/job/split-quorum/35/console ./output/HZ/HzMember3HZBB/hs_err_pid2302.log ./output/HZ/HzMember1HZBB/hs_err_pid2301.log ./output/HZ/HzMember2HZBB/hs_err_pid2318.log ./output/HZ/HzMember3HZBB/core.2302 ./output/HZ/HzMember1HZBB/core.2301 ./output/HZ/HzMember2HZBB/core.2318 I also a different test which also crashes the JVM http://jenkins.hazelcast.com/view/wan/job/wan-sync-oome/28/console ``` # Problematic frame: # J 4782 C2 com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization /Data;)V (128 bytes) @ 0x00007f6841db8a69 [0x00007f6841db8960+0x109] ``` probably has same root cause http://54.147.27.51/~jenkins/workspace/wan-sync-oome/4.0-SNAPSHOT/2019_10_23-10_10_48/sync-oome/ output/B/HzMember2B/core.2476
defect
sigsegv com hazelcast map impl record recordreaderwriter writerecord lcom hazelcast nio objectdataoutput lcom hazelcast map impl record record lcom hazelcast nio serialization different test are crashing the member jvm a fatal error has been detected by the java runtime environment sigsegv at pc pid tid jre version java tm se runtime environment build java vm java hotspot tm bit server vm mixed mode linux compressed oops problematic frame j com hazelcast map impl record recordreaderwriter writerecord lcom hazelcast nio objectdataoutput lcom hazelcast map impl record record lcom hazelcast nio serialization data v bytes core dump written default location home user hz root core or core if you would like to submit a bug report please visit native frames j compiled java code j interpreted vv vm code c native code j com hazelcast map impl record recordreaderwriter writerecord lcom hazelcast nio objectdataoutput lcom hazelcast map impl record record lcom hazelcast nio serialization d ata v bytes j com hazelcast map impl operation putallbackupoperation writeinternal lcom hazelcast nio objectdataoutput v bytes j com hazelcast spi impl operationservice operation writedata lcom hazelcast nio objectdataoutput v bytes j com hazelcast internal serialization impl enterprisedataserializableserializer write lcom hazelcast nio objectdataoutput ljava lang object v bytes j com hazelcast spi impl operationservice impl operations backup writeinternal lcom hazelcast nio objectdataoutput v bytes j com hazelcast spi impl operationservice operation writedata lcom hazelcast nio objectdataoutput v bytes j com hazelcast internal serialization impl enterprisedataserializableserializer write lcom hazelcast nio objectdataoutput ljava lang object v bytes j com hazelcast internal serialization impl abstractserializationservice tobytes ljava lang object izlcom hazelcast partition partitioningstrategy ljava nio byteorder b bytes j com hazelcast spi impl operationservice impl outboundoperationhandler send lcom hazelcast spi impl operationservice operation lcom hazelcast cluster address z bytes j com hazelcast spi impl operationservice impl operationrunnerimpl handleresponse lcom hazelcast spi impl operationservice operation v bytes j com hazelcast spi impl operationservice impl operationrunnerimpl call lcom hazelcast spi impl operationservice operation v bytes j com hazelcast spi impl operationexecutor impl operationthread process ljava lang object v bytes j com hazelcast spi impl operationexecutor impl operationthread run v bytes v stubroutines call stub danny conlon pm core dump in this run output hz hs err log output hz hs err log output hz hs err log output hz core output hz core output hz core i also a different test which also crashes the jvm problematic frame j com hazelcast map impl record recordreaderwriter writerecord lcom hazelcast nio objectdataoutput lcom hazelcast map impl record record lcom hazelcast nio serialization data v bytes probably has same root cause output b core
1
45,492
12,827,141,583
IssuesEvent
2020-07-06 17:53:45
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
[SCREENREADER]: Alert Boxes - Role "alert" MUST announce itself to assistive tech on page load, and on change
508-defect-2 508/Accessibility design system
# [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2) <!-- Enter an issue title using the format [ERROR TYPE]: Brief description of the problem --- [SCREENREADER]: Edit buttons need aria-label for context [KEYBOARD]: Add another user link will not receive keyboard focus [AXE-CORE]: Heading levels should increase by one [COGNITION]: Error messages should be more specific [COLOR]: Blue button on blue background does not have sufficient contrast ratio --- --> <!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. --> **Feedback framework** - **❗️ Must** for if the feedback must be applied - **⚠️ Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Description <!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. --> Our [Alertbox error variant](https://design.va.gov/components/alertboxes#error-alert) of the Alertbox component contains a `role="alert"` for announcing important content to screen readers immediately. In testing, I discovered that several screen reader + browser combinations do not announce these alerts when they are rendered on first page load: * Win10 + Firefox + NVDA * Win10 + IE11 + JAWS * iOS + VoiceOver + Mobile Safari We need to come up with a strategy to inject the `role="alert"` after the component is rendered or another way to trigger a change in the component so this type of alert is read out regardless of when it is added to the document object model (DOM). ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** _Trevor_ ## Acceptance Criteria <!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. --> - [ ] Error alerts are read out on first page render - [ ] Error alerts are read out when they are dynamically added to a page ## Environment * https://design.va.gov/components/alertboxes#error-alert ## WCAG or Vendor Guidance (optional) * [Name, Role, Value: Understanding SC 4.1.2](https://www.w3.org/TR/UNDERSTANDING-WCAG20/ensure-compat-rsv.html) ## Screenshots or Trace Logs <!-- Drop any screenshots or error logs that might be useful for debugging --> ![Screen Shot 2020-06-19 at 12 21 55 PM](https://user-images.githubusercontent.com/934879/85162917-b34f2a80-b227-11ea-8bac-f86429c10e71.png)
1.0
[SCREENREADER]: Alert Boxes - Role "alert" MUST announce itself to assistive tech on page load, and on change - # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2) <!-- Enter an issue title using the format [ERROR TYPE]: Brief description of the problem --- [SCREENREADER]: Edit buttons need aria-label for context [KEYBOARD]: Add another user link will not receive keyboard focus [AXE-CORE]: Heading levels should increase by one [COGNITION]: Error messages should be more specific [COLOR]: Blue button on blue background does not have sufficient contrast ratio --- --> <!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. --> **Feedback framework** - **❗️ Must** for if the feedback must be applied - **⚠️ Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Description <!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. --> Our [Alertbox error variant](https://design.va.gov/components/alertboxes#error-alert) of the Alertbox component contains a `role="alert"` for announcing important content to screen readers immediately. In testing, I discovered that several screen reader + browser combinations do not announce these alerts when they are rendered on first page load: * Win10 + Firefox + NVDA * Win10 + IE11 + JAWS * iOS + VoiceOver + Mobile Safari We need to come up with a strategy to inject the `role="alert"` after the component is rendered or another way to trigger a change in the component so this type of alert is read out regardless of when it is added to the document object model (DOM). ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** _Trevor_ ## Acceptance Criteria <!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. --> - [ ] Error alerts are read out on first page render - [ ] Error alerts are read out when they are dynamically added to a page ## Environment * https://design.va.gov/components/alertboxes#error-alert ## WCAG or Vendor Guidance (optional) * [Name, Role, Value: Understanding SC 4.1.2](https://www.w3.org/TR/UNDERSTANDING-WCAG20/ensure-compat-rsv.html) ## Screenshots or Trace Logs <!-- Drop any screenshots or error logs that might be useful for debugging --> ![Screen Shot 2020-06-19 at 12 21 55 PM](https://user-images.githubusercontent.com/934879/85162917-b34f2a80-b227-11ea-8bac-f86429c10e71.png)
defect
alert boxes role alert must announce itself to assistive tech on page load and on change enter an issue title using the format brief description of the problem edit buttons need aria label for context add another user link will not receive keyboard focus heading levels should increase by one error messages should be more specific blue button on blue background does not have sufficient contrast ratio feedback framework ❗️ must for if the feedback must be applied ⚠️ should if the feedback is best practice ✔️ consider for suggestions enhancements description our of the alertbox component contains a role alert for announcing important content to screen readers immediately in testing i discovered that several screen reader browser combinations do not announce these alerts when they are rendered on first page load firefox nvda jaws ios voiceover mobile safari we need to come up with a strategy to inject the role alert after the component is rendered or another way to trigger a change in the component so this type of alert is read out regardless of when it is added to the document object model dom point of contact if this issue is being opened by a vfs team member please add a point of contact usually this is the same person who enters the issue ticket vfs point of contact trevor acceptance criteria error alerts are read out on first page render error alerts are read out when they are dynamically added to a page environment wcag or vendor guidance optional screenshots or trace logs
1
40,714
10,140,875,954
IssuesEvent
2019-08-03 08:30:55
STEllAR-GROUP/hpx
https://api.github.com/repos/STEllAR-GROUP/hpx
closed
HPX MPI use incorrect
category: parcel transport tag: wontfix type: defect
In the CMakeLists for the MPI parcelport, variables such as MPI_CXX_LIBRARIES are used. HPX does not use the C++ interface to MPI and therefore linking to MPI_CXX_LIBRARIES is incorrect. It is sufficient to use the MPI_C_LIBRARIES and MPI_C_INCLUDE_PATH etc etc
1.0
HPX MPI use incorrect - In the CMakeLists for the MPI parcelport, variables such as MPI_CXX_LIBRARIES are used. HPX does not use the C++ interface to MPI and therefore linking to MPI_CXX_LIBRARIES is incorrect. It is sufficient to use the MPI_C_LIBRARIES and MPI_C_INCLUDE_PATH etc etc
defect
hpx mpi use incorrect in the cmakelists for the mpi parcelport variables such as mpi cxx libraries are used hpx does not use the c interface to mpi and therefore linking to mpi cxx libraries is incorrect it is sufficient to use the mpi c libraries and mpi c include path etc etc
1
97,165
12,217,474,400
IssuesEvent
2020-05-01 17:16:35
factn/resilience-app
https://api.github.com/repos/factn/resilience-app
opened
Organizer can switch volunteers from active to inactive
Design Needed
**Description** As an organizer, I can switch volunteers from active to inactive status and back so that I can control who is eligible to volunteer in my organization (for health and safety reasons). [Airtable](https://airtable.com/tblRjd2McKgRW9MKj/viwlRRNI0PZIxrGkc/recMPlGUt4v47ORpO?blocks=hide)
1.0
Organizer can switch volunteers from active to inactive - **Description** As an organizer, I can switch volunteers from active to inactive status and back so that I can control who is eligible to volunteer in my organization (for health and safety reasons). [Airtable](https://airtable.com/tblRjd2McKgRW9MKj/viwlRRNI0PZIxrGkc/recMPlGUt4v47ORpO?blocks=hide)
non_defect
organizer can switch volunteers from active to inactive description as an organizer i can switch volunteers from active to inactive status and back so that i can control who is eligible to volunteer in my organization for health and safety reasons
0
751,202
26,232,935,256
IssuesEvent
2023-01-05 03:00:59
AleoHQ/leo
https://api.github.com/repos/AleoHQ/leo
closed
[Bug] compiling 18 `i32` exponent operations produces 30gb project folder
bug fuzz-bug priority-medium
## 🐛 Bug Report this seems to have basically the same cause as the OOM errors. every time you do an `i32 ** i32` operation with input it adds about 1gb to the proving key size. the following code leads to a proving key that's over 16gb and a json that's over 13gb. i could have done more iterations, but this program took 50gb of ram and anything more than that just OOM's me. #### Code snippet to reproduce ``` function main(x: i32) { for i in 0..18 { let y = x ** 2; } } ``` ## Your Environment - leo commit 7e24b38620f30dda816b1a76ae4e6486b8a838e0 - rustc version 1.55.0-nightly - Windows 10.0.19043 (Windows 10 Pro) [64-bit]
1.0
[Bug] compiling 18 `i32` exponent operations produces 30gb project folder - ## 🐛 Bug Report this seems to have basically the same cause as the OOM errors. every time you do an `i32 ** i32` operation with input it adds about 1gb to the proving key size. the following code leads to a proving key that's over 16gb and a json that's over 13gb. i could have done more iterations, but this program took 50gb of ram and anything more than that just OOM's me. #### Code snippet to reproduce ``` function main(x: i32) { for i in 0..18 { let y = x ** 2; } } ``` ## Your Environment - leo commit 7e24b38620f30dda816b1a76ae4e6486b8a838e0 - rustc version 1.55.0-nightly - Windows 10.0.19043 (Windows 10 Pro) [64-bit]
non_defect
compiling exponent operations produces project folder 🐛 bug report this seems to have basically the same cause as the oom errors every time you do an operation with input it adds about to the proving key size the following code leads to a proving key that s over and a json that s over i could have done more iterations but this program took of ram and anything more than that just oom s me code snippet to reproduce function main x for i in let y x your environment leo commit rustc version nightly windows windows pro
0
276,513
23,996,942,497
IssuesEvent
2022-09-14 08:20:35
INL/corpus-frontend
https://api.github.com/repos/INL/corpus-frontend
closed
Improve date filter
enhancement status: finished/testing
3 fields for from and to. Still formatted yyyy-mm-dd, but with just separate input fields. Ensure the placeholders for these fields are the lowest and highest possible values.
1.0
Improve date filter - 3 fields for from and to. Still formatted yyyy-mm-dd, but with just separate input fields. Ensure the placeholders for these fields are the lowest and highest possible values.
non_defect
improve date filter fields for from and to still formatted yyyy mm dd but with just separate input fields ensure the placeholders for these fields are the lowest and highest possible values
0
9,995
2,616,018,664
IssuesEvent
2015-03-02 01:00:16
jasonhall/bwapi
https://api.github.com/repos/jasonhall/bwapi
closed
Local PC has some issues (instability)
auto-migrated Component-Logic Milestone-MajorRelease Performance Priority-Medium Type-Defect Usability
``` Will have to address these issues at some point. ``` Original issue reported on code.google.com by `AHeinerm` on 4 Jul 2012 at 3:22
1.0
Local PC has some issues (instability) - ``` Will have to address these issues at some point. ``` Original issue reported on code.google.com by `AHeinerm` on 4 Jul 2012 at 3:22
defect
local pc has some issues instability will have to address these issues at some point original issue reported on code google com by aheinerm on jul at
1
216,223
24,247,373,959
IssuesEvent
2022-09-27 11:43:14
SmartBear/zephyr-scale-junit-integration
https://api.github.com/repos/SmartBear/zephyr-scale-junit-integration
closed
CVE-2019-10202 (High) detected in jackson-databind-2.9.4.jar - autoclosed
security vulnerability
## CVE-2019-10202 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /rxml/jackson/core/jackson-databind/2.9.4/jackson-databind-2.9.4.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/SmartBear/zephyr-scale-junit-integration/commit/a0028c5cb64062650f7c73dc13cd92022e8a1fac">a0028c5cb64062650f7c73dc13cd92022e8a1fac</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A series of deserialization vulnerabilities have been discovered in Codehaus 1.9.x implemented in EAP 7. This CVE fixes CVE-2017-17485, CVE-2017-7525, CVE-2017-15095, CVE-2018-5968, CVE-2018-7489, CVE-2018-1000873, CVE-2019-12086 reported for FasterXML jackson-databind by implementing a whitelist approach that will mitigate these vulnerabilities and future ones alike. <p>Publish Date: 2019-10-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10202>CVE-2019-10202</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://access.redhat.com/errata/RHSA-2019:2938">https://access.redhat.com/errata/RHSA-2019:2938</a></p> <p>Release Date: 2019-10-01</p> <p>Fix Resolution: 2.9.9</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
True
CVE-2019-10202 (High) detected in jackson-databind-2.9.4.jar - autoclosed - ## CVE-2019-10202 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /rxml/jackson/core/jackson-databind/2.9.4/jackson-databind-2.9.4.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/SmartBear/zephyr-scale-junit-integration/commit/a0028c5cb64062650f7c73dc13cd92022e8a1fac">a0028c5cb64062650f7c73dc13cd92022e8a1fac</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A series of deserialization vulnerabilities have been discovered in Codehaus 1.9.x implemented in EAP 7. This CVE fixes CVE-2017-17485, CVE-2017-7525, CVE-2017-15095, CVE-2018-5968, CVE-2018-7489, CVE-2018-1000873, CVE-2019-12086 reported for FasterXML jackson-databind by implementing a whitelist approach that will mitigate these vulnerabilities and future ones alike. <p>Publish Date: 2019-10-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10202>CVE-2019-10202</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://access.redhat.com/errata/RHSA-2019:2938">https://access.redhat.com/errata/RHSA-2019:2938</a></p> <p>Release Date: 2019-10-01</p> <p>Fix Resolution: 2.9.9</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
non_defect
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library rxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details a series of deserialization vulnerabilities have been discovered in codehaus x implemented in eap this cve fixes cve cve cve cve cve cve cve reported for fasterxml jackson databind by implementing a whitelist approach that will mitigate these vulnerabilities and future ones alike publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution check this box to open an automated fix pr
0
122,917
17,771,630,761
IssuesEvent
2021-08-30 14:17:03
Tim-sandbox/barista
https://api.github.com/repos/Tim-sandbox/barista
opened
CVE-2021-23364 (Medium) detected in browserslist-4.14.2.tgz, browserslist-4.14.6.tgz
security vulnerability
## CVE-2021-23364 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>browserslist-4.14.2.tgz</b>, <b>browserslist-4.14.6.tgz</b></p></summary> <p> <details><summary><b>browserslist-4.14.2.tgz</b></p></summary> <p>Share target browsers between different front-end tools, like Autoprefixer, Stylelint and babel-env-preset</p> <p>Library home page: <a href="https://registry.npmjs.org/browserslist/-/browserslist-4.14.2.tgz">https://registry.npmjs.org/browserslist/-/browserslist-4.14.2.tgz</a></p> <p>Path to dependency file: barista/barista-docs/package.json</p> <p>Path to vulnerable library: barista/barista-docs/node_modules/react-dev-utils/node_modules/browserslist/package.json</p> <p> Dependency Hierarchy: - core-2.0.0-beta.5.tgz (Root Library) - react-dev-utils-11.0.4.tgz - :x: **browserslist-4.14.2.tgz** (Vulnerable Library) </details> <details><summary><b>browserslist-4.14.6.tgz</b></p></summary> <p>Share target browsers between different front-end tools, like Autoprefixer, Stylelint and babel-env-preset</p> <p>Library home page: <a href="https://registry.npmjs.org/browserslist/-/browserslist-4.14.6.tgz">https://registry.npmjs.org/browserslist/-/browserslist-4.14.6.tgz</a></p> <p>Path to dependency file: barista/barista-web/package.json</p> <p>Path to vulnerable library: barista/barista-web/node_modules/browserslist/package.json</p> <p> Dependency Hierarchy: - build-angular-0.1102.13.tgz (Root Library) - :x: **browserslist-4.14.6.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Tim-sandbox/barista/commit/2b8e77b2ff0d688bfd2ffb44061287e82fa71967">2b8e77b2ff0d688bfd2ffb44061287e82fa71967</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package browserslist from 4.0.0 and before 4.16.5 are vulnerable to Regular Expression Denial of Service (ReDoS) during parsing of queries. <p>Publish Date: 2021-04-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23364>CVE-2021-23364</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23364">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23364</a></p> <p>Release Date: 2021-04-28</p> <p>Fix Resolution: browserslist - 4.16.5</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"browserslist","packageVersion":"4.14.2","packageFilePaths":["/barista-docs/package.json"],"isTransitiveDependency":true,"dependencyTree":"@docusaurus/core:2.0.0-beta.5;react-dev-utils:11.0.4;browserslist:4.14.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"browserslist - 4.16.5"},{"packageType":"javascript/Node.js","packageName":"browserslist","packageVersion":"4.14.6","packageFilePaths":["/barista-web/package.json"],"isTransitiveDependency":true,"dependencyTree":"@angular-devkit/build-angular:0.1102.13;browserslist:4.14.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"browserslist - 4.16.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23364","vulnerabilityDetails":"The package browserslist from 4.0.0 and before 4.16.5 are vulnerable to Regular Expression Denial of Service (ReDoS) during parsing of queries.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23364","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-23364 (Medium) detected in browserslist-4.14.2.tgz, browserslist-4.14.6.tgz - ## CVE-2021-23364 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>browserslist-4.14.2.tgz</b>, <b>browserslist-4.14.6.tgz</b></p></summary> <p> <details><summary><b>browserslist-4.14.2.tgz</b></p></summary> <p>Share target browsers between different front-end tools, like Autoprefixer, Stylelint and babel-env-preset</p> <p>Library home page: <a href="https://registry.npmjs.org/browserslist/-/browserslist-4.14.2.tgz">https://registry.npmjs.org/browserslist/-/browserslist-4.14.2.tgz</a></p> <p>Path to dependency file: barista/barista-docs/package.json</p> <p>Path to vulnerable library: barista/barista-docs/node_modules/react-dev-utils/node_modules/browserslist/package.json</p> <p> Dependency Hierarchy: - core-2.0.0-beta.5.tgz (Root Library) - react-dev-utils-11.0.4.tgz - :x: **browserslist-4.14.2.tgz** (Vulnerable Library) </details> <details><summary><b>browserslist-4.14.6.tgz</b></p></summary> <p>Share target browsers between different front-end tools, like Autoprefixer, Stylelint and babel-env-preset</p> <p>Library home page: <a href="https://registry.npmjs.org/browserslist/-/browserslist-4.14.6.tgz">https://registry.npmjs.org/browserslist/-/browserslist-4.14.6.tgz</a></p> <p>Path to dependency file: barista/barista-web/package.json</p> <p>Path to vulnerable library: barista/barista-web/node_modules/browserslist/package.json</p> <p> Dependency Hierarchy: - build-angular-0.1102.13.tgz (Root Library) - :x: **browserslist-4.14.6.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Tim-sandbox/barista/commit/2b8e77b2ff0d688bfd2ffb44061287e82fa71967">2b8e77b2ff0d688bfd2ffb44061287e82fa71967</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package browserslist from 4.0.0 and before 4.16.5 are vulnerable to Regular Expression Denial of Service (ReDoS) during parsing of queries. <p>Publish Date: 2021-04-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23364>CVE-2021-23364</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23364">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23364</a></p> <p>Release Date: 2021-04-28</p> <p>Fix Resolution: browserslist - 4.16.5</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"browserslist","packageVersion":"4.14.2","packageFilePaths":["/barista-docs/package.json"],"isTransitiveDependency":true,"dependencyTree":"@docusaurus/core:2.0.0-beta.5;react-dev-utils:11.0.4;browserslist:4.14.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"browserslist - 4.16.5"},{"packageType":"javascript/Node.js","packageName":"browserslist","packageVersion":"4.14.6","packageFilePaths":["/barista-web/package.json"],"isTransitiveDependency":true,"dependencyTree":"@angular-devkit/build-angular:0.1102.13;browserslist:4.14.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"browserslist - 4.16.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23364","vulnerabilityDetails":"The package browserslist from 4.0.0 and before 4.16.5 are vulnerable to Regular Expression Denial of Service (ReDoS) during parsing of queries.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23364","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_defect
cve medium detected in browserslist tgz browserslist tgz cve medium severity vulnerability vulnerable libraries browserslist tgz browserslist tgz browserslist tgz share target browsers between different front end tools like autoprefixer stylelint and babel env preset library home page a href path to dependency file barista barista docs package json path to vulnerable library barista barista docs node modules react dev utils node modules browserslist package json dependency hierarchy core beta tgz root library react dev utils tgz x browserslist tgz vulnerable library browserslist tgz share target browsers between different front end tools like autoprefixer stylelint and babel env preset library home page a href path to dependency file barista barista web package json path to vulnerable library barista barista web node modules browserslist package json dependency hierarchy build angular tgz root library x browserslist tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package browserslist from and before are vulnerable to regular expression denial of service redos during parsing of queries publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution browserslist isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree docusaurus core beta react dev utils browserslist isminimumfixversionavailable true minimumfixversion browserslist packagetype javascript node js packagename browserslist packageversion packagefilepaths istransitivedependency true dependencytree angular devkit build angular browserslist isminimumfixversionavailable true minimumfixversion browserslist basebranches vulnerabilityidentifier cve vulnerabilitydetails the package browserslist from and before are vulnerable to regular expression denial of service redos during parsing of queries vulnerabilityurl
0
1,273
2,775,119,966
IssuesEvent
2015-05-04 14:21:03
a85/POSTMan-Chrome-Extension
https://api.github.com/repos/a85/POSTMan-Chrome-Extension
closed
Resize history/collections panel
Usability
Well, that's quiet self-explanatory. I sometimes work with split windows and I can't resize the left panel containing the history/collections of the request which is kinda big with a 1366x800 resolution.
True
Resize history/collections panel - Well, that's quiet self-explanatory. I sometimes work with split windows and I can't resize the left panel containing the history/collections of the request which is kinda big with a 1366x800 resolution.
non_defect
resize history collections panel well that s quiet self explanatory i sometimes work with split windows and i can t resize the left panel containing the history collections of the request which is kinda big with a resolution
0
351,766
32,025,571,470
IssuesEvent
2023-09-22 08:35:50
onmetal/cephlet
https://api.github.com/repos/onmetal/cephlet
closed
CreateVolume Method Testing
enhancement integration-tests testing
# Summary Test CreateVolume Method ## Description - Create test cases to validate the functionality of the `CreateVolume` method. - Utilize the `go-ceph` library to verify that volumes are properly created in the Ceph cluster with different configurations. - Check for error conditions and boundary cases.
2.0
CreateVolume Method Testing - # Summary Test CreateVolume Method ## Description - Create test cases to validate the functionality of the `CreateVolume` method. - Utilize the `go-ceph` library to verify that volumes are properly created in the Ceph cluster with different configurations. - Check for error conditions and boundary cases.
non_defect
createvolume method testing summary test createvolume method description create test cases to validate the functionality of the createvolume method utilize the go ceph library to verify that volumes are properly created in the ceph cluster with different configurations check for error conditions and boundary cases
0
228,196
17,422,683,589
IssuesEvent
2021-08-04 04:48:14
fangwei123456/spikingjelly
https://api.github.com/repos/fangwei123456/spikingjelly
closed
type object 'CIFAR10DVS' has no attribute 'downloadable'
bug documentation good first issue
When I execute ``` from spikingjelly.datasets.cifar10_dvs import CIFAR10DVS from spikingjelly.datasets.dvs128_gesture import DVS128Gesture print('CIFAR10-DVS downloadable', CIFAR10DVS.downloadable()) print('resource, url, md5/n', CIFAR10DVS.resource_url_md5()) print('DVS128Gesture downloadable', DVS128Gesture.downloadable()) print('resource, url, md5/n', DVS128Gesture.resource_url_md5()) ``` throws the following error: ``` E:\SoftWare\Anaconda3\lib\site-packages\torchaudio\extension\extension.py:14: UserWarning: torchaudio C++ extension is not available. warnings.warn('torchaudio C++ extension is not available.') E:\SoftWare\Anaconda3\lib\site-packages\torchaudio\backend\utils.py:88: UserWarning: No audio backend is available. warnings.warn('No audio backend is available.') Traceback (most recent call last): File "D:/Project/rpg_e2vid/test.py", line 4, in <module> print('CIFAR10-DVS downloadable', CIFAR10DVS.downloadable()) AttributeError: type object 'CIFAR10DVS' has no attribute 'downloadable' ```
1.0
type object 'CIFAR10DVS' has no attribute 'downloadable' - When I execute ``` from spikingjelly.datasets.cifar10_dvs import CIFAR10DVS from spikingjelly.datasets.dvs128_gesture import DVS128Gesture print('CIFAR10-DVS downloadable', CIFAR10DVS.downloadable()) print('resource, url, md5/n', CIFAR10DVS.resource_url_md5()) print('DVS128Gesture downloadable', DVS128Gesture.downloadable()) print('resource, url, md5/n', DVS128Gesture.resource_url_md5()) ``` throws the following error: ``` E:\SoftWare\Anaconda3\lib\site-packages\torchaudio\extension\extension.py:14: UserWarning: torchaudio C++ extension is not available. warnings.warn('torchaudio C++ extension is not available.') E:\SoftWare\Anaconda3\lib\site-packages\torchaudio\backend\utils.py:88: UserWarning: No audio backend is available. warnings.warn('No audio backend is available.') Traceback (most recent call last): File "D:/Project/rpg_e2vid/test.py", line 4, in <module> print('CIFAR10-DVS downloadable', CIFAR10DVS.downloadable()) AttributeError: type object 'CIFAR10DVS' has no attribute 'downloadable' ```
non_defect
type object has no attribute downloadable when i execute from spikingjelly datasets dvs import from spikingjelly datasets gesture import print dvs downloadable downloadable print resource url n resource url print downloadable downloadable print resource url n resource url throws the following error e software lib site packages torchaudio extension extension py userwarning torchaudio c extension is not available warnings warn torchaudio c extension is not available e software lib site packages torchaudio backend utils py userwarning no audio backend is available warnings warn no audio backend is available traceback most recent call last file d project rpg test py line in print dvs downloadable downloadable attributeerror type object has no attribute downloadable
0
815,216
30,541,509,073
IssuesEvent
2023-07-19 21:54:09
open-source-uc/planner
https://api.github.com/repos/open-source-uc/planner
closed
[PAN-19] Manejo de roles
Medium priority Backend
El sistema debe reconocer los distintos roles del sistema (invitado, usuario, moderador y administrador) y debe haber una forma establecida para revisar el rol desde las rutas del back. <sub>From [SyncLinear.com](https://synclinear.com) | [PAN-19](https://linear.app/planner-uc/issue/PAN-19/manejo-de-roles)</sub>
1.0
[PAN-19] Manejo de roles - El sistema debe reconocer los distintos roles del sistema (invitado, usuario, moderador y administrador) y debe haber una forma establecida para revisar el rol desde las rutas del back. <sub>From [SyncLinear.com](https://synclinear.com) | [PAN-19](https://linear.app/planner-uc/issue/PAN-19/manejo-de-roles)</sub>
non_defect
manejo de roles el sistema debe reconocer los distintos roles del sistema invitado usuario moderador y administrador y debe haber una forma establecida para revisar el rol desde las rutas del back from
0
13,840
2,787,896,197
IssuesEvent
2015-05-08 09:46:03
OpenMS/OpenMS
https://api.github.com/repos/OpenMS/OpenMS
closed
TheoreticalSpectrumGenerator only supports DeltaMass but no NeutralLoss [294]
1.8 Bug-Reports defect minor OpenMS - library
Submitted by timosachsenberg on 2011-03-16 11:11:26 e.g. Modification with DeltaMass M & NeutralLoss M should only result in a modified precursor mass but the MS2 is also shifted.
1.0
TheoreticalSpectrumGenerator only supports DeltaMass but no NeutralLoss [294] - Submitted by timosachsenberg on 2011-03-16 11:11:26 e.g. Modification with DeltaMass M & NeutralLoss M should only result in a modified precursor mass but the MS2 is also shifted.
defect
theoreticalspectrumgenerator only supports deltamass but no neutralloss submitted by timosachsenberg on e g modification with deltamass m neutralloss m should only result in a modified precursor mass but the is also shifted
1
345,470
30,815,813,173
IssuesEvent
2023-08-01 13:25:00
rizinorg/rizin
https://api.github.com/repos/rizinorg/rizin
closed
Incorrect disassembling of tableswitch jvm instruction
test-required disassembly java
Rizin calculates operand alignment based on jvm->section [/librz/asm/arch/java/jvm.c](https://github.com/rizinorg/rizin/blob/d1fd135df322beb93a88eb038c1da1777c82a780/librz/asm/arch/java/jvm.c#L51), but in [librz/analysis/p/analysis_java.c](https://github.com/rizinorg/rizin/blob/d1fd135df322beb93a88eb038c1da1777c82a780/librz/analysis/p/analysis_java.c#L62C11-L62C11) section is set to first matching section that contains instruction address, which is incorrect, because sometimes there are multiple overlapping sections. ### Work environment | Questions | Answers |------------------------------------------------------|-------------------- | OS/arch/bits (mandatory) | macos 13.4.1, x86_64 | File format of the file you reverse (mandatory) | java .class | Architecture/bits of the file (mandatory) | java | `rizin -v` full output, **not truncated** (mandatory) | rizin 0.6.0 @ darwin-x86-64 ### Steps to reproduce the behavior - Run `rizin -qc 's 18677; pd 10' SbtParser.class` ### Expected behavior Rizin outputs something like ``` iload 1 istore 2 iload 2 tableswitch 0 :L1 1 :L2 default :L3 .end tableswitch ``` (This is output from jadx-gui) ### Actual behavior Rizin incorrectly disassembles tableswitch: <img width="609" alt="Снимок экрана 2023-07-26 в 16 04 00" src="https://github.com/rizinorg/rizin/assets/42062111/eb287591-8435-4a03-a2aa-a59d4e4ac4dd"> ### Additional Logs, screenshots, source code, configuration dump, ... [SbtParser.class.zip](https://github.com/rizinorg/rizin/files/12170195/SbtParser.class.zip)
1.0
Incorrect disassembling of tableswitch jvm instruction - Rizin calculates operand alignment based on jvm->section [/librz/asm/arch/java/jvm.c](https://github.com/rizinorg/rizin/blob/d1fd135df322beb93a88eb038c1da1777c82a780/librz/asm/arch/java/jvm.c#L51), but in [librz/analysis/p/analysis_java.c](https://github.com/rizinorg/rizin/blob/d1fd135df322beb93a88eb038c1da1777c82a780/librz/analysis/p/analysis_java.c#L62C11-L62C11) section is set to first matching section that contains instruction address, which is incorrect, because sometimes there are multiple overlapping sections. ### Work environment | Questions | Answers |------------------------------------------------------|-------------------- | OS/arch/bits (mandatory) | macos 13.4.1, x86_64 | File format of the file you reverse (mandatory) | java .class | Architecture/bits of the file (mandatory) | java | `rizin -v` full output, **not truncated** (mandatory) | rizin 0.6.0 @ darwin-x86-64 ### Steps to reproduce the behavior - Run `rizin -qc 's 18677; pd 10' SbtParser.class` ### Expected behavior Rizin outputs something like ``` iload 1 istore 2 iload 2 tableswitch 0 :L1 1 :L2 default :L3 .end tableswitch ``` (This is output from jadx-gui) ### Actual behavior Rizin incorrectly disassembles tableswitch: <img width="609" alt="Снимок экрана 2023-07-26 в 16 04 00" src="https://github.com/rizinorg/rizin/assets/42062111/eb287591-8435-4a03-a2aa-a59d4e4ac4dd"> ### Additional Logs, screenshots, source code, configuration dump, ... [SbtParser.class.zip](https://github.com/rizinorg/rizin/files/12170195/SbtParser.class.zip)
non_defect
incorrect disassembling of tableswitch jvm instruction rizin calculates operand alignment based on jvm section but in section is set to first matching section that contains instruction address which is incorrect because sometimes there are multiple overlapping sections work environment questions answers os arch bits mandatory macos file format of the file you reverse mandatory java class architecture bits of the file mandatory java rizin v full output not truncated mandatory rizin darwin steps to reproduce the behavior run rizin qc s pd sbtparser class expected behavior rizin outputs something like iload istore iload tableswitch default end tableswitch this is output from jadx gui actual behavior rizin incorrectly disassembles tableswitch img width alt снимок экрана в src additional logs screenshots source code configuration dump
0
106,426
11,488,184,530
IssuesEvent
2020-02-11 13:28:05
spring-projects/spring-boot
https://api.github.com/repos/spring-projects/spring-boot
closed
Document how to register a blocking health contributor with the reactive registry
type: documentation
In order to add my custom health indicator I injected `HealthContributorRegistry` and then registered my contributor. I did this because I saw that `HealthContributorRegistry` was autoconfigured (required for `HealthEndpoint`) Later on I've found that my custom contributor was not present - I realized that `ReactiveHealthContributorRegistry` was used instead. I wonder if it would be feasible to update `ReactiveHealthContributorRegistry` when `HealthContributorRegistry` is being updated. Current setup seems to be unintuitive - there are no clear guidelines what should be configured and how reactive parts interact with traditional ones.
1.0
Document how to register a blocking health contributor with the reactive registry - In order to add my custom health indicator I injected `HealthContributorRegistry` and then registered my contributor. I did this because I saw that `HealthContributorRegistry` was autoconfigured (required for `HealthEndpoint`) Later on I've found that my custom contributor was not present - I realized that `ReactiveHealthContributorRegistry` was used instead. I wonder if it would be feasible to update `ReactiveHealthContributorRegistry` when `HealthContributorRegistry` is being updated. Current setup seems to be unintuitive - there are no clear guidelines what should be configured and how reactive parts interact with traditional ones.
non_defect
document how to register a blocking health contributor with the reactive registry in order to add my custom health indicator i injected healthcontributorregistry and then registered my contributor i did this because i saw that healthcontributorregistry was autoconfigured required for healthendpoint later on i ve found that my custom contributor was not present i realized that reactivehealthcontributorregistry was used instead i wonder if it would be feasible to update reactivehealthcontributorregistry when healthcontributorregistry is being updated current setup seems to be unintuitive there are no clear guidelines what should be configured and how reactive parts interact with traditional ones
0
57,301
15,729,894,962
IssuesEvent
2021-03-29 15:19:42
Kytech/xbox360wirelesschatpad
https://api.github.com/repos/Kytech/xbox360wirelesschatpad
closed
ERROR: Wireless Receiver Not Found.
Type-Defect auto-migrated
``` win 7 x64 installed filter, vjoy no problem, run the chatpad exe and get the following output, with all boxes greyed out up top [04/03/2014 1:00:16 AM] - ERROR: Wireless Receiver Not Found. [04/03/2014 1:00:16 AM] - WARNING: Failed to Acquire vJoy Gamepad Number 4. [04/03/2014 1:00:16 AM] - WARNING: Failed to Acquire vJoy Gamepad Number 3. [04/03/2014 1:00:16 AM] - WARNING: Failed to Acquire vJoy Gamepad Number 2. any ideas? thanks for this, it's an exciting idea to solve the problem of "couch internet"! ``` Original issue reported on code.google.com by `aurishal...@gmail.com` on 4 Mar 2014 at 5:06
1.0
ERROR: Wireless Receiver Not Found. - ``` win 7 x64 installed filter, vjoy no problem, run the chatpad exe and get the following output, with all boxes greyed out up top [04/03/2014 1:00:16 AM] - ERROR: Wireless Receiver Not Found. [04/03/2014 1:00:16 AM] - WARNING: Failed to Acquire vJoy Gamepad Number 4. [04/03/2014 1:00:16 AM] - WARNING: Failed to Acquire vJoy Gamepad Number 3. [04/03/2014 1:00:16 AM] - WARNING: Failed to Acquire vJoy Gamepad Number 2. any ideas? thanks for this, it's an exciting idea to solve the problem of "couch internet"! ``` Original issue reported on code.google.com by `aurishal...@gmail.com` on 4 Mar 2014 at 5:06
defect
error wireless receiver not found win installed filter vjoy no problem run the chatpad exe and get the following output with all boxes greyed out up top error wireless receiver not found warning failed to acquire vjoy gamepad number warning failed to acquire vjoy gamepad number warning failed to acquire vjoy gamepad number any ideas thanks for this it s an exciting idea to solve the problem of couch internet original issue reported on code google com by aurishal gmail com on mar at
1
220,261
17,173,510,902
IssuesEvent
2021-07-15 08:33:19
NetApp/harvest
https://api.github.com/repos/NetApp/harvest
closed
NetApp Detail: Volume Dashboard / space used column does not reflect current fill grade as graphic
bug status/testme
Volume dashboard shows a incomplete grade of filling in the table of volumes **To Reproduce** Open the NetApp Detail: Volume **Expected behavior** Last column in table should reflect the current fill-state of volume in conjunction to value **Actual behavior** Graph shows always the same fill-state ![image](https://user-images.githubusercontent.com/33655308/125610269-b482e3f6-f915-43a8-8912-ae5359b8be6f.png) **Possible solution, workaround, fix** Set an override property for the "space used" column "Standard options > Max" and set the value to 100 ![image](https://user-images.githubusercontent.com/33655308/125610384-3bded168-8e64-4cf9-8f1d-7590f7396f82.png)
1.0
NetApp Detail: Volume Dashboard / space used column does not reflect current fill grade as graphic - Volume dashboard shows a incomplete grade of filling in the table of volumes **To Reproduce** Open the NetApp Detail: Volume **Expected behavior** Last column in table should reflect the current fill-state of volume in conjunction to value **Actual behavior** Graph shows always the same fill-state ![image](https://user-images.githubusercontent.com/33655308/125610269-b482e3f6-f915-43a8-8912-ae5359b8be6f.png) **Possible solution, workaround, fix** Set an override property for the "space used" column "Standard options > Max" and set the value to 100 ![image](https://user-images.githubusercontent.com/33655308/125610384-3bded168-8e64-4cf9-8f1d-7590f7396f82.png)
non_defect
netapp detail volume dashboard space used column does not reflect current fill grade as graphic volume dashboard shows a incomplete grade of filling in the table of volumes to reproduce open the netapp detail volume expected behavior last column in table should reflect the current fill state of volume in conjunction to value actual behavior graph shows always the same fill state possible solution workaround fix set an override property for the space used column standard options max and set the value to
0
74,984
25,463,950,621
IssuesEvent
2022-11-25 00:35:26
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
shgo is not correctly passing jac to minimizer
defect scipy.optimize Documentation query
I am getting the "_minimize_slsqp() got multiple values for argument 'jac' " error for my toy example below. Is this a bug in shgo or am I doing something wrong? Please advise if there is a workaround. ``` from scipy.optimize import shgo def objective(x): return 3*x[0]*x[0] + 2*x[0] + 5 def objective_jac(x): return 6*x[0] + 2 if __name__ == "__main__": bounds = [(-100, 100)] result = shgo(objective, bounds, options={'jac':objective_jac}) print(result) ```
1.0
shgo is not correctly passing jac to minimizer - I am getting the "_minimize_slsqp() got multiple values for argument 'jac' " error for my toy example below. Is this a bug in shgo or am I doing something wrong? Please advise if there is a workaround. ``` from scipy.optimize import shgo def objective(x): return 3*x[0]*x[0] + 2*x[0] + 5 def objective_jac(x): return 6*x[0] + 2 if __name__ == "__main__": bounds = [(-100, 100)] result = shgo(objective, bounds, options={'jac':objective_jac}) print(result) ```
defect
shgo is not correctly passing jac to minimizer i am getting the minimize slsqp got multiple values for argument jac error for my toy example below is this a bug in shgo or am i doing something wrong please advise if there is a workaround from scipy optimize import shgo def objective x return x x x def objective jac x return x if name main bounds result shgo objective bounds options jac objective jac print result
1
329,218
28,208,727,469
IssuesEvent
2023-04-05 00:58:15
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
opened
Adopt split repo: python-eventarc-publishing
migration:samples:generated migration:workaround:none migration:library:gapic_auto migration:testing:unit migration:issues:none
Migrate the split-repo https://github.com/googleapis/python-eventarc-publishing to https://github.com/googleapis/google-cloud-python. To do the actual migration, we need to ensure we can clear any of the following tags in this issue that describe the state of the source repo: `testing:system`, `samples:manual`,`workaround:owlbot`.
1.0
Adopt split repo: python-eventarc-publishing - Migrate the split-repo https://github.com/googleapis/python-eventarc-publishing to https://github.com/googleapis/google-cloud-python. To do the actual migration, we need to ensure we can clear any of the following tags in this issue that describe the state of the source repo: `testing:system`, `samples:manual`,`workaround:owlbot`.
non_defect
adopt split repo python eventarc publishing migrate the split repo to to do the actual migration we need to ensure we can clear any of the following tags in this issue that describe the state of the source repo testing system samples manual workaround owlbot
0
12,756
2,715,987,873
IssuesEvent
2015-04-10 16:18:01
codenameone/CodenameOne
https://api.github.com/repos/codenameone/CodenameOne
closed
BlackBerry 5 BrowserComponent disappears on transition
Priority-High Type-Defect
Original [issue 491](https://code.google.com/p/codenameone/issues/detail?id=491) created by codenameone on 2013-01-17T20:58:24.000Z: The BB 5 BrowserComponent disappears when transitioning into a form. E.g. the following will display &quot;Hello World&quot; briefly, but then disappear. Form hi = new Form(&quot;Hi World&quot;); BrowserComponent b = new BrowserComponent(); b.setPage('&lt;html&gt;&lt;body&gt;Hello World&lt;/body&gt;&lt;/html&gt;', 'file:///'); hi.setLayout(new BorderPane()); hi.addComponent(BorderLayout.CENTER, b); hi.show(); However, if I remove and re-add the browser component inside the onShowCompleted() event of the form, it shows up properly. I'm guessing this has something to do with the PeerComponent code in the BlackBerry implementation. I'm looking through the createNativePeer() method and looking at the implementation of peer components, but, being new to BB, it will be a bit of a hill to climb to nail this down. I'm keen to hunt this one down, so any pointers or suggestions on places to look, much appreciated.
1.0
BlackBerry 5 BrowserComponent disappears on transition - Original [issue 491](https://code.google.com/p/codenameone/issues/detail?id=491) created by codenameone on 2013-01-17T20:58:24.000Z: The BB 5 BrowserComponent disappears when transitioning into a form. E.g. the following will display &quot;Hello World&quot; briefly, but then disappear. Form hi = new Form(&quot;Hi World&quot;); BrowserComponent b = new BrowserComponent(); b.setPage('&lt;html&gt;&lt;body&gt;Hello World&lt;/body&gt;&lt;/html&gt;', 'file:///'); hi.setLayout(new BorderPane()); hi.addComponent(BorderLayout.CENTER, b); hi.show(); However, if I remove and re-add the browser component inside the onShowCompleted() event of the form, it shows up properly. I'm guessing this has something to do with the PeerComponent code in the BlackBerry implementation. I'm looking through the createNativePeer() method and looking at the implementation of peer components, but, being new to BB, it will be a bit of a hill to climb to nail this down. I'm keen to hunt this one down, so any pointers or suggestions on places to look, much appreciated.
defect
blackberry browsercomponent disappears on transition original created by codenameone on the bb browsercomponent disappears when transitioning into a form e g the following will display quot hello world quot briefly but then disappear form hi new form quot hi world quot browsercomponent b new browsercomponent b setpage lt html gt lt body gt hello world lt body gt lt html gt file hi setlayout new borderpane hi addcomponent borderlayout center b hi show however if i remove and re add the browser component inside the onshowcompleted event of the form it shows up properly i m guessing this has something to do with the peercomponent code in the blackberry implementation i m looking through the createnativepeer method and looking at the implementation of peer components but being new to bb it will be a bit of a hill to climb to nail this down i m keen to hunt this one down so any pointers or suggestions on places to look much appreciated
1
185,877
14,384,142,981
IssuesEvent
2020-12-02 10:04:02
WoWManiaUK/Redemption
https://api.github.com/repos/WoWManiaUK/Redemption
closed
[Raid] Felmyst's Demonic Vapors
Fixed on PTR - Tester Confirmed
**Links:** https://wowwiki.fandom.com/wiki/Felmyst https://youtu.be/pcv5LlawB1A?t=862 https://www.wow-mania.com/armory/?search=demonic+vapor#spells https://www.wow-mania.com/armory/?npc=25268 **What is Happening:** The fight function mostly fine, but 1 major issue persists. When air phase begins, Felmyst will use her first aerial-ability roughly 5 seconds after leaving melee range; the [Demonic Vapor](https://www.wow-mania.com/armory/?search=demonic+vapor#spells). However, it is currently targetting every single raid member with a nature damage beam that leaves behind a nature damage trail. The stacking trails of damage quickly overwhelm the entire room and the more players are present, the more chaotic it is. **What Should happen:** > She will target a random player for a couple seconds, then cast it under them. The spell follows that player for about ten seconds, leaving the DoT trail and summoning skeletons soon after. 1. The spell is supposed to target only a single raid member and chase them for a short duration (duration and damage are correct for 1 person), before choosing a new target and repeating the process (Behaves somewhat like the spikes in the Anub'Arak encounter). After 2 Demonic Vapors, she starts her fly-bys that spread the Fog of Corruption (fixed previously). Behavior is clearly seen from ~14:20 onward in [this video](https://youtu.be/pcv5LlawB1A?t=862). 2. The trail is supposed to spawn ~10 [Unyielding Dead](https://www.wow-mania.com/armory/?npc=25268) as it is forming (14:37 in the video), and additional skeletons if players touch the trail afterwards (similar to Yogg-Saron's clouds). There used to be a fairly bad issue with the number of skeletons spawned by this ability (most likely due to the spell targeting all raid members instead of only 1). *Less important than 1
1.0
[Raid] Felmyst's Demonic Vapors - **Links:** https://wowwiki.fandom.com/wiki/Felmyst https://youtu.be/pcv5LlawB1A?t=862 https://www.wow-mania.com/armory/?search=demonic+vapor#spells https://www.wow-mania.com/armory/?npc=25268 **What is Happening:** The fight function mostly fine, but 1 major issue persists. When air phase begins, Felmyst will use her first aerial-ability roughly 5 seconds after leaving melee range; the [Demonic Vapor](https://www.wow-mania.com/armory/?search=demonic+vapor#spells). However, it is currently targetting every single raid member with a nature damage beam that leaves behind a nature damage trail. The stacking trails of damage quickly overwhelm the entire room and the more players are present, the more chaotic it is. **What Should happen:** > She will target a random player for a couple seconds, then cast it under them. The spell follows that player for about ten seconds, leaving the DoT trail and summoning skeletons soon after. 1. The spell is supposed to target only a single raid member and chase them for a short duration (duration and damage are correct for 1 person), before choosing a new target and repeating the process (Behaves somewhat like the spikes in the Anub'Arak encounter). After 2 Demonic Vapors, she starts her fly-bys that spread the Fog of Corruption (fixed previously). Behavior is clearly seen from ~14:20 onward in [this video](https://youtu.be/pcv5LlawB1A?t=862). 2. The trail is supposed to spawn ~10 [Unyielding Dead](https://www.wow-mania.com/armory/?npc=25268) as it is forming (14:37 in the video), and additional skeletons if players touch the trail afterwards (similar to Yogg-Saron's clouds). There used to be a fairly bad issue with the number of skeletons spawned by this ability (most likely due to the spell targeting all raid members instead of only 1). *Less important than 1
non_defect
felmyst s demonic vapors links what is happening the fight function mostly fine but major issue persists when air phase begins felmyst will use her first aerial ability roughly seconds after leaving melee range the however it is currently targetting every single raid member with a nature damage beam that leaves behind a nature damage trail the stacking trails of damage quickly overwhelm the entire room and the more players are present the more chaotic it is what should happen she will target a random player for a couple seconds then cast it under them the spell follows that player for about ten seconds leaving the dot trail and summoning skeletons soon after the spell is supposed to target only a single raid member and chase them for a short duration duration and damage are correct for person before choosing a new target and repeating the process behaves somewhat like the spikes in the anub arak encounter after demonic vapors she starts her fly bys that spread the fog of corruption fixed previously behavior is clearly seen from onward in the trail is supposed to spawn as it is forming in the video and additional skeletons if players touch the trail afterwards similar to yogg saron s clouds there used to be a fairly bad issue with the number of skeletons spawned by this ability most likely due to the spell targeting all raid members instead of only less important than
0
48,534
13,113,835,143
IssuesEvent
2020-08-05 06:29:51
naev/naev
https://api.github.com/repos/naev/naev
closed
When a faction is not <known>, its ships, systems, and assets are still clearly marked
Priority-High Type-Defect
When a faction lacks the `<known>` tag, the faction's presence is listed as "unknown". But the name and/or logo of the faction appears when targeting a ship or planet of that faction, or selecting a system controlled by the faction on the map.
1.0
When a faction is not <known>, its ships, systems, and assets are still clearly marked - When a faction lacks the `<known>` tag, the faction's presence is listed as "unknown". But the name and/or logo of the faction appears when targeting a ship or planet of that faction, or selecting a system controlled by the faction on the map.
defect
when a faction is not its ships systems and assets are still clearly marked when a faction lacks the tag the faction s presence is listed as unknown but the name and or logo of the faction appears when targeting a ship or planet of that faction or selecting a system controlled by the faction on the map
1
58,101
16,342,448,659
IssuesEvent
2021-05-13 00:19:41
darshan-hpc/darshan
https://api.github.com/repos/darshan-hpc/darshan
closed
incorrect reporting of bytes read and written by HMMER application
defect wrapper libraries
In GitLab by @shanedsnyder on Sep 24, 2015, 16:27 See http://hmmer.janelia.org/. Problem reported by Chris Daley. Example execution reads gigabytes of data and writes 19 MiB of data, but Darshan reports 8 bytes read and roughly 1 MiB written. Application is using fread and fwrite for I/O, and everything shows up as expected in strace as read and write operations. This indicates that there may be a bug in how Darshan records the sizes of fread and fwrite operations. Removing from 2.2.9 milestone for now, but we should still try to produce a prototype to test with hmmer.
1.0
incorrect reporting of bytes read and written by HMMER application - In GitLab by @shanedsnyder on Sep 24, 2015, 16:27 See http://hmmer.janelia.org/. Problem reported by Chris Daley. Example execution reads gigabytes of data and writes 19 MiB of data, but Darshan reports 8 bytes read and roughly 1 MiB written. Application is using fread and fwrite for I/O, and everything shows up as expected in strace as read and write operations. This indicates that there may be a bug in how Darshan records the sizes of fread and fwrite operations. Removing from 2.2.9 milestone for now, but we should still try to produce a prototype to test with hmmer.
defect
incorrect reporting of bytes read and written by hmmer application in gitlab by shanedsnyder on sep see problem reported by chris daley example execution reads gigabytes of data and writes mib of data but darshan reports bytes read and roughly mib written application is using fread and fwrite for i o and everything shows up as expected in strace as read and write operations this indicates that there may be a bug in how darshan records the sizes of fread and fwrite operations removing from milestone for now but we should still try to produce a prototype to test with hmmer
1
31,214
6,447,816,237
IssuesEvent
2017-08-14 09:13:17
PowerDNS/pdns
https://api.github.com/repos/PowerDNS/pdns
closed
IPv6 addresses improperly encoded in carbon metrics
defect dnsdist
- Program: dnsdist - Issue type: Bug report ### Short description When logging to carbon, IPv6 addresses are enclosed in square brackets ([]) within metric names. Square brackets have special meaning to many storage engines and frontends. ### Environment - Operating system: Ubuntu 16.04 - Software version: dnsdist 1.1.0 - Software source: repo ### Steps to reproduce 1. Configure dnsdist with an IPv6 interface and logging to a carbon server 2. Observe the metric names within your carbon instance 3. IPv4 addresses are encoded as `10_0_0_1:53_tcp`, IPv6 addresses are encoded as `[2001:db8::1]:53_tcp` ### Expected behaviour A recommended encoding would be `2001:db8::1_53_tcp`. It is really a toss-up on the colons becoming ambiguous or breaking with the v4 encoding.
1.0
IPv6 addresses improperly encoded in carbon metrics - - Program: dnsdist - Issue type: Bug report ### Short description When logging to carbon, IPv6 addresses are enclosed in square brackets ([]) within metric names. Square brackets have special meaning to many storage engines and frontends. ### Environment - Operating system: Ubuntu 16.04 - Software version: dnsdist 1.1.0 - Software source: repo ### Steps to reproduce 1. Configure dnsdist with an IPv6 interface and logging to a carbon server 2. Observe the metric names within your carbon instance 3. IPv4 addresses are encoded as `10_0_0_1:53_tcp`, IPv6 addresses are encoded as `[2001:db8::1]:53_tcp` ### Expected behaviour A recommended encoding would be `2001:db8::1_53_tcp`. It is really a toss-up on the colons becoming ambiguous or breaking with the v4 encoding.
defect
addresses improperly encoded in carbon metrics program dnsdist issue type bug report short description when logging to carbon addresses are enclosed in square brackets within metric names square brackets have special meaning to many storage engines and frontends environment operating system ubuntu software version dnsdist software source repo steps to reproduce configure dnsdist with an interface and logging to a carbon server observe the metric names within your carbon instance addresses are encoded as tcp addresses are encoded as tcp expected behaviour a recommended encoding would be tcp it is really a toss up on the colons becoming ambiguous or breaking with the encoding
1
79,750
28,780,723,527
IssuesEvent
2023-05-02 00:14:26
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
transactionCoroutine maybe? does not wait for transaction to commit before returning
T: Defect
### Expected behavior Wait for transaction to commit before returning (async) control back to to consumer. ### Actual behavior Control is returned back to consumer before transaction has committed. ### Steps to reproduce the problem I've been using the transactionCoroutine wrapper to manage a transaction against a postgres database connection pool with r2dbc and r2dbc pool. I have some integration tests that run (in a transaction) some commands before committing. Something along the lines of: ``` val config = DefaultConfiguration() .set(pool) .set(JDBCUtils.dialect(pool)) val dsl = DSL.using(config) val updatedCount = dsl.transactionCoroutine { config -> val transactionDsl = config.dsl() ... some stuff transactionalDsl.update(...).awaitLast() } verifyTransactionResults(dsl) ``` Intermittently, the verification of the results will fail in a way that indicates that the transaction hadn't committed. If I turn OFF connection pooling I find that the intermittency goes away and the tests pass as expected. If I delay / sleep for a few MS before checking the intermittency goes away. An alternative implementation of transactionCoroutine using `.awaitLast()` rather than `.awaitFirstOrNull()` seems to solve the problem. In a tight loop, I can reproduce this every time. However, when trying against h2 I was unable to reproduce (though could be a function of removing the network latency). With H2, the transactionCoroutine implementation finished significantly faster. <details> <summary>full repro code here</summary> dependencies: ``` implementation("io.r2dbc:r2dbc-postgresql:0.8.13.RELEASE") implementation("io.r2dbc:r2dbc-pool:1.0.0.RELEASE") implementation("org.jetbrains.kotlinx:kotlinx-coroutines-core-jvm:1.6.4") implementation("org.jetbrains.kotlinx:kotlinx-coroutines-jdk8:1.6.4") implementation("org.jooq:jooq-kotlin:3.18.0") implementation("org.jooq:jooq:3.18.0") implementation("org.jooq:jooq-kotlin-coroutines:3.18.0") implementation("org.jooq:jooq-postgres-extensions:3.18.0") testRuntimeOnly("org.junit.jupiter:junit-jupiter-engine:5.9.2") testImplementation("org.junit.jupiter:junit-jupiter-api:5.9.2") ``` test code: ``` import io.r2dbc.pool.ConnectionPool import io.r2dbc.pool.ConnectionPoolConfiguration import io.r2dbc.spi.ConnectionFactories import io.r2dbc.spi.ConnectionFactoryOptions import kotlinx.coroutines.reactive.awaitFirstOrNull import kotlinx.coroutines.reactive.awaitLast import kotlinx.coroutines.reactor.mono import kotlinx.coroutines.runBlocking import org.jooq.DSLContext import org.jooq.impl.DSL import org.jooq.impl.DSL.count import org.jooq.impl.DSL.field import org.jooq.impl.DSL.inline import org.jooq.impl.DSL.table import org.jooq.impl.DefaultConfiguration import org.jooq.impl.SQLDataType.INTEGER import org.jooq.impl.SQLDataType.VARCHAR import org.jooq.kotlin.coroutines.transactionCoroutine import org.jooq.tools.jdbc.JDBCUtils import org.junit.jupiter.api.Assertions.assertEquals import org.junit.jupiter.api.Assertions.assertNotNull import org.junit.jupiter.api.Test import java.time.Duration import java.util.UUID class ConnectionWrapperTest { private val connectionFactoryOptions = ConnectionFactoryOptions.builder() .option(ConnectionFactoryOptions.DRIVER, "postgresql") .option(ConnectionFactoryOptions.HOST, "localhost") .option(ConnectionFactoryOptions.PORT, 5432) .option(ConnectionFactoryOptions.USER, "user") .option(ConnectionFactoryOptions.DATABASE, "test") .build() private val tableName = "testing12345" private val idCol = "id" private val valueCol = "val" private suspend fun validate(wrapper: ConnectionWrapper, tableName: String, uuid: String) { val returned = wrapper.dbQuery { select(field(idCol), field(valueCol)) .from(table(tableName)) .where(field(valueCol).eq(inline(uuid))) .awaitFirstOrNull() } assertNotNull(returned) } private suspend fun updateValue(dsl: DSLContext, tableName: String, uuid: String): Int? { return dsl.update(table(tableName)) .set(field(valueCol), uuid) .awaitLast() } @Test fun `transaction using transactionCoroutine`() { val connectionWrapper = ConnectionWrapper(connectionFactoryOptions) runBlocking { setup(connectionWrapper, tableName) } repeat(1_000) { runBlocking { val uuid = UUID.randomUUID().toString() val updated = connectionWrapper.dbQuery { transactionCoroutine { config -> val transactionDsl = config.dsl() updateValue(transactionDsl, tableName, uuid) } } assertEquals(1, updated) validate(connectionWrapper, tableName, uuid) } } connectionWrapper.shutdown() } @Test fun `transaction using alternative`() = runBlocking { val connectionWrapper = ConnectionWrapper(connectionFactoryOptions) runBlocking { setup(connectionWrapper, tableName) } repeat(1_000) { val uuid = UUID.randomUUID().toString() val updated = connectionWrapper.dbQuery { transactionPublisher { config -> val transactionDsl = config.dsl() mono { updateValue(transactionDsl, tableName, uuid) } // use await last instead of await first }.awaitLast() } assertEquals(1, updated) validate(connectionWrapper, tableName, uuid) } connectionWrapper.shutdown() } private suspend fun setup(wrapper: ConnectionWrapper, tableName: String) { wrapper.dbQuery { createTableIfNotExists(tableName) .column(idCol, INTEGER) .column(valueCol, VARCHAR(255)) .awaitFirstOrNull() val field = count().`as`("COUNT") val record = select(field) .from(table(tableName)) .awaitFirstOrNull() if (record == null || record[field] == 0) { insertInto(table(tableName)) .columns(field(idCol), field(valueCol)) .values(1, "old value") .awaitFirstOrNull() } } } } class ConnectionWrapper(options: ConnectionFactoryOptions) { private val factory = ConnectionFactories.get(options) private val configuration = ConnectionPoolConfiguration.builder(factory) .maxIdleTime(Duration.ofMillis(60000)) // 1 min .maxSize(10) .initialSize(10) .minIdle(10) .build() private val pool = ConnectionPool(configuration) suspend fun <T> dbQuery(block: suspend DSLContext.() -> T): T { val config = DefaultConfiguration() .set(pool) .set(JDBCUtils.dialect(pool)) val context = DSL.using(config) return block(context) } fun shutdown() { pool.dispose() } } ``` </details> ### jOOQ Version 3.18.0 ### Database product and version PostgreSQL 14.7 (Homebrew) on aarch64-apple-darwin22.1.0 ### Java Version 17 ### OS Version Mac ### JDBC driver name and version (include name if unofficial driver) io.r2dbc:r2dbc-postgres:0.8.13.RELEASE
1.0
transactionCoroutine maybe? does not wait for transaction to commit before returning - ### Expected behavior Wait for transaction to commit before returning (async) control back to to consumer. ### Actual behavior Control is returned back to consumer before transaction has committed. ### Steps to reproduce the problem I've been using the transactionCoroutine wrapper to manage a transaction against a postgres database connection pool with r2dbc and r2dbc pool. I have some integration tests that run (in a transaction) some commands before committing. Something along the lines of: ``` val config = DefaultConfiguration() .set(pool) .set(JDBCUtils.dialect(pool)) val dsl = DSL.using(config) val updatedCount = dsl.transactionCoroutine { config -> val transactionDsl = config.dsl() ... some stuff transactionalDsl.update(...).awaitLast() } verifyTransactionResults(dsl) ``` Intermittently, the verification of the results will fail in a way that indicates that the transaction hadn't committed. If I turn OFF connection pooling I find that the intermittency goes away and the tests pass as expected. If I delay / sleep for a few MS before checking the intermittency goes away. An alternative implementation of transactionCoroutine using `.awaitLast()` rather than `.awaitFirstOrNull()` seems to solve the problem. In a tight loop, I can reproduce this every time. However, when trying against h2 I was unable to reproduce (though could be a function of removing the network latency). With H2, the transactionCoroutine implementation finished significantly faster. <details> <summary>full repro code here</summary> dependencies: ``` implementation("io.r2dbc:r2dbc-postgresql:0.8.13.RELEASE") implementation("io.r2dbc:r2dbc-pool:1.0.0.RELEASE") implementation("org.jetbrains.kotlinx:kotlinx-coroutines-core-jvm:1.6.4") implementation("org.jetbrains.kotlinx:kotlinx-coroutines-jdk8:1.6.4") implementation("org.jooq:jooq-kotlin:3.18.0") implementation("org.jooq:jooq:3.18.0") implementation("org.jooq:jooq-kotlin-coroutines:3.18.0") implementation("org.jooq:jooq-postgres-extensions:3.18.0") testRuntimeOnly("org.junit.jupiter:junit-jupiter-engine:5.9.2") testImplementation("org.junit.jupiter:junit-jupiter-api:5.9.2") ``` test code: ``` import io.r2dbc.pool.ConnectionPool import io.r2dbc.pool.ConnectionPoolConfiguration import io.r2dbc.spi.ConnectionFactories import io.r2dbc.spi.ConnectionFactoryOptions import kotlinx.coroutines.reactive.awaitFirstOrNull import kotlinx.coroutines.reactive.awaitLast import kotlinx.coroutines.reactor.mono import kotlinx.coroutines.runBlocking import org.jooq.DSLContext import org.jooq.impl.DSL import org.jooq.impl.DSL.count import org.jooq.impl.DSL.field import org.jooq.impl.DSL.inline import org.jooq.impl.DSL.table import org.jooq.impl.DefaultConfiguration import org.jooq.impl.SQLDataType.INTEGER import org.jooq.impl.SQLDataType.VARCHAR import org.jooq.kotlin.coroutines.transactionCoroutine import org.jooq.tools.jdbc.JDBCUtils import org.junit.jupiter.api.Assertions.assertEquals import org.junit.jupiter.api.Assertions.assertNotNull import org.junit.jupiter.api.Test import java.time.Duration import java.util.UUID class ConnectionWrapperTest { private val connectionFactoryOptions = ConnectionFactoryOptions.builder() .option(ConnectionFactoryOptions.DRIVER, "postgresql") .option(ConnectionFactoryOptions.HOST, "localhost") .option(ConnectionFactoryOptions.PORT, 5432) .option(ConnectionFactoryOptions.USER, "user") .option(ConnectionFactoryOptions.DATABASE, "test") .build() private val tableName = "testing12345" private val idCol = "id" private val valueCol = "val" private suspend fun validate(wrapper: ConnectionWrapper, tableName: String, uuid: String) { val returned = wrapper.dbQuery { select(field(idCol), field(valueCol)) .from(table(tableName)) .where(field(valueCol).eq(inline(uuid))) .awaitFirstOrNull() } assertNotNull(returned) } private suspend fun updateValue(dsl: DSLContext, tableName: String, uuid: String): Int? { return dsl.update(table(tableName)) .set(field(valueCol), uuid) .awaitLast() } @Test fun `transaction using transactionCoroutine`() { val connectionWrapper = ConnectionWrapper(connectionFactoryOptions) runBlocking { setup(connectionWrapper, tableName) } repeat(1_000) { runBlocking { val uuid = UUID.randomUUID().toString() val updated = connectionWrapper.dbQuery { transactionCoroutine { config -> val transactionDsl = config.dsl() updateValue(transactionDsl, tableName, uuid) } } assertEquals(1, updated) validate(connectionWrapper, tableName, uuid) } } connectionWrapper.shutdown() } @Test fun `transaction using alternative`() = runBlocking { val connectionWrapper = ConnectionWrapper(connectionFactoryOptions) runBlocking { setup(connectionWrapper, tableName) } repeat(1_000) { val uuid = UUID.randomUUID().toString() val updated = connectionWrapper.dbQuery { transactionPublisher { config -> val transactionDsl = config.dsl() mono { updateValue(transactionDsl, tableName, uuid) } // use await last instead of await first }.awaitLast() } assertEquals(1, updated) validate(connectionWrapper, tableName, uuid) } connectionWrapper.shutdown() } private suspend fun setup(wrapper: ConnectionWrapper, tableName: String) { wrapper.dbQuery { createTableIfNotExists(tableName) .column(idCol, INTEGER) .column(valueCol, VARCHAR(255)) .awaitFirstOrNull() val field = count().`as`("COUNT") val record = select(field) .from(table(tableName)) .awaitFirstOrNull() if (record == null || record[field] == 0) { insertInto(table(tableName)) .columns(field(idCol), field(valueCol)) .values(1, "old value") .awaitFirstOrNull() } } } } class ConnectionWrapper(options: ConnectionFactoryOptions) { private val factory = ConnectionFactories.get(options) private val configuration = ConnectionPoolConfiguration.builder(factory) .maxIdleTime(Duration.ofMillis(60000)) // 1 min .maxSize(10) .initialSize(10) .minIdle(10) .build() private val pool = ConnectionPool(configuration) suspend fun <T> dbQuery(block: suspend DSLContext.() -> T): T { val config = DefaultConfiguration() .set(pool) .set(JDBCUtils.dialect(pool)) val context = DSL.using(config) return block(context) } fun shutdown() { pool.dispose() } } ``` </details> ### jOOQ Version 3.18.0 ### Database product and version PostgreSQL 14.7 (Homebrew) on aarch64-apple-darwin22.1.0 ### Java Version 17 ### OS Version Mac ### JDBC driver name and version (include name if unofficial driver) io.r2dbc:r2dbc-postgres:0.8.13.RELEASE
defect
transactioncoroutine maybe does not wait for transaction to commit before returning expected behavior wait for transaction to commit before returning async control back to to consumer actual behavior control is returned back to consumer before transaction has committed steps to reproduce the problem i ve been using the transactioncoroutine wrapper to manage a transaction against a postgres database connection pool with and pool i have some integration tests that run in a transaction some commands before committing something along the lines of val config defaultconfiguration set pool set jdbcutils dialect pool val dsl dsl using config val updatedcount dsl transactioncoroutine config val transactiondsl config dsl some stuff transactionaldsl update awaitlast verifytransactionresults dsl intermittently the verification of the results will fail in a way that indicates that the transaction hadn t committed if i turn off connection pooling i find that the intermittency goes away and the tests pass as expected if i delay sleep for a few ms before checking the intermittency goes away an alternative implementation of transactioncoroutine using awaitlast rather than awaitfirstornull seems to solve the problem in a tight loop i can reproduce this every time however when trying against i was unable to reproduce though could be a function of removing the network latency with the transactioncoroutine implementation finished significantly faster full repro code here dependencies implementation io postgresql release implementation io pool release implementation org jetbrains kotlinx kotlinx coroutines core jvm implementation org jetbrains kotlinx kotlinx coroutines implementation org jooq jooq kotlin implementation org jooq jooq implementation org jooq jooq kotlin coroutines implementation org jooq jooq postgres extensions testruntimeonly org junit jupiter junit jupiter engine testimplementation org junit jupiter junit jupiter api test code import io pool connectionpool import io pool connectionpoolconfiguration import io spi connectionfactories import io spi connectionfactoryoptions import kotlinx coroutines reactive awaitfirstornull import kotlinx coroutines reactive awaitlast import kotlinx coroutines reactor mono import kotlinx coroutines runblocking import org jooq dslcontext import org jooq impl dsl import org jooq impl dsl count import org jooq impl dsl field import org jooq impl dsl inline import org jooq impl dsl table import org jooq impl defaultconfiguration import org jooq impl sqldatatype integer import org jooq impl sqldatatype varchar import org jooq kotlin coroutines transactioncoroutine import org jooq tools jdbc jdbcutils import org junit jupiter api assertions assertequals import org junit jupiter api assertions assertnotnull import org junit jupiter api test import java time duration import java util uuid class connectionwrappertest private val connectionfactoryoptions connectionfactoryoptions builder option connectionfactoryoptions driver postgresql option connectionfactoryoptions host localhost option connectionfactoryoptions port option connectionfactoryoptions user user option connectionfactoryoptions database test build private val tablename private val idcol id private val valuecol val private suspend fun validate wrapper connectionwrapper tablename string uuid string val returned wrapper dbquery select field idcol field valuecol from table tablename where field valuecol eq inline uuid awaitfirstornull assertnotnull returned private suspend fun updatevalue dsl dslcontext tablename string uuid string int return dsl update table tablename set field valuecol uuid awaitlast test fun transaction using transactioncoroutine val connectionwrapper connectionwrapper connectionfactoryoptions runblocking setup connectionwrapper tablename repeat runblocking val uuid uuid randomuuid tostring val updated connectionwrapper dbquery transactioncoroutine config val transactiondsl config dsl updatevalue transactiondsl tablename uuid assertequals updated validate connectionwrapper tablename uuid connectionwrapper shutdown test fun transaction using alternative runblocking val connectionwrapper connectionwrapper connectionfactoryoptions runblocking setup connectionwrapper tablename repeat val uuid uuid randomuuid tostring val updated connectionwrapper dbquery transactionpublisher config val transactiondsl config dsl mono updatevalue transactiondsl tablename uuid use await last instead of await first awaitlast assertequals updated validate connectionwrapper tablename uuid connectionwrapper shutdown private suspend fun setup wrapper connectionwrapper tablename string wrapper dbquery createtableifnotexists tablename column idcol integer column valuecol varchar awaitfirstornull val field count as count val record select field from table tablename awaitfirstornull if record null record insertinto table tablename columns field idcol field valuecol values old value awaitfirstornull class connectionwrapper options connectionfactoryoptions private val factory connectionfactories get options private val configuration connectionpoolconfiguration builder factory maxidletime duration ofmillis min maxsize initialsize minidle build private val pool connectionpool configuration suspend fun dbquery block suspend dslcontext t t val config defaultconfiguration set pool set jdbcutils dialect pool val context dsl using config return block context fun shutdown pool dispose jooq version database product and version postgresql homebrew on apple java version os version mac jdbc driver name and version include name if unofficial driver io postgres release
1
10,478
27,020,930,225
IssuesEvent
2023-02-11 02:00:39
facebook/react-native
https://api.github.com/repos/facebook/react-native
closed
Nested TextInput inside Text looses focus on key stroke
Stale Component: TextInput Priority: Low Platform: All Tech: React Native Core Type: Old Architecture Impact: Bug
### Description Hello, I have a nested Text setup with Text and Text Inputs like this: ``` <Text> <Text/> <View> <TextInput/> </View> </Text> ``` From the styling and text wrapping perspective it works great, but unfortunately the Text Input looses focus on every keystroke. Is this a common thing or am I missing something here? ### Version 0.64.3 ### Output of `npx react-native info` System: OS: macOS 12.3.1 CPU: (8) arm64 Apple M1 Memory: 140.03 MB / 8.00 GB Shell: 5.8 - /bin/zsh Binaries: Node: 17.8.0 - /opt/homebrew/bin/node Yarn: Not Found npm: 8.5.5 - /opt/homebrew/bin/npm Watchman: Not Found Managers: CocoaPods: 1.11.3 - /usr/local/bin/pod SDKs: iOS SDK: Platforms: DriverKit 21.4, iOS 15.4, macOS 12.3, tvOS 15.4, watchOS 8.5 Android SDK: Not Found IDEs: Android Studio: Not Found Xcode: 13.3/13E113 - /usr/bin/xcodebuild Languages: Java: Not Found npmPackages: @react-native-community/cli: Not Found react: ^17.0.1 => 17.0.2 react-native: 0.64.3 => 0.64.3 react-native-macos: Not Found npmGlobalPackages: *react-native*: Not Found ### Steps to reproduce Setup a simple React Native App with a Text Input inside a Text component ### Snack, code example, screenshot, or link to a repository _No response_
1.0
Nested TextInput inside Text looses focus on key stroke - ### Description Hello, I have a nested Text setup with Text and Text Inputs like this: ``` <Text> <Text/> <View> <TextInput/> </View> </Text> ``` From the styling and text wrapping perspective it works great, but unfortunately the Text Input looses focus on every keystroke. Is this a common thing or am I missing something here? ### Version 0.64.3 ### Output of `npx react-native info` System: OS: macOS 12.3.1 CPU: (8) arm64 Apple M1 Memory: 140.03 MB / 8.00 GB Shell: 5.8 - /bin/zsh Binaries: Node: 17.8.0 - /opt/homebrew/bin/node Yarn: Not Found npm: 8.5.5 - /opt/homebrew/bin/npm Watchman: Not Found Managers: CocoaPods: 1.11.3 - /usr/local/bin/pod SDKs: iOS SDK: Platforms: DriverKit 21.4, iOS 15.4, macOS 12.3, tvOS 15.4, watchOS 8.5 Android SDK: Not Found IDEs: Android Studio: Not Found Xcode: 13.3/13E113 - /usr/bin/xcodebuild Languages: Java: Not Found npmPackages: @react-native-community/cli: Not Found react: ^17.0.1 => 17.0.2 react-native: 0.64.3 => 0.64.3 react-native-macos: Not Found npmGlobalPackages: *react-native*: Not Found ### Steps to reproduce Setup a simple React Native App with a Text Input inside a Text component ### Snack, code example, screenshot, or link to a repository _No response_
non_defect
nested textinput inside text looses focus on key stroke description hello i have a nested text setup with text and text inputs like this from the styling and text wrapping perspective it works great but unfortunately the text input looses focus on every keystroke is this a common thing or am i missing something here version output of npx react native info system os macos cpu apple memory mb gb shell bin zsh binaries node opt homebrew bin node yarn not found npm opt homebrew bin npm watchman not found managers cocoapods usr local bin pod sdks ios sdk platforms driverkit ios macos tvos watchos android sdk not found ides android studio not found xcode usr bin xcodebuild languages java not found npmpackages react native community cli not found react react native react native macos not found npmglobalpackages react native not found steps to reproduce setup a simple react native app with a text input inside a text component snack code example screenshot or link to a repository no response
0
34,503
7,452,434,087
IssuesEvent
2018-03-29 08:22:51
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
valdkonna ja kirjeldusüksuse vahelist seost ei saa tekitada
P: normal R: duplicate T: defect
**Reported by katrin vesterblom on 5 Jun 2013 13:19 UTC** rahvusarhiiv.tietotest.ee, sisse loginud test.test timebox-i järgi peaks olema realiseeritud valdkonna ja ühe KÜ vahelise seose loomine. Praktikas see ei õnnestu: KÜ vormil pole praegu üldse Valdkonna ega Valdkonnamärksõna sakki, akent, vms. Valdkonna vaates on küll olemas võimalus seostada KÜ-ga, aga ükskõik milliseid leidandmeid seostamiseks otsin (ERA.1, TLA.777, Politseitalitus), midagi ma sinna ei saa. TLA.777 annab üsna ruttu teate "Tulemusi ei leitud", kuigi selliste leidandmetega kirjeldusüksus (arhiiv) on olemas. ERA.1 ja Politseitalitus jätavad pikaks ajaks ette teate "Otsib..." ja ei jõua kuskile välja. ERA.1.1.1 annab ka üsna ruttu teate "Tulemusi ei leitud", kuigi ka selline kirjeldusüksus (selliste leidandmetega) on olemas.
1.0
valdkonna ja kirjeldusüksuse vahelist seost ei saa tekitada - **Reported by katrin vesterblom on 5 Jun 2013 13:19 UTC** rahvusarhiiv.tietotest.ee, sisse loginud test.test timebox-i järgi peaks olema realiseeritud valdkonna ja ühe KÜ vahelise seose loomine. Praktikas see ei õnnestu: KÜ vormil pole praegu üldse Valdkonna ega Valdkonnamärksõna sakki, akent, vms. Valdkonna vaates on küll olemas võimalus seostada KÜ-ga, aga ükskõik milliseid leidandmeid seostamiseks otsin (ERA.1, TLA.777, Politseitalitus), midagi ma sinna ei saa. TLA.777 annab üsna ruttu teate "Tulemusi ei leitud", kuigi selliste leidandmetega kirjeldusüksus (arhiiv) on olemas. ERA.1 ja Politseitalitus jätavad pikaks ajaks ette teate "Otsib..." ja ei jõua kuskile välja. ERA.1.1.1 annab ka üsna ruttu teate "Tulemusi ei leitud", kuigi ka selline kirjeldusüksus (selliste leidandmetega) on olemas.
defect
valdkonna ja kirjeldusüksuse vahelist seost ei saa tekitada reported by katrin vesterblom on jun utc rahvusarhiiv tietotest ee sisse loginud test test timebox i järgi peaks olema realiseeritud valdkonna ja ühe kü vahelise seose loomine praktikas see ei õnnestu kü vormil pole praegu üldse valdkonna ega valdkonnamärksõna sakki akent vms valdkonna vaates on küll olemas võimalus seostada kü ga aga ükskõik milliseid leidandmeid seostamiseks otsin era tla politseitalitus midagi ma sinna ei saa tla annab üsna ruttu teate tulemusi ei leitud kuigi selliste leidandmetega kirjeldusüksus arhiiv on olemas era ja politseitalitus jätavad pikaks ajaks ette teate otsib ja ei jõua kuskile välja era annab ka üsna ruttu teate tulemusi ei leitud kuigi ka selline kirjeldusüksus selliste leidandmetega on olemas
1
9,679
2,615,165,207
IssuesEvent
2015-03-01 06:45:22
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
Reaver is skipping correct pin due to fake WSC NACK Respond
auto-migrated Priority-Triage Type-Defect
``` Hi there I have a problem whit Reaver problem is that Reaver is skipping pins due to Fake nack responds received after M4 message after checking second half of pin in previous attempt. Signal strength:58db. Wireless card used: Realtek RTL8187 built-in Toshiba Satellite in monitor mode Router Model TP-Link TL-WR720N (TP-LINK Wireless Router WR720N) distance about 10-15 meters Latest version of Reaver downloaded from Backtrack Linux Repo using apt-get update apt-get install Reaver commands. 6. Please describe what you think the issue is. I think there must be an additional check that previous half of pin was correct or not. [+] Trying pin 43620852 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 43620869 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 43620876 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 43620883 [+] Sending EAPOL START request root@bt:~# reaver -i mon0 -b 64:70:02:DD:F4:8C -vv --nack ``` Original issue reported on code.google.com by `kostad...@yahoo.com` on 23 Jan 2013 at 7:33
1.0
Reaver is skipping correct pin due to fake WSC NACK Respond - ``` Hi there I have a problem whit Reaver problem is that Reaver is skipping pins due to Fake nack responds received after M4 message after checking second half of pin in previous attempt. Signal strength:58db. Wireless card used: Realtek RTL8187 built-in Toshiba Satellite in monitor mode Router Model TP-Link TL-WR720N (TP-LINK Wireless Router WR720N) distance about 10-15 meters Latest version of Reaver downloaded from Backtrack Linux Repo using apt-get update apt-get install Reaver commands. 6. Please describe what you think the issue is. I think there must be an additional check that previous half of pin was correct or not. [+] Trying pin 43620852 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 43620869 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 43620876 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 43620883 [+] Sending EAPOL START request root@bt:~# reaver -i mon0 -b 64:70:02:DD:F4:8C -vv --nack ``` Original issue reported on code.google.com by `kostad...@yahoo.com` on 23 Jan 2013 at 7:33
defect
reaver is skipping correct pin due to fake wsc nack respond hi there i have a problem whit reaver problem is that reaver is skipping pins due to fake nack responds received after message after checking second half of pin in previous attempt signal strength wireless card used realtek built in toshiba satellite in monitor mode router model tp link tl tp link wireless router distance about meters latest version of reaver downloaded from backtrack linux repo using apt get update apt get install reaver commands please describe what you think the issue is i think there must be an additional check that previous half of pin was correct or not trying pin sending eapol start request received identity request sending identity response received identity request sending identity response received message sending message received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request received identity request sending identity response received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request received identity request sending identity response received identity request sending identity response received message sending message received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request root bt reaver i b dd vv nack original issue reported on code google com by kostad yahoo com on jan at
1
196,787
14,889,996,791
IssuesEvent
2021-01-20 22:17:21
golang/go
https://api.github.com/repos/golang/go
closed
std,cmd: add test to ensure that all bundled packages are in sync with go.mod version
NeedsFix Testing release-blocker
We've had issues in the past where version skew could be introduced if people forgot to update one part of the distribution without also updating another (e.g., versions in src/go.mod and src/cmd/go.mod). A very effective solution was adding a test that fails when skew is detected. I believe we still need to address this for net/http's bundled copies of `golang.org/x/net/http2` and `golang.org/x/net/internal/socks`. Right now, it's possible to update the bundled copy independently from the src/go.mod version, but it's not clear that flexibility is something we want to keep. See #41375 where this came up. Also relevant is #25285. I plan to revisit my review of [CL 189818](https://golang.org/cl/189818) (which fixes issue #32031), the trade-offs have changed by now (especially given #41330). /cc @toothrot @bcmills @FiloSottile
1.0
std,cmd: add test to ensure that all bundled packages are in sync with go.mod version - We've had issues in the past where version skew could be introduced if people forgot to update one part of the distribution without also updating another (e.g., versions in src/go.mod and src/cmd/go.mod). A very effective solution was adding a test that fails when skew is detected. I believe we still need to address this for net/http's bundled copies of `golang.org/x/net/http2` and `golang.org/x/net/internal/socks`. Right now, it's possible to update the bundled copy independently from the src/go.mod version, but it's not clear that flexibility is something we want to keep. See #41375 where this came up. Also relevant is #25285. I plan to revisit my review of [CL 189818](https://golang.org/cl/189818) (which fixes issue #32031), the trade-offs have changed by now (especially given #41330). /cc @toothrot @bcmills @FiloSottile
non_defect
std cmd add test to ensure that all bundled packages are in sync with go mod version we ve had issues in the past where version skew could be introduced if people forgot to update one part of the distribution without also updating another e g versions in src go mod and src cmd go mod a very effective solution was adding a test that fails when skew is detected i believe we still need to address this for net http s bundled copies of golang org x net and golang org x net internal socks right now it s possible to update the bundled copy independently from the src go mod version but it s not clear that flexibility is something we want to keep see where this came up also relevant is i plan to revisit my review of which fixes issue the trade offs have changed by now especially given cc toothrot bcmills filosottile
0
124,557
16,614,262,668
IssuesEvent
2021-06-02 14:55:04
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
opened
Grayer Undead Hairs
:art:character designing😶 :exclamation: priority high :grey_exclamation: priority low
<!-- DO NOT REMOVE PRE-EXISTING LINES IF YOU WANT TO SUGGEST A FEW THINGS, OPEN A NEW ISSUE PER EVERY SUGGESTION ---------------------------------------------------------------------------------------------------------- --> **Describe your suggestion in full detail below:** So we have the undead trait in `wc_undead_modifier.txt` and it makes the skin pale and green, but I think we also should make hairs grayer and paler too.
1.0
Grayer Undead Hairs - <!-- DO NOT REMOVE PRE-EXISTING LINES IF YOU WANT TO SUGGEST A FEW THINGS, OPEN A NEW ISSUE PER EVERY SUGGESTION ---------------------------------------------------------------------------------------------------------- --> **Describe your suggestion in full detail below:** So we have the undead trait in `wc_undead_modifier.txt` and it makes the skin pale and green, but I think we also should make hairs grayer and paler too.
non_defect
grayer undead hairs do not remove pre existing lines if you want to suggest a few things open a new issue per every suggestion describe your suggestion in full detail below so we have the undead trait in wc undead modifier txt and it makes the skin pale and green but i think we also should make hairs grayer and paler too
0
169,524
13,150,495,196
IssuesEvent
2020-08-09 11:53:44
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
ToDate("2012-03-01","XXX") test failure & wasted conversion attempts
bug expression language tests
I'm investigating a local test failure which I'm unsure whether is specific to OS X or just how timezones are set up on my particular machine. The failing test is: Assert.assertEquals(invoke("toDate", "2012-03-01","XXX"), invoke("toDate", "2012-03-01")); which returns 2012-03-01T01:00:00 instead of 2012-03-01T00:00:00 This appears to have something to do with the default fallthrough case here, perhaps due to Daylight Savings Time: https://github.com/OpenRefine/OpenRefine/blob/306b541c699e7e87d5e08be1310575b85b15dbf0/main/src/com/google/refine/expr/functions/ToDate.java#L170-L172 While investigating this, I came across this code: https://github.com/OpenRefine/OpenRefine/blob/306b541c699e7e87d5e08be1310575b85b15dbf0/main/src/com/google/refine/expr/functions/ToDate.java#L135-L145 which I don't completely understand, but in the test case creates an array with 160 copies of the (invalid) code "XXX" which it then uses to attempt to parse the date 160 times, always failing, before falling back to the code above
1.0
ToDate("2012-03-01","XXX") test failure & wasted conversion attempts - I'm investigating a local test failure which I'm unsure whether is specific to OS X or just how timezones are set up on my particular machine. The failing test is: Assert.assertEquals(invoke("toDate", "2012-03-01","XXX"), invoke("toDate", "2012-03-01")); which returns 2012-03-01T01:00:00 instead of 2012-03-01T00:00:00 This appears to have something to do with the default fallthrough case here, perhaps due to Daylight Savings Time: https://github.com/OpenRefine/OpenRefine/blob/306b541c699e7e87d5e08be1310575b85b15dbf0/main/src/com/google/refine/expr/functions/ToDate.java#L170-L172 While investigating this, I came across this code: https://github.com/OpenRefine/OpenRefine/blob/306b541c699e7e87d5e08be1310575b85b15dbf0/main/src/com/google/refine/expr/functions/ToDate.java#L135-L145 which I don't completely understand, but in the test case creates an array with 160 copies of the (invalid) code "XXX" which it then uses to attempt to parse the date 160 times, always failing, before falling back to the code above
non_defect
todate xxx test failure wasted conversion attempts i m investigating a local test failure which i m unsure whether is specific to os x or just how timezones are set up on my particular machine the failing test is assert assertequals invoke todate xxx invoke todate which returns instead of this appears to have something to do with the default fallthrough case here perhaps due to daylight savings time while investigating this i came across this code which i don t completely understand but in the test case creates an array with copies of the invalid code xxx which it then uses to attempt to parse the date times always failing before falling back to the code above
0
197,228
14,913,889,912
IssuesEvent
2021-01-22 14:43:59
golang/go
https://api.github.com/repos/golang/go
closed
net: TestReadFromTimeoutFluctuation and TestWriteTimeoutFluctuation trybot flake on OpenBSD
NeedsFix OS-OpenBSD Testing help wanted
Test flake on OpenBSD: https://storage.googleapis.com/go-build-log/4f0cbd68/openbsd-amd64-60_a0ca068c.log ``` --- FAIL: TestReadFromTimeoutFluctuation (1.48s) timeout_test.go:727: ReadFrom took over 1s; expected 0.1s --- FAIL: TestWriteTimeoutFluctuation (1.96s) timeout_test.go:769: Write took over 1s; expected 0.1s FAIL FAIL net 33.085s ```
1.0
net: TestReadFromTimeoutFluctuation and TestWriteTimeoutFluctuation trybot flake on OpenBSD - Test flake on OpenBSD: https://storage.googleapis.com/go-build-log/4f0cbd68/openbsd-amd64-60_a0ca068c.log ``` --- FAIL: TestReadFromTimeoutFluctuation (1.48s) timeout_test.go:727: ReadFrom took over 1s; expected 0.1s --- FAIL: TestWriteTimeoutFluctuation (1.96s) timeout_test.go:769: Write took over 1s; expected 0.1s FAIL FAIL net 33.085s ```
non_defect
net testreadfromtimeoutfluctuation and testwritetimeoutfluctuation trybot flake on openbsd test flake on openbsd fail testreadfromtimeoutfluctuation timeout test go readfrom took over expected fail testwritetimeoutfluctuation timeout test go write took over expected fail fail net
0
71,045
23,422,919,106
IssuesEvent
2022-08-14 00:44:35
jezzsantos/automate
https://api.github.com/repos/jezzsantos/automate
closed
Switching pattern on ID, not Name
defect-design
We should be switching pattern based on ID not based on its name. `automate edit switch <PATTERNID>`
1.0
Switching pattern on ID, not Name - We should be switching pattern based on ID not based on its name. `automate edit switch <PATTERNID>`
defect
switching pattern on id not name we should be switching pattern based on id not based on its name automate edit switch
1
18,040
12,743,597,370
IssuesEvent
2020-06-26 10:44:44
OpenLiberty/blogs
https://api.github.com/repos/OpenLiberty/blogs
closed
Refine instructions on how to contribute blog posts
infrastructure
Right now the instructions are in a rough state, but are enough to get someone going with blog development. 1. Update OpenLiberty/openliberty.io README to point to this repository to find blog authoring instructions. 2. Update the Dockerfile's comments to reflect how to use the docker image. File here https://github.com/OpenLiberty/openliberty.io/blob/master/Dockerfile 3. Decide if we want people to build the docker image or pull from dockerhub (at the moment hosted on kinueng/openliberty.io) 4. Add maybe some instructions about using the staging site (openlibertydev) after using the docker image. 5. Evaluate who must review pull requests to this repository FYI @lauracowen
1.0
Refine instructions on how to contribute blog posts - Right now the instructions are in a rough state, but are enough to get someone going with blog development. 1. Update OpenLiberty/openliberty.io README to point to this repository to find blog authoring instructions. 2. Update the Dockerfile's comments to reflect how to use the docker image. File here https://github.com/OpenLiberty/openliberty.io/blob/master/Dockerfile 3. Decide if we want people to build the docker image or pull from dockerhub (at the moment hosted on kinueng/openliberty.io) 4. Add maybe some instructions about using the staging site (openlibertydev) after using the docker image. 5. Evaluate who must review pull requests to this repository FYI @lauracowen
non_defect
refine instructions on how to contribute blog posts right now the instructions are in a rough state but are enough to get someone going with blog development update openliberty openliberty io readme to point to this repository to find blog authoring instructions update the dockerfile s comments to reflect how to use the docker image file here decide if we want people to build the docker image or pull from dockerhub at the moment hosted on kinueng openliberty io add maybe some instructions about using the staging site openlibertydev after using the docker image evaluate who must review pull requests to this repository fyi lauracowen
0
182,706
21,673,925,228
IssuesEvent
2022-05-08 12:06:05
turkdevops/vscode
https://api.github.com/repos/turkdevops/vscode
closed
CVE-2021-23362 (Medium) detected in hosted-git-info-2.5.0.tgz - autoclosed
security vulnerability
## CVE-2021-23362 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.5.0.tgz</b></p></summary> <p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p> <p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.5.0.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.5.0.tgz</a></p> <p>Path to dependency file: /extensions/emmet/package.json</p> <p>Path to vulnerable library: /extensions/emmet/node_modules/hosted-git-info/package.json</p> <p> Dependency Hierarchy: - vscode-1.0.1.tgz (Root Library) - gulp-untar-0.0.4.tgz - gulp-util-2.2.20.tgz - dateformat-1.0.12.tgz - meow-3.7.0.tgz - normalize-package-data-2.4.0.tgz - :x: **hosted-git-info-2.5.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/vscode/commit/9fd6b056a06e14655f4f0b0f631d670b24878828">9fd6b056a06e14655f4f0b0f631d670b24878828</a></p> <p>Found in base branch: <b>webview-views</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution (hosted-git-info): 2.8.9</p> <p>Direct dependency fix Resolution (vscode): 1.0.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23362 (Medium) detected in hosted-git-info-2.5.0.tgz - autoclosed - ## CVE-2021-23362 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.5.0.tgz</b></p></summary> <p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p> <p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.5.0.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.5.0.tgz</a></p> <p>Path to dependency file: /extensions/emmet/package.json</p> <p>Path to vulnerable library: /extensions/emmet/node_modules/hosted-git-info/package.json</p> <p> Dependency Hierarchy: - vscode-1.0.1.tgz (Root Library) - gulp-untar-0.0.4.tgz - gulp-util-2.2.20.tgz - dateformat-1.0.12.tgz - meow-3.7.0.tgz - normalize-package-data-2.4.0.tgz - :x: **hosted-git-info-2.5.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/vscode/commit/9fd6b056a06e14655f4f0b0f631d670b24878828">9fd6b056a06e14655f4f0b0f631d670b24878828</a></p> <p>Found in base branch: <b>webview-views</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution (hosted-git-info): 2.8.9</p> <p>Direct dependency fix Resolution (vscode): 1.0.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in hosted git info tgz autoclosed cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file extensions emmet package json path to vulnerable library extensions emmet node modules hosted git info package json dependency hierarchy vscode tgz root library gulp untar tgz gulp util tgz dateformat tgz meow tgz normalize package data tgz x hosted git info tgz vulnerable library found in head commit a href found in base branch webview views vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info direct dependency fix resolution vscode step up your open source security game with whitesource
0
20,546
3,373,914,740
IssuesEvent
2015-11-24 10:21:53
NetCDF4Excel/project
https://api.github.com/repos/NetCDF4Excel/project
closed
Run time error 53: File not found: VbNc.dll
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1.Open NetCDF4Excel_2007.xlsm 2.Enable macros (NetCDF) 3.Open File from NetCDF What is the expected output? What do you see instead? The expected output is the file opening up. Instead I get an error message: Run time error 53: File not found: VbNc.dll I have confirmed that the VbNc.dll file is present under the NETCDF4 Excel directory. What version of the product are you using? On what operating system? Version 2.1 on Excel 2007 on Windows 8 ``` Original issue reported on code.google.com by `wertt...@gmail.com` on 16 Jan 2014 at 9:40
1.0
Run time error 53: File not found: VbNc.dll - ``` What steps will reproduce the problem? 1.Open NetCDF4Excel_2007.xlsm 2.Enable macros (NetCDF) 3.Open File from NetCDF What is the expected output? What do you see instead? The expected output is the file opening up. Instead I get an error message: Run time error 53: File not found: VbNc.dll I have confirmed that the VbNc.dll file is present under the NETCDF4 Excel directory. What version of the product are you using? On what operating system? Version 2.1 on Excel 2007 on Windows 8 ``` Original issue reported on code.google.com by `wertt...@gmail.com` on 16 Jan 2014 at 9:40
defect
run time error file not found vbnc dll what steps will reproduce the problem open xlsm enable macros netcdf open file from netcdf what is the expected output what do you see instead the expected output is the file opening up instead i get an error message run time error file not found vbnc dll i have confirmed that the vbnc dll file is present under the excel directory what version of the product are you using on what operating system version on excel on windows original issue reported on code google com by wertt gmail com on jan at
1
251,177
18,940,924,267
IssuesEvent
2021-11-18 02:41:12
spacetelescope/romancal
https://api.github.com/repos/spacetelescope/romancal
opened
Update References General Doc for Jump and Ramp
documentation
_Issue [RCAL-250](https://jira.stsci.edu/browse/RCAL-250) was created on JIRA by [Paul Huwe](https://jira.stsci.edu/secure/ViewProfile.jspa?name=phuwe):_ The references_general.rst file needs to be updated to properly reflect the jump and ramp fitting requirements.
1.0
Update References General Doc for Jump and Ramp - _Issue [RCAL-250](https://jira.stsci.edu/browse/RCAL-250) was created on JIRA by [Paul Huwe](https://jira.stsci.edu/secure/ViewProfile.jspa?name=phuwe):_ The references_general.rst file needs to be updated to properly reflect the jump and ramp fitting requirements.
non_defect
update references general doc for jump and ramp issue was created on jira by the references general rst file needs to be updated to properly reflect the jump and ramp fitting requirements
0
57,734
16,017,984,244
IssuesEvent
2021-04-20 18:31:47
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Can't share screen in desktop app
T-Defect
<!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P --> <!-- Please report security issues by email to security@matrix.org --> <!-- This is a bug report template. By following the instructions below and filling out the sections with your information, you will help the us to get all the necessary data to fix your issue. You can also preview your report before submitting it. You may remove sections that aren't relevant to your particular case. Text between <!-- and --​> marks will be invisible in the report. --> ### Description I can not share my screen in a video call. The application can't find any application windows or screens. ### Steps to reproduce - Open a video call with someone else - Click on the screen share button Describe how what happens differs from what you expected. I can't select an application window to share. The loading screen never finishes. Logs being sent: yes ### Version information - **Platform**: Desktop For the desktop app: - **OS**: Windows 10 - **Version**: 1.7.25
1.0
Can't share screen in desktop app - <!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P --> <!-- Please report security issues by email to security@matrix.org --> <!-- This is a bug report template. By following the instructions below and filling out the sections with your information, you will help the us to get all the necessary data to fix your issue. You can also preview your report before submitting it. You may remove sections that aren't relevant to your particular case. Text between <!-- and --​> marks will be invisible in the report. --> ### Description I can not share my screen in a video call. The application can't find any application windows or screens. ### Steps to reproduce - Open a video call with someone else - Click on the screen share button Describe how what happens differs from what you expected. I can't select an application window to share. The loading screen never finishes. Logs being sent: yes ### Version information - **Platform**: Desktop For the desktop app: - **OS**: Windows 10 - **Version**: 1.7.25
defect
can t share screen in desktop app this is a bug report template by following the instructions below and filling out the sections with your information you will help the us to get all the necessary data to fix your issue you can also preview your report before submitting it you may remove sections that aren t relevant to your particular case text between marks will be invisible in the report description i can not share my screen in a video call the application can t find any application windows or screens steps to reproduce open a video call with someone else click on the screen share button describe how what happens differs from what you expected i can t select an application window to share the loading screen never finishes logs being sent yes version information platform desktop for the desktop app os windows version
1
49,464
6,027,321,858
IssuesEvent
2017-06-08 13:32:55
ionic-team/ionic
https://api.github.com/repos/ionic-team/ionic
closed
ionic2 - bug(overlays) - After clicking to dismiss an overlay, the click propagates to whatever is beneath it.
needs:devicetesting v2
#### Short description of the problem: If I have an overlay open (modal confirmed, not sure about others), and click a dismiss button on it, the transition will run and then once it is dismissed it will click whatever is beneath it, such as an input, etc. This happens on iOS. Not sure about any other platforms. #### What behavior are you expecting? I expect there to not be duplicate clicks. **Steps to reproduce:** 1. Create a page with some inputs on it (I am specifically using standard textareas) 2. Create a modal with a dismiss button directly over one of the textareas 3. Open the modal, click the dismiss button, and observe that the textarea is focused and the keyboard appears **Run `ionic info` from terminal/cmd prompt:** (paste output below) Cordova CLI: 6.2.0 Gulp version: CLI version 3.9.1 Gulp local: Local version 3.9.1 Ionic Framework Version: 2.0.0-rc.1 Ionic CLI Version: 2.1.0 Ionic App Lib Version: 2.1.0-beta.1 ios-deploy version: 1.8.4 ios-sim version: 5.0.6 OS: Mac OS X El Capitan Node Version: v4.5.0 Xcode version: Xcode 8.0 Build version 8A218a
1.0
ionic2 - bug(overlays) - After clicking to dismiss an overlay, the click propagates to whatever is beneath it. - #### Short description of the problem: If I have an overlay open (modal confirmed, not sure about others), and click a dismiss button on it, the transition will run and then once it is dismissed it will click whatever is beneath it, such as an input, etc. This happens on iOS. Not sure about any other platforms. #### What behavior are you expecting? I expect there to not be duplicate clicks. **Steps to reproduce:** 1. Create a page with some inputs on it (I am specifically using standard textareas) 2. Create a modal with a dismiss button directly over one of the textareas 3. Open the modal, click the dismiss button, and observe that the textarea is focused and the keyboard appears **Run `ionic info` from terminal/cmd prompt:** (paste output below) Cordova CLI: 6.2.0 Gulp version: CLI version 3.9.1 Gulp local: Local version 3.9.1 Ionic Framework Version: 2.0.0-rc.1 Ionic CLI Version: 2.1.0 Ionic App Lib Version: 2.1.0-beta.1 ios-deploy version: 1.8.4 ios-sim version: 5.0.6 OS: Mac OS X El Capitan Node Version: v4.5.0 Xcode version: Xcode 8.0 Build version 8A218a
non_defect
bug overlays after clicking to dismiss an overlay the click propagates to whatever is beneath it short description of the problem if i have an overlay open modal confirmed not sure about others and click a dismiss button on it the transition will run and then once it is dismissed it will click whatever is beneath it such as an input etc this happens on ios not sure about any other platforms what behavior are you expecting i expect there to not be duplicate clicks steps to reproduce create a page with some inputs on it i am specifically using standard textareas create a modal with a dismiss button directly over one of the textareas open the modal click the dismiss button and observe that the textarea is focused and the keyboard appears run ionic info from terminal cmd prompt paste output below cordova cli gulp version cli version gulp local local version ionic framework version rc ionic cli version ionic app lib version beta ios deploy version ios sim version os mac os x el capitan node version xcode version xcode build version
0
77,575
27,057,314,359
IssuesEvent
2023-02-13 17:01:48
microsoft/TypeScript
https://api.github.com/repos/microsoft/TypeScript
closed
Missing NOTICE file makes it harder to satisfy terms of license
Not a Defect
# Bug Report ### 🔎 Search Terms NOTICE file ### 🕗 Version & Regression Information I encountered the problem in the 4.9.5 release. I also examined the latest code in the public rep. I reviewed the FAQ for entries about NOTICE file. ### 🙁 Actual behavior There are copyright, patent, trademark, and attribution notices from the Source form of the Work, and the work does not include a NOTICE text file as part of its distribution. The license used by the project is Apache Public License 2.0: https://github.com/microsoft/TypeScript/blob/main/LICENSE.txt That contains these clauses: `You must retain, in the Source form of any Derivative Works that You distribute, all copyright, patent, trademark, and attribution notices from the Source form of the Work, excluding those notices that do not pertain to any part of the Derivative Works; and If the Work includes a "NOTICE" text file as part of its distribution, then any Derivative Works that You distribute must include a readable copy of the attribution notices contained within such NOTICE file, excluding those notices that do not pertain to any part of the Derivative Works` ### 🙂 Expected behavior <!-- What you expected to happen instead, and why --> The work should contain a NOTICE text file per the convention documented here: https://www.apache.org/licenses/LICENSE-2.0#apply
1.0
Missing NOTICE file makes it harder to satisfy terms of license - # Bug Report ### 🔎 Search Terms NOTICE file ### 🕗 Version & Regression Information I encountered the problem in the 4.9.5 release. I also examined the latest code in the public rep. I reviewed the FAQ for entries about NOTICE file. ### 🙁 Actual behavior There are copyright, patent, trademark, and attribution notices from the Source form of the Work, and the work does not include a NOTICE text file as part of its distribution. The license used by the project is Apache Public License 2.0: https://github.com/microsoft/TypeScript/blob/main/LICENSE.txt That contains these clauses: `You must retain, in the Source form of any Derivative Works that You distribute, all copyright, patent, trademark, and attribution notices from the Source form of the Work, excluding those notices that do not pertain to any part of the Derivative Works; and If the Work includes a "NOTICE" text file as part of its distribution, then any Derivative Works that You distribute must include a readable copy of the attribution notices contained within such NOTICE file, excluding those notices that do not pertain to any part of the Derivative Works` ### 🙂 Expected behavior <!-- What you expected to happen instead, and why --> The work should contain a NOTICE text file per the convention documented here: https://www.apache.org/licenses/LICENSE-2.0#apply
defect
missing notice file makes it harder to satisfy terms of license bug report 🔎 search terms notice file 🕗 version regression information i encountered the problem in the release i also examined the latest code in the public rep i reviewed the faq for entries about notice file 🙁 actual behavior there are copyright patent trademark and attribution notices from the source form of the work and the work does not include a notice text file as part of its distribution the license used by the project is apache public license that contains these clauses you must retain in the source form of any derivative works that you distribute all copyright patent trademark and attribution notices from the source form of the work excluding those notices that do not pertain to any part of the derivative works and if the work includes a notice text file as part of its distribution then any derivative works that you distribute must include a readable copy of the attribution notices contained within such notice file excluding those notices that do not pertain to any part of the derivative works 🙂 expected behavior the work should contain a notice text file per the convention documented here
1
57,745
16,025,170,288
IssuesEvent
2021-04-21 08:11:27
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Auto-switching to rooms in spaces is disorienting/broken
A-Spaces S-Major T-Defect Z-Rageshake
It's also unreliable if you are viewing a favourite room and you try to navigate to a different space
1.0
Auto-switching to rooms in spaces is disorienting/broken - It's also unreliable if you are viewing a favourite room and you try to navigate to a different space
defect
auto switching to rooms in spaces is disorienting broken it s also unreliable if you are viewing a favourite room and you try to navigate to a different space
1
55,011
14,121,057,162
IssuesEvent
2020-11-09 00:39:48
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
solve_ivp fails to solve with non-zero initial time, large parameters, and methods other than LSODA
defect scipy.integrate
As the title says, `scipy.integrate.solve_ivp` fail to solve with a non-zero initial time, large parameters, and methods other than LSODA. Remove any one of those conditions and it solves. Here is an ODE that demonstrates the problem. ``` import numpy as np import scipy.integrate c_0 = np.array([1.0e6]) def ode(t, y): c = y[0] cdot = C / c RHS = np.array([cdot]) return(RHS) ``` Attempting to integrate from a non-zero initial time fails to solve. For example, ``` C = 1e50 t_span = (1.0, 2.0) ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'Radau') print(ivp_sol.message) ``` prints the following to the screen ``` Required step size is less than spacing between numbers. ``` Attempting to integrate from a zero initial time or with a smaller value of `C` runs fine. For example, ``` C = 1e50 t_span = (0.0, 2.0) ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'Radau') print(ivp_sol.message) C = 1e25 t_span = (1.0, 2.0) ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'Radau') print(ivp_sol.message) ``` prints the following to the screen ``` The solver successfully reached the end of the integration interval. The solver successfully reached the end of the integration interval. ``` This problem seems to involve all methods, except for LSODA, although LSODA still complains about small step sizes. Running ``` C = 1e50 t_span = (1.0, 2.0) ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'LSODA') print(ivp_sol.message) ``` results in several warnings like ``` lsoda-- warning..internal t (=r1) and h (=r2) are�� such that in the machine, t + h = t on the next step (h = step size). solver will continue anyway�� in above, r1 = 0.1000000000000D+01 r2 = 0.3162277663331D-39 ``` but it completes with ``` The solver successfully reached the end of the integration interval. ``` It’s not surprising that some methods might work better than others for stiff problems, but it should not matter if the initial time is zero or not. Here are my version numbers. ``` import sys, scipy, numpy; print(scipy.__version__, numpy.__version__, sys.version_info) ``` produces ``` ('1.2.1', '1.16.4', sys.version_info(major=2, minor=7, micro=16, releaselevel='final', serial=0)) ```
1.0
solve_ivp fails to solve with non-zero initial time, large parameters, and methods other than LSODA - As the title says, `scipy.integrate.solve_ivp` fail to solve with a non-zero initial time, large parameters, and methods other than LSODA. Remove any one of those conditions and it solves. Here is an ODE that demonstrates the problem. ``` import numpy as np import scipy.integrate c_0 = np.array([1.0e6]) def ode(t, y): c = y[0] cdot = C / c RHS = np.array([cdot]) return(RHS) ``` Attempting to integrate from a non-zero initial time fails to solve. For example, ``` C = 1e50 t_span = (1.0, 2.0) ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'Radau') print(ivp_sol.message) ``` prints the following to the screen ``` Required step size is less than spacing between numbers. ``` Attempting to integrate from a zero initial time or with a smaller value of `C` runs fine. For example, ``` C = 1e50 t_span = (0.0, 2.0) ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'Radau') print(ivp_sol.message) C = 1e25 t_span = (1.0, 2.0) ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'Radau') print(ivp_sol.message) ``` prints the following to the screen ``` The solver successfully reached the end of the integration interval. The solver successfully reached the end of the integration interval. ``` This problem seems to involve all methods, except for LSODA, although LSODA still complains about small step sizes. Running ``` C = 1e50 t_span = (1.0, 2.0) ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'LSODA') print(ivp_sol.message) ``` results in several warnings like ``` lsoda-- warning..internal t (=r1) and h (=r2) are�� such that in the machine, t + h = t on the next step (h = step size). solver will continue anyway�� in above, r1 = 0.1000000000000D+01 r2 = 0.3162277663331D-39 ``` but it completes with ``` The solver successfully reached the end of the integration interval. ``` It’s not surprising that some methods might work better than others for stiff problems, but it should not matter if the initial time is zero or not. Here are my version numbers. ``` import sys, scipy, numpy; print(scipy.__version__, numpy.__version__, sys.version_info) ``` produces ``` ('1.2.1', '1.16.4', sys.version_info(major=2, minor=7, micro=16, releaselevel='final', serial=0)) ```
defect
solve ivp fails to solve with non zero initial time large parameters and methods other than lsoda as the title says scipy integrate solve ivp fail to solve with a non zero initial time large parameters and methods other than lsoda remove any one of those conditions and it solves here is an ode that demonstrates the problem import numpy as np import scipy integrate c np array def ode t y c y cdot c c rhs np array return rhs attempting to integrate from a non zero initial time fails to solve for example c t span ivp sol scipy integrate solve ivp ode t span c method radau print ivp sol message prints the following to the screen required step size is less than spacing between numbers attempting to integrate from a zero initial time or with a smaller value of c runs fine for example c t span ivp sol scipy integrate solve ivp ode t span c method radau print ivp sol message c t span ivp sol scipy integrate solve ivp ode t span c method radau print ivp sol message prints the following to the screen the solver successfully reached the end of the integration interval the solver successfully reached the end of the integration interval this problem seems to involve all methods except for lsoda although lsoda still complains about small step sizes running c t span ivp sol scipy integrate solve ivp ode t span c method lsoda print ivp sol message results in several warnings like lsoda warning internal t and h are�� such that in the machine t h t on the next step h step size solver will continue anyway�� in above but it completes with the solver successfully reached the end of the integration interval it’s not surprising that some methods might work better than others for stiff problems but it should not matter if the initial time is zero or not here are my version numbers import sys scipy numpy print scipy version numpy version sys version info produces sys version info major minor micro releaselevel final serial
1
391,455
26,893,542,131
IssuesEvent
2023-02-06 10:40:26
Arquisoft/lomap_en1b
https://api.github.com/repos/Arquisoft/lomap_en1b
closed
Adding names and emails to documentation
documentation:page_facing_up::pen:
Every team member should add his/her name and email into a new section at the end of the README and merge it into develop branch via pull requests. When the README is finished, it should be pushed into master.
1.0
Adding names and emails to documentation - Every team member should add his/her name and email into a new section at the end of the README and merge it into develop branch via pull requests. When the README is finished, it should be pushed into master.
non_defect
adding names and emails to documentation every team member should add his her name and email into a new section at the end of the readme and merge it into develop branch via pull requests when the readme is finished it should be pushed into master
0
37,687
8,474,803,116
IssuesEvent
2018-10-24 17:08:25
brainvisa/testbidon
https://api.github.com/repos/brainvisa/testbidon
closed
OpenSlide plugin: ineffecient read when used with borders
Category: soma-io Component: Resolution Priority: Normal Status: Closed Tracker: Defect
--- Author Name: **Souedet, Nicolas** (Souedet, Nicolas) Original Redmine Issue: 14573, https://bioproj.extra.cea.fr/redmine/issues/14573 Original Date: 2016-03-18 Original Assignee: Souedet, Nicolas --- Openslide and especially CZI format is based on tiled images. To fill memory with borders and because the soma-io openslide plugin does not support strides yet, VolumeFormatReader reads line after line to fill memory. But to read a part of the line, it is necessary to read a full tile and eventually to uncompress data. Each tile is put in the openslide cache, but the size of the openslide cache is limited, and often not big enough to store all the tiles of a line. To fix this problem, it is necessary to add support for strides in the soma-io plugin and to read data in a tiled way.
1.0
OpenSlide plugin: ineffecient read when used with borders - --- Author Name: **Souedet, Nicolas** (Souedet, Nicolas) Original Redmine Issue: 14573, https://bioproj.extra.cea.fr/redmine/issues/14573 Original Date: 2016-03-18 Original Assignee: Souedet, Nicolas --- Openslide and especially CZI format is based on tiled images. To fill memory with borders and because the soma-io openslide plugin does not support strides yet, VolumeFormatReader reads line after line to fill memory. But to read a part of the line, it is necessary to read a full tile and eventually to uncompress data. Each tile is put in the openslide cache, but the size of the openslide cache is limited, and often not big enough to store all the tiles of a line. To fix this problem, it is necessary to add support for strides in the soma-io plugin and to read data in a tiled way.
defect
openslide plugin ineffecient read when used with borders author name souedet nicolas souedet nicolas original redmine issue original date original assignee souedet nicolas openslide and especially czi format is based on tiled images to fill memory with borders and because the soma io openslide plugin does not support strides yet volumeformatreader reads line after line to fill memory but to read a part of the line it is necessary to read a full tile and eventually to uncompress data each tile is put in the openslide cache but the size of the openslide cache is limited and often not big enough to store all the tiles of a line to fix this problem it is necessary to add support for strides in the soma io plugin and to read data in a tiled way
1
249,513
26,945,769,816
IssuesEvent
2023-02-08 07:57:03
jgeraigery/pyadi-iio
https://api.github.com/repos/jgeraigery/pyadi-iio
opened
paramiko-3.0.0-py3-none-any.whl: 1 vulnerabilities (highest severity is: 4.8)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>paramiko-3.0.0-py3-none-any.whl</b></p></summary> <p></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/pyadi-iio/commit/5be07ac2fc964a356b857941afa1941e3aa207ec">5be07ac2fc964a356b857941afa1941e3aa207ec</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (paramiko version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2023-23931](https://www.mend.io/vulnerability-database/CVE-2023-23931) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.8 | cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2023-23931</summary> ### Vulnerable Library - <b>cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl</b></p> <p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/fc/b2/3b946e24de214fc49adeefeea6214bcbc4bce2bd745877f074d1dd13c9a2/cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl">https://files.pythonhosted.org/packages/fc/b2/3b946e24de214fc49adeefeea6214bcbc4bce2bd745877f074d1dd13c9a2/cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - paramiko-3.0.0-py3-none-any.whl (Root Library) - :x: **cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/pyadi-iio/commit/5be07ac2fc964a356b857941afa1941e3aa207ec">5be07ac2fc964a356b857941afa1941e3aa207ec</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> cryptography is a package designed to expose cryptographic primitives and recipes to Python developers. In affected versions `Cipher.update_into` would accept Python objects which implement the buffer protocol, but provide only immutable buffers. This would allow immutable objects (such as `bytes`) to be mutated, thus violating fundamental rules of Python and resulting in corrupted output. This now correctly raises an exception. This issue has been present since `update_into` was originally introduced in cryptography 1.8. <p>Publish Date: 2023-02-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23931>CVE-2023-23931</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>4.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-23931">https://www.cve.org/CVERecord?id=CVE-2023-23931</a></p> <p>Release Date: 2023-02-07</p> <p>Fix Resolution: cryptography - 39.0.1</p> </p> <p></p> </details>
True
paramiko-3.0.0-py3-none-any.whl: 1 vulnerabilities (highest severity is: 4.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>paramiko-3.0.0-py3-none-any.whl</b></p></summary> <p></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/pyadi-iio/commit/5be07ac2fc964a356b857941afa1941e3aa207ec">5be07ac2fc964a356b857941afa1941e3aa207ec</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (paramiko version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2023-23931](https://www.mend.io/vulnerability-database/CVE-2023-23931) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.8 | cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2023-23931</summary> ### Vulnerable Library - <b>cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl</b></p> <p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/fc/b2/3b946e24de214fc49adeefeea6214bcbc4bce2bd745877f074d1dd13c9a2/cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl">https://files.pythonhosted.org/packages/fc/b2/3b946e24de214fc49adeefeea6214bcbc4bce2bd745877f074d1dd13c9a2/cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - paramiko-3.0.0-py3-none-any.whl (Root Library) - :x: **cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/pyadi-iio/commit/5be07ac2fc964a356b857941afa1941e3aa207ec">5be07ac2fc964a356b857941afa1941e3aa207ec</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> cryptography is a package designed to expose cryptographic primitives and recipes to Python developers. In affected versions `Cipher.update_into` would accept Python objects which implement the buffer protocol, but provide only immutable buffers. This would allow immutable objects (such as `bytes`) to be mutated, thus violating fundamental rules of Python and resulting in corrupted output. This now correctly raises an exception. This issue has been present since `update_into` was originally introduced in cryptography 1.8. <p>Publish Date: 2023-02-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23931>CVE-2023-23931</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>4.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-23931">https://www.cve.org/CVERecord?id=CVE-2023-23931</a></p> <p>Release Date: 2023-02-07</p> <p>Fix Resolution: cryptography - 39.0.1</p> </p> <p></p> </details>
non_defect
paramiko none any whl vulnerabilities highest severity is vulnerable library paramiko none any whl path to dependency file requirements txt path to vulnerable library requirements txt found in head commit a href vulnerabilities cve severity cvss dependency type fixed in paramiko version remediation available medium cryptography manylinux whl transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library cryptography manylinux whl cryptography is a package which provides cryptographic recipes and primitives to python developers library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy paramiko none any whl root library x cryptography manylinux whl vulnerable library found in head commit a href found in base branch master vulnerability details cryptography is a package designed to expose cryptographic primitives and recipes to python developers in affected versions cipher update into would accept python objects which implement the buffer protocol but provide only immutable buffers this would allow immutable objects such as bytes to be mutated thus violating fundamental rules of python and resulting in corrupted output this now correctly raises an exception this issue has been present since update into was originally introduced in cryptography publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cryptography
0
57,550
15,855,675,157
IssuesEvent
2021-04-08 00:26:55
STEllAR-GROUP/hpx
https://api.github.com/repos/STEllAR-GROUP/hpx
closed
HPX-1.6.0 fails to build on Windows 10
compiler: msvc type: defect
## Expected Behavior Successful build using Visual Studio 2019 ## Actual Behavior Build failed with the following errors: Error C2182 'abstract declarator': illegal use of type 'void' 1d_stencil_channel (Examples\1D Stencil\1d_stencil_channel\1d_stencil_channel) C:\Users\phili\Documents\cpp\hpx-1.6.0\libs\core\functional\include\hpx\functional\traits\is_invocable.hpp Note: error points to line 81 of is_invocable.hpp ## Steps to Reproduce the Problem Follow all steps in the HPX manual regarding: How to build HPX under Windows 10 x64 with Visual Studio 2015 ## Specifications - HPX Version: 1.6.0 - Platform (compiler, OS): Visual Studio 2019, Windows 10. - Setting compiler to use C++14 as well as C++17 makes no difference.
1.0
HPX-1.6.0 fails to build on Windows 10 - ## Expected Behavior Successful build using Visual Studio 2019 ## Actual Behavior Build failed with the following errors: Error C2182 'abstract declarator': illegal use of type 'void' 1d_stencil_channel (Examples\1D Stencil\1d_stencil_channel\1d_stencil_channel) C:\Users\phili\Documents\cpp\hpx-1.6.0\libs\core\functional\include\hpx\functional\traits\is_invocable.hpp Note: error points to line 81 of is_invocable.hpp ## Steps to Reproduce the Problem Follow all steps in the HPX manual regarding: How to build HPX under Windows 10 x64 with Visual Studio 2015 ## Specifications - HPX Version: 1.6.0 - Platform (compiler, OS): Visual Studio 2019, Windows 10. - Setting compiler to use C++14 as well as C++17 makes no difference.
defect
hpx fails to build on windows expected behavior successful build using visual studio actual behavior build failed with the following errors error abstract declarator illegal use of type void stencil channel examples stencil stencil channel stencil channel c users phili documents cpp hpx libs core functional include hpx functional traits is invocable hpp note error points to line of is invocable hpp steps to reproduce the problem follow all steps in the hpx manual regarding how to build hpx under windows with visual studio specifications hpx version platform compiler os visual studio windows setting compiler to use c as well as c makes no difference
1
68,944
21,994,370,730
IssuesEvent
2022-05-26 03:44:26
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
opened
[🐛 Bug]: Open.Selenium.InvalidStateException was occurred using IE.
I-defect needs-triaging
### What happened? Open.Selenium.InvalidStateException was occurred using IE. For details and reproduction methods, I wonder if you check items after 'How can we reproduce the issue?'. This exception is occured at InternetExplorerDriver, is not occured at EdgeDriver. I test this 👆 change to EdgeDriver class from InternetExplorerDriver class, test has all passed. Sorry for my poor english, thank you. ### How can we reproduce the issue? ```shell The implementation is like this: private readonly IWebDriver _webDriver; // selenium process run class field. ... public ctor(string webDriverDirectoryPath) { // get webdriver directory absolute path from ctor argument. _webDriver = new InternetExplorerDriver(webDriverDirectoryPath); } public void Init() { // this method is navigating method for looping invoke method page. } // this method called by other class using foreach. public void Invoke(TestCase testCase) { ... _webDriver .FindElement(By.Id("{Input html tag id}")) .Clear(); // error has occurred this point. _webDriver .FindElement(By.Id("{Input html tag id}")) .SendKeys(testCase.Value); ... var selectElement = new SelectElement(_webDriver .FindElement(By.Id("{Select html tag id}")) ); selectElement.SelectByIndex(testCase.SelectIndex); // this method not work. however using for loop and SendKey(Keys.ArrowDown) is work. } That👆 code runnning at Xunit testing. ```cs var testCases = new[] { new TestCase(1, 1, "hoge", "001", "51000", "Yamada"), new TestCase(2, 1, "fuga", "002", "51001", "Tanaka"), new TestCase(3, 1, "piyo", "003", "51002", "Suzuki") }; using var invoker = new SeleniumInvoker(); invoker.Init(); var result1 = invoker.Invoke(testCases[0]); Assert.Equal("OK", result1.CheckResult); Assert.Equal("Yamada", result1.CustomerName); Assert.Empty(result1.ErrorMessages); var result2 = invoker.GetResult(testCases[1]); Assert.Equal("NG", result2.CheckResult); Assert.Equal("Tanaka", result2.CustomerName); Assert.Equal("Invalid Code", result2.ErrorMessages.Single()); var result3 = invoker.GetResult(testCases[2]); Assert.Equal("NG", result3.CheckResult); Assert.Equal("Suzuki", result3.CustomerName); Assert.Equal("No Account", result3.ErrorMessages.Single()); ``` ``` ### Relevant log output ```shell Xunit failed message: OpenQA.Selenium.InvalidElementStateException A JavaScript error was encountered clearing the element. The driver assumes this is because the element is hidden, disabled or read-only, and it must not be to clear the element. at OpenQA.Selenium.WebDriver.UnpackAndThrowOnError(Response errorResponse) at OpenQA.Selenium.WebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters) at OpenQA.Selenium.WebDriver.InternalExecute(String driverCommandToExecute, Dictionary`2 parameters) at OpenQA.Selenium.WebElement.Execute(String commandToExecute, Dictionary`2 parameters) at OpenQA.Selenium.WebElement.Clear() ``` ``` ### Operating System Windows 10, Version 21H1 , OS Build 19043.1706 ### Selenium version C#, Selenium.Support 4.1.1, Selenium.WebDriver 4.1.1 ### What are the browser(s) and version(s) where you see this issue? Internet Explorer Version 21H1 (OS Build 19043.1706) ### What are the browser driver(s) and version(s) where you see this issue? IEDriverServer.exe 4.0.0.0 (64-bit) ### Are you using Selenium Grid? _No response_
1.0
[🐛 Bug]: Open.Selenium.InvalidStateException was occurred using IE. - ### What happened? Open.Selenium.InvalidStateException was occurred using IE. For details and reproduction methods, I wonder if you check items after 'How can we reproduce the issue?'. This exception is occured at InternetExplorerDriver, is not occured at EdgeDriver. I test this 👆 change to EdgeDriver class from InternetExplorerDriver class, test has all passed. Sorry for my poor english, thank you. ### How can we reproduce the issue? ```shell The implementation is like this: private readonly IWebDriver _webDriver; // selenium process run class field. ... public ctor(string webDriverDirectoryPath) { // get webdriver directory absolute path from ctor argument. _webDriver = new InternetExplorerDriver(webDriverDirectoryPath); } public void Init() { // this method is navigating method for looping invoke method page. } // this method called by other class using foreach. public void Invoke(TestCase testCase) { ... _webDriver .FindElement(By.Id("{Input html tag id}")) .Clear(); // error has occurred this point. _webDriver .FindElement(By.Id("{Input html tag id}")) .SendKeys(testCase.Value); ... var selectElement = new SelectElement(_webDriver .FindElement(By.Id("{Select html tag id}")) ); selectElement.SelectByIndex(testCase.SelectIndex); // this method not work. however using for loop and SendKey(Keys.ArrowDown) is work. } That👆 code runnning at Xunit testing. ```cs var testCases = new[] { new TestCase(1, 1, "hoge", "001", "51000", "Yamada"), new TestCase(2, 1, "fuga", "002", "51001", "Tanaka"), new TestCase(3, 1, "piyo", "003", "51002", "Suzuki") }; using var invoker = new SeleniumInvoker(); invoker.Init(); var result1 = invoker.Invoke(testCases[0]); Assert.Equal("OK", result1.CheckResult); Assert.Equal("Yamada", result1.CustomerName); Assert.Empty(result1.ErrorMessages); var result2 = invoker.GetResult(testCases[1]); Assert.Equal("NG", result2.CheckResult); Assert.Equal("Tanaka", result2.CustomerName); Assert.Equal("Invalid Code", result2.ErrorMessages.Single()); var result3 = invoker.GetResult(testCases[2]); Assert.Equal("NG", result3.CheckResult); Assert.Equal("Suzuki", result3.CustomerName); Assert.Equal("No Account", result3.ErrorMessages.Single()); ``` ``` ### Relevant log output ```shell Xunit failed message: OpenQA.Selenium.InvalidElementStateException A JavaScript error was encountered clearing the element. The driver assumes this is because the element is hidden, disabled or read-only, and it must not be to clear the element. at OpenQA.Selenium.WebDriver.UnpackAndThrowOnError(Response errorResponse) at OpenQA.Selenium.WebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters) at OpenQA.Selenium.WebDriver.InternalExecute(String driverCommandToExecute, Dictionary`2 parameters) at OpenQA.Selenium.WebElement.Execute(String commandToExecute, Dictionary`2 parameters) at OpenQA.Selenium.WebElement.Clear() ``` ``` ### Operating System Windows 10, Version 21H1 , OS Build 19043.1706 ### Selenium version C#, Selenium.Support 4.1.1, Selenium.WebDriver 4.1.1 ### What are the browser(s) and version(s) where you see this issue? Internet Explorer Version 21H1 (OS Build 19043.1706) ### What are the browser driver(s) and version(s) where you see this issue? IEDriverServer.exe 4.0.0.0 (64-bit) ### Are you using Selenium Grid? _No response_
defect
open selenium invalidstateexception was occurred using ie what happened open selenium invalidstateexception was occurred using ie for details and reproduction methods i wonder if you check items after how can we reproduce the issue this exception is occured at internetexplorerdriver is not occured at edgedriver i test this 👆 change to edgedriver class from internetexplorerdriver class test has all passed sorry for my poor english thank you how can we reproduce the issue shell the implementation is like this private readonly iwebdriver webdriver selenium process run class field public ctor string webdriverdirectorypath get webdriver directory absolute path from ctor argument webdriver new internetexplorerdriver webdriverdirectorypath public void init this method is navigating method for looping invoke method page this method called by other class using foreach public void invoke testcase testcase webdriver findelement by id input html tag id clear error has occurred this point webdriver findelement by id input html tag id sendkeys testcase value var selectelement new selectelement webdriver findelement by id select html tag id selectelement selectbyindex testcase selectindex this method not work however using for loop and sendkey keys arrowdown is work that👆 code runnning at xunit testing cs var testcases new new testcase hoge yamada new testcase fuga tanaka new testcase piyo suzuki using var invoker new seleniuminvoker invoker init var invoker invoke testcases assert equal ok checkresult assert equal yamada customername assert empty errormessages var invoker getresult testcases assert equal ng checkresult assert equal tanaka customername assert equal invalid code errormessages single var invoker getresult testcases assert equal ng checkresult assert equal suzuki customername assert equal no account errormessages single relevant log output shell xunit failed message openqa selenium invalidelementstateexception a javascript error was encountered clearing the element the driver assumes this is because the element is hidden disabled or read only and it must not be to clear the element at openqa selenium webdriver unpackandthrowonerror response errorresponse at openqa selenium webdriver execute string drivercommandtoexecute dictionary parameters at openqa selenium webdriver internalexecute string drivercommandtoexecute dictionary parameters at openqa selenium webelement execute string commandtoexecute dictionary parameters at openqa selenium webelement clear operating system windows version os build selenium version c selenium support selenium webdriver what are the browser s and version s where you see this issue internet explorer version os build what are the browser driver s and version s where you see this issue iedriverserver exe bit are you using selenium grid no response
1
36,852
8,167,589,516
IssuesEvent
2018-08-26 00:57:07
idaholab/raven
https://api.github.com/repos/idaholab/raven
opened
codecoverage is months old
defect devel master priority_critical
-------- Issue Description -------- ##### What did you expect to see happen? The code coverage script to be functional. The code coverage is months old since the check_coverage.sh is broken. This is causing the master branch to be red since a while. ##### What did you see instead? The code coverage script is currently broken ##### Do you have a suggested fix for the development team? ##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [ ] 1. Is it tagged with a type: defect or improvement? - [ ] 2. Is it tagged with a priority: critical, normal or minor? - [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)? - [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
1.0
codecoverage is months old - -------- Issue Description -------- ##### What did you expect to see happen? The code coverage script to be functional. The code coverage is months old since the check_coverage.sh is broken. This is causing the master branch to be red since a while. ##### What did you see instead? The code coverage script is currently broken ##### Do you have a suggested fix for the development team? ##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [ ] 1. Is it tagged with a type: defect or improvement? - [ ] 2. Is it tagged with a priority: critical, normal or minor? - [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)? - [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
defect
codecoverage is months old issue description what did you expect to see happen the code coverage script to be functional the code coverage is months old since the check coverage sh is broken this is causing the master branch to be red since a while what did you see instead the code coverage script is currently broken do you have a suggested fix for the development team please attach the input file s that generate this error the simpler the input the faster we can find the issue for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided
1
431
2,535,850,448
IssuesEvent
2015-01-26 08:28:05
localwiki/localwiki
https://api.github.com/repos/localwiki/localwiki
closed
django-taggit not quite working with versionutils
defect versioning
We're not using django-taggit, but it's worth noting that the following doesn't work right: ``` class Page(models.Model): ... tags = TaggableManager() versioning.register(Page) versioning.register(Tag) p_h = p.versions.most_recent() p_h.tags.all() # <-- this will always just display an empty set. ``` Everything works right for non-historical instances. And the Tag object gets versioned correctly.
1.0
django-taggit not quite working with versionutils - We're not using django-taggit, but it's worth noting that the following doesn't work right: ``` class Page(models.Model): ... tags = TaggableManager() versioning.register(Page) versioning.register(Tag) p_h = p.versions.most_recent() p_h.tags.all() # <-- this will always just display an empty set. ``` Everything works right for non-historical instances. And the Tag object gets versioned correctly.
defect
django taggit not quite working with versionutils we re not using django taggit but it s worth noting that the following doesn t work right class page models model tags taggablemanager versioning register page versioning register tag p h p versions most recent p h tags all this will always just display an empty set everything works right for non historical instances and the tag object gets versioned correctly
1
77,244
26,875,431,888
IssuesEvent
2023-02-05 00:38:46
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
BUG: numerical issues for cdf/ppf of the betaprime distribution
defect scipy.stats
### Describe your issue. The CDF of `stats.betaprime` loses precision if relatively small values of `b` are used. ``` from scipy import stats from scipy import special as sc a, b = 0.05, 0.1 x = 1e22 stats.betaprime.cdf(x, a, b) # 1.0 sc.betainc(a, b, x/(1.+x)) # 1.0, this is used as the CDF ``` The correct value is ~0.997881: ``` import mpmath as mp mp.dps = 25 a, b = mp.mpf('0.05'), mp.mpf('0.1') x = mp.mpf('1e22') y = mp.fdiv(x, mp.fadd(mp.mpf('1.0'), x, dps=25), dps=25) mp.betainc(a, b, 0.0, y, regularized=True) # mpf('0.99788115003523326') ``` To compute the correct value without increased precision, one needs to do the following (see https://dlmf.nist.gov/8.17, Eq 8.17.4): `stats.beta.sf(1/(1+x), b, a) # 0.9978811466052919` Similarly, the PPF added in #17562 will run into a similar issue for such values. If we continue the example for above: ``` a, b = 0.05, 0.1 x = 1e22 p = stats.beta.sf(1/(1+x), b, a) r = sc.betaincinv(a, b, p) r/(1-r) # 9007199254740991.0, this is the ppf value computed in 17562 r/(1-r) / x # 9.007199254740991e-07, the value is way too small ``` The fix is to solve `stats.beta.sf(1/(1+x), b, a)` for `x`: ` 1/stats.beta.isf(p, b, a) - 1 # 1.0000000000001254e+22 ` ### Reproducing Code Example ```python see above ``` ### Error message ```shell no error, just inaccurate output ``` ### SciPy/NumPy/Python version information 1.9.3 1.22.4 sys.version_info(major=3, minor=9, micro=7, releaselevel='final', serial=0)
1.0
BUG: numerical issues for cdf/ppf of the betaprime distribution - ### Describe your issue. The CDF of `stats.betaprime` loses precision if relatively small values of `b` are used. ``` from scipy import stats from scipy import special as sc a, b = 0.05, 0.1 x = 1e22 stats.betaprime.cdf(x, a, b) # 1.0 sc.betainc(a, b, x/(1.+x)) # 1.0, this is used as the CDF ``` The correct value is ~0.997881: ``` import mpmath as mp mp.dps = 25 a, b = mp.mpf('0.05'), mp.mpf('0.1') x = mp.mpf('1e22') y = mp.fdiv(x, mp.fadd(mp.mpf('1.0'), x, dps=25), dps=25) mp.betainc(a, b, 0.0, y, regularized=True) # mpf('0.99788115003523326') ``` To compute the correct value without increased precision, one needs to do the following (see https://dlmf.nist.gov/8.17, Eq 8.17.4): `stats.beta.sf(1/(1+x), b, a) # 0.9978811466052919` Similarly, the PPF added in #17562 will run into a similar issue for such values. If we continue the example for above: ``` a, b = 0.05, 0.1 x = 1e22 p = stats.beta.sf(1/(1+x), b, a) r = sc.betaincinv(a, b, p) r/(1-r) # 9007199254740991.0, this is the ppf value computed in 17562 r/(1-r) / x # 9.007199254740991e-07, the value is way too small ``` The fix is to solve `stats.beta.sf(1/(1+x), b, a)` for `x`: ` 1/stats.beta.isf(p, b, a) - 1 # 1.0000000000001254e+22 ` ### Reproducing Code Example ```python see above ``` ### Error message ```shell no error, just inaccurate output ``` ### SciPy/NumPy/Python version information 1.9.3 1.22.4 sys.version_info(major=3, minor=9, micro=7, releaselevel='final', serial=0)
defect
bug numerical issues for cdf ppf of the betaprime distribution describe your issue the cdf of stats betaprime loses precision if relatively small values of b are used from scipy import stats from scipy import special as sc a b x stats betaprime cdf x a b sc betainc a b x x this is used as the cdf the correct value is import mpmath as mp mp dps a b mp mpf mp mpf x mp mpf y mp fdiv x mp fadd mp mpf x dps dps mp betainc a b y regularized true mpf to compute the correct value without increased precision one needs to do the following see eq stats beta sf x b a similarly the ppf added in will run into a similar issue for such values if we continue the example for above a b x p stats beta sf x b a r sc betaincinv a b p r r this is the ppf value computed in r r x the value is way too small the fix is to solve stats beta sf x b a for x stats beta isf p b a reproducing code example python see above error message shell no error just inaccurate output scipy numpy python version information sys version info major minor micro releaselevel final serial
1
29,730
5,845,704,384
IssuesEvent
2017-05-10 14:45:08
OpenMS/OpenMS
https://api.github.com/repos/OpenMS/OpenMS
opened
Invalid mzIdentML created from idXML with missing proteins
defect
In idXML (or internally in OpenMS) it is possible to have peptides without protein associations (e.g. from the search engine or due to filtering). However, in mzIdentML this is no longer allowed from version 1.1.0 on - which is the current version that OpenMS writes. If peptides without associated proteins are present in an idXML file, converting the file to mzid with IDFileConverter will produce an invalid result, without any warning or error. Running "FileInfo -v" on the output file will give error messages like this: `Validation error in file '...' line ... column 34: element 'cvParam' is not allowed for content model '(PeptideEvidenceRef+,Fragmentation?,(cvParam|userParam))'` See also this comment and the following ones: https://github.com/OpenMS/OpenMS/pull/2528#issuecomment-299894675
1.0
Invalid mzIdentML created from idXML with missing proteins - In idXML (or internally in OpenMS) it is possible to have peptides without protein associations (e.g. from the search engine or due to filtering). However, in mzIdentML this is no longer allowed from version 1.1.0 on - which is the current version that OpenMS writes. If peptides without associated proteins are present in an idXML file, converting the file to mzid with IDFileConverter will produce an invalid result, without any warning or error. Running "FileInfo -v" on the output file will give error messages like this: `Validation error in file '...' line ... column 34: element 'cvParam' is not allowed for content model '(PeptideEvidenceRef+,Fragmentation?,(cvParam|userParam))'` See also this comment and the following ones: https://github.com/OpenMS/OpenMS/pull/2528#issuecomment-299894675
defect
invalid mzidentml created from idxml with missing proteins in idxml or internally in openms it is possible to have peptides without protein associations e g from the search engine or due to filtering however in mzidentml this is no longer allowed from version on which is the current version that openms writes if peptides without associated proteins are present in an idxml file converting the file to mzid with idfileconverter will produce an invalid result without any warning or error running fileinfo v on the output file will give error messages like this validation error in file line column element cvparam is not allowed for content model peptideevidenceref fragmentation cvparam userparam see also this comment and the following ones
1
63,049
17,358,250,464
IssuesEvent
2021-07-29 16:51:29
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
When 1:1 video call is full screen, the dialpad and "three dots" buttons don't work
A-VoIP P2 T-Defect X-Needs-Investigation
### Description When 1:1 video call is full screen, the dialpad and "three dots" buttons don't work ### Steps to reproduce - start a video call - go full screen (top right arrows) - notice the controls work, but the two buttons in the bottom corners don't work ### Version information <!-- IMPORTANT: please answer the following questions, to help us narrow down the problem --> Element Nightly version: 2021050501 macos
1.0
When 1:1 video call is full screen, the dialpad and "three dots" buttons don't work - ### Description When 1:1 video call is full screen, the dialpad and "three dots" buttons don't work ### Steps to reproduce - start a video call - go full screen (top right arrows) - notice the controls work, but the two buttons in the bottom corners don't work ### Version information <!-- IMPORTANT: please answer the following questions, to help us narrow down the problem --> Element Nightly version: 2021050501 macos
defect
when video call is full screen the dialpad and three dots buttons don t work description when video call is full screen the dialpad and three dots buttons don t work steps to reproduce start a video call go full screen top right arrows notice the controls work but the two buttons in the bottom corners don t work version information element nightly version macos
1
58,456
16,542,689,457
IssuesEvent
2021-05-27 18:59:50
idaholab/moose
https://api.github.com/repos/idaholab/moose
closed
UserObjectInterface::hasUserObject<T> does not check type
C: MOOSE P: normal T: defect
## Bug Description `UserObjectInterface::hasUserObject<T>` does not see if the base UO is of type T. ## Steps to Reproduce See the method. ## Impact Method does not perform the intended behavior.
1.0
UserObjectInterface::hasUserObject<T> does not check type - ## Bug Description `UserObjectInterface::hasUserObject<T>` does not see if the base UO is of type T. ## Steps to Reproduce See the method. ## Impact Method does not perform the intended behavior.
defect
userobjectinterface hasuserobject does not check type bug description userobjectinterface hasuserobject does not see if the base uo is of type t steps to reproduce see the method impact method does not perform the intended behavior
1
180,990
21,630,924,595
IssuesEvent
2022-05-05 09:37:14
Nivaskumark/packages_apps_settings_A10_r33_CVE-2020-0188
https://api.github.com/repos/Nivaskumark/packages_apps_settings_A10_r33_CVE-2020-0188
opened
CVE-2020-0394 (High) detected in Settingsandroid-10.0.0_r33
security vulnerability
## CVE-2020-0394 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Settingsandroid-10.0.0_r33</b></p></summary> <p> <p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Settings>https://android.googlesource.com/platform/packages/apps/Settings</a></p> <p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/packages_apps_settings_A10_r33_CVE-2020-0188/commit/f3df08e4562c757ffb3a076c5898906fdc1afde6">f3df08e4562c757ffb3a076c5898906fdc1afde6</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/settings/bluetooth/BluetoothPairingDialog.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In onCreate of BluetoothPairingDialog.java, there is a possible tapjacking vector due to an insecure default value. This could lead to local escalation of privilege and untrusted devices accessing contact lists with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-8.0 Android-8.1 Android-9 Android-10 Android-11Android ID: A-155648639 <p>Publish Date: 2020-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0394>CVE-2020-0394</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2020-09-01">https://source.android.com/security/bulletin/2020-09-01</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: android-8.0.0_r50,android-8.1.0_r80,android-9.0.0_r60</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-0394 (High) detected in Settingsandroid-10.0.0_r33 - ## CVE-2020-0394 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Settingsandroid-10.0.0_r33</b></p></summary> <p> <p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Settings>https://android.googlesource.com/platform/packages/apps/Settings</a></p> <p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/packages_apps_settings_A10_r33_CVE-2020-0188/commit/f3df08e4562c757ffb3a076c5898906fdc1afde6">f3df08e4562c757ffb3a076c5898906fdc1afde6</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/settings/bluetooth/BluetoothPairingDialog.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In onCreate of BluetoothPairingDialog.java, there is a possible tapjacking vector due to an insecure default value. This could lead to local escalation of privilege and untrusted devices accessing contact lists with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-8.0 Android-8.1 Android-9 Android-10 Android-11Android ID: A-155648639 <p>Publish Date: 2020-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0394>CVE-2020-0394</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2020-09-01">https://source.android.com/security/bulletin/2020-09-01</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: android-8.0.0_r50,android-8.1.0_r80,android-9.0.0_r60</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in settingsandroid cve high severity vulnerability vulnerable library settingsandroid library home page a href found in head commit a href found in base branch master vulnerable source files src com android settings bluetooth bluetoothpairingdialog java vulnerability details in oncreate of bluetoothpairingdialog java there is a possible tapjacking vector due to an insecure default value this could lead to local escalation of privilege and untrusted devices accessing contact lists with no additional execution privileges needed user interaction is needed for exploitation product androidversions android android android android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android android android step up your open source security game with whitesource
0
41,057
10,279,024,021
IssuesEvent
2019-08-25 19:18:40
ase379/gpprofile2017
https://api.github.com/repos/ase379/gpprofile2017
closed
Huge prf might crash under 64 Bit
defect
Using a huge prf file (> 2GB size) sometimes crashes. After a restart, the load works.
1.0
Huge prf might crash under 64 Bit - Using a huge prf file (> 2GB size) sometimes crashes. After a restart, the load works.
defect
huge prf might crash under bit using a huge prf file size sometimes crashes after a restart the load works
1
1,355
2,603,840,141
IssuesEvent
2015-02-24 18:14:24
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳包皮内长小水泡
auto-migrated Priority-Medium Type-Defect
``` 沈阳包皮内长小水泡〓沈陽軍區政治部醫院性病〓TEL:024-3102 3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位� ��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:12
1.0
沈阳包皮内长小水泡 - ``` 沈阳包皮内长小水泡〓沈陽軍區政治部醫院性病〓TEL:024-3102 3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位� ��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:12
defect
沈阳包皮内长小水泡 沈阳包皮内长小水泡〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位� �� 。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 original issue reported on code google com by gmail com on jun at
1
102,600
8,850,246,949
IssuesEvent
2019-01-08 12:43:13
Kademi/kademi-dev
https://api.github.com/repos/Kademi/kademi-dev
closed
Update Icons
Ready to Test - Dev
Some Kcom2 components are using Bootstrap Glyphicon, it should be converted to font awesome icon https://github.com/Kademi/projects/issues/925
1.0
Update Icons - Some Kcom2 components are using Bootstrap Glyphicon, it should be converted to font awesome icon https://github.com/Kademi/projects/issues/925
non_defect
update icons some components are using bootstrap glyphicon it should be converted to font awesome icon
0
626,006
19,783,643,178
IssuesEvent
2022-01-18 02:12:07
Baystation12/Baystation12
https://api.github.com/repos/Baystation12/Baystation12
closed
Shield generator not clear for user when it doesn't have power
Oversight :bug: Priority: Low BINGO!
#### Description of issue When the shield generator does not have a cable dot-knot under it, it won't receive power. However, rather than act unpowered or rather than prompting the user, it will attempt to start, immediately fail, and show "Recovering..." #### Difference between expected and actual behavior Actual: 1. Shield interface can be interacted with when unpowered. 2. Shield can be turned on when unpowered. 3. Shield instantly fails when unpowered. 4. Shield will appear to be damaged to the point of failure and turn off. Expected: 1. Shield will prompt user in redtext, show an interface error message that blocks the shield configuration, or show an "INSUFFICIENT POWER" state rather than a "RECOVERING" state in UI. #### Steps to reproduce Place shield generator on no wires, a straight wire, or a curved wire - anything without a wire dot in the center of the tile. #### Specific information for locating N/A #### Length of time in which bug has been known to occur Forever. #### Client version, Server revision & Game ID Client Version: 511 Server Revision: 326f3982b969cc893743f7dbf88f125bf1830860 - dev - Game ID: bQ2-dege Current map: SEV Torch #### Issue bing Please check whatever applies. More checkboxes checked increase your chances of the issue being looked at sooner. - [X] Issue could be reproduced at least once - [X] Issue could be reproduced by different players - [X] Issue could be reproduced in multiple rounds - [X] Issue happened in a recent (less than 7 days ago) round - [X] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
1.0
Shield generator not clear for user when it doesn't have power - #### Description of issue When the shield generator does not have a cable dot-knot under it, it won't receive power. However, rather than act unpowered or rather than prompting the user, it will attempt to start, immediately fail, and show "Recovering..." #### Difference between expected and actual behavior Actual: 1. Shield interface can be interacted with when unpowered. 2. Shield can be turned on when unpowered. 3. Shield instantly fails when unpowered. 4. Shield will appear to be damaged to the point of failure and turn off. Expected: 1. Shield will prompt user in redtext, show an interface error message that blocks the shield configuration, or show an "INSUFFICIENT POWER" state rather than a "RECOVERING" state in UI. #### Steps to reproduce Place shield generator on no wires, a straight wire, or a curved wire - anything without a wire dot in the center of the tile. #### Specific information for locating N/A #### Length of time in which bug has been known to occur Forever. #### Client version, Server revision & Game ID Client Version: 511 Server Revision: 326f3982b969cc893743f7dbf88f125bf1830860 - dev - Game ID: bQ2-dege Current map: SEV Torch #### Issue bing Please check whatever applies. More checkboxes checked increase your chances of the issue being looked at sooner. - [X] Issue could be reproduced at least once - [X] Issue could be reproduced by different players - [X] Issue could be reproduced in multiple rounds - [X] Issue happened in a recent (less than 7 days ago) round - [X] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
non_defect
shield generator not clear for user when it doesn t have power description of issue when the shield generator does not have a cable dot knot under it it won t receive power however rather than act unpowered or rather than prompting the user it will attempt to start immediately fail and show recovering difference between expected and actual behavior actual shield interface can be interacted with when unpowered shield can be turned on when unpowered shield instantly fails when unpowered shield will appear to be damaged to the point of failure and turn off expected shield will prompt user in redtext show an interface error message that blocks the shield configuration or show an insufficient power state rather than a recovering state in ui steps to reproduce place shield generator on no wires a straight wire or a curved wire anything without a wire dot in the center of the tile specific information for locating n a length of time in which bug has been known to occur forever client version server revision game id client version server revision dev game id dege current map sev torch issue bing please check whatever applies more checkboxes checked increase your chances of the issue being looked at sooner issue could be reproduced at least once issue could be reproduced by different players issue could be reproduced in multiple rounds issue happened in a recent less than days ago round
0
50,250
13,187,401,370
IssuesEvent
2020-08-13 03:17:59
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
tableio and generated nugen files (Trac #386)
Migrated from Trac defect other
Hi Alex, you asked me to write a ticket about my problem with tableio and generated nugen files. I understand that I am not supposed to submit tickets as user "icecube", but it seems as if I do not have the needed permissions. I apologize for any inconvenience in this context. You find a copy of the description I have sent to the dataclass mailing list below. I just have to add, that the suggestion of Jakob works. By simply filtering NO_INTERACTION events by using the neutrino_event_type of the I3MCTree I could avoid the problem. However there is one minor inconsistency, which I do not understand. As described there have been many generated MC files, which have been processed without any problem, although containing NO_INTERACTION files. Thanks for looking at it. Best Hans =============================== Hi dataclass, I am currently experiencing some troubles with Tableio (rootwriter) in the context of generated nugen files (5000 frames each). The problem does not occur when running on level2 processed files (i.e. IC79 datasets 6457, 6005). The generated datasets I tried: IC79: http://internal.icecube.wisc.edu/simulation/dataset/7747 IC80+DC6: http://internal.icecube.wisc.edu/simulation/dataset/2476 some files are processed by rootwriter smoothly, others produce the following error message: /home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost' /home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was never properly closed. This is a BUG! This message appears i.e. when running on file /data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000700.i3.gz but not on i.e. /data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000800.i3.gz Same story for dataset 7747: /home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino- generator/7747/00000-00999/nugen_nue_IC79.007747.000080.i3.gz /home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost' /home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was never properly closed. This is a BUG! whereas i.e. the files 90-99 work fine: /home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino- generator/7747/00000-00999/nugen_nue_IC79.007747.000099.i3.gz I3Tray finishing... writer: 50000 calls to physics 225.93s user 0.65s system You can have a look at my bare tableio scripts at /home/hmniederhausen/scripts/2476_rootwriter.py /home/hmniederhausen/scripts/7747_generated_rootwriter.py -f insert_integer_in_range(0,10) // -f 8 gives error, -f 9 works fine Am I doing something wrong? I am wondering what makes the difference between generated and level2 processed files when it comes down to tableio. However I need to run some root scripts on the generated 7747 to debug some problems. I would be glad if somebody has an idea. Thanks in advance. Cheers, Hans <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/386 , reported by hans.niederhausen and owned by olivas_</summary> <p> ```json { "status": "closed", "changetime": "2015-03-06T20:02:40", "description": "Hi Alex,\n\nyou asked me to write a ticket about my problem with tableio and generated nugen files. \n\nI understand that I am not supposed to submit tickets as user \"icecube\", but it seems as if I do not have the needed permissions. I apologize for any inconvenience in this context.\n\nYou find a copy of the description I have sent to the dataclass mailing list below. I just have to add, that the suggestion of Jakob works. By simply filtering NO_INTERACTION events by using the neutrino_event_type of the I3MCTree I could avoid the problem.\n\nHowever there is one minor inconsistency, which I do not understand. As described there have been many generated MC files, which have been processed without any problem, although containing NO_INTERACTION files.\n\nThanks for looking at it.\n\nBest\nHans\n\n===============================\n\nHi dataclass,\n\nI am currently experiencing some troubles with Tableio (rootwriter) in the context of generated nugen files (5000 frames each). The problem does not occur when running on level2 processed files (i.e. IC79 datasets 6457, 6005).\n\nThe generated datasets I tried:\n IC79: http://internal.icecube.wisc.edu/simulation/dataset/7747\n IC80+DC6: http://internal.icecube.wisc.edu/simulation/dataset/2476\n\nsome files are processed by rootwriter smoothly, others produce the following error message: \n\t/home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost'\n\t/home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was \tnever properly closed. This is a BUG! \n\nThis message appears i.e. when running on file \n\t/data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000700.i3.gz \n\nbut not on i.e.\n\t/data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000800.i3.gz \n\nSame story for dataset 7747:\n\t/home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino-\t\t\t \n\tgenerator/7747/00000-00999/nugen_nue_IC79.007747.000080.i3.gz\n\t/home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost'\n\t/home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was \tnever properly closed. This is a BUG!\n\nwhereas i.e. the files 90-99 work fine:\n\t/home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino-\t\n\tgenerator/7747/00000-00999/nugen_nue_IC79.007747.000099.i3.gz\n\tI3Tray finishing...\n writer: 50000 calls to physics 225.93s user 0.65s system\n\nYou can have a look at my bare tableio scripts at \n/home/hmniederhausen/scripts/2476_rootwriter.py \n/home/hmniederhausen/scripts/7747_generated_rootwriter.py -f insert_integer_in_range(0,10) // -f 8 gives error, -f 9 works fine\n\n\nAm I doing something wrong? I am wondering what makes the difference between generated and level2 processed files when it comes down to tableio. However I need to run some root scripts on the generated 7747 to debug some problems. I would be glad if somebody has an idea.\n\nThanks in advance.\n\nCheers,\nHans", "reporter": "hans.niederhausen", "cc": "", "resolution": "wontfix", "_ts": "1425672160951089", "component": "other", "summary": "tableio and generated nugen files", "priority": "minor", "keywords": "tableio generated", "time": "2012-03-28T02:06:12", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
tableio and generated nugen files (Trac #386) - Hi Alex, you asked me to write a ticket about my problem with tableio and generated nugen files. I understand that I am not supposed to submit tickets as user "icecube", but it seems as if I do not have the needed permissions. I apologize for any inconvenience in this context. You find a copy of the description I have sent to the dataclass mailing list below. I just have to add, that the suggestion of Jakob works. By simply filtering NO_INTERACTION events by using the neutrino_event_type of the I3MCTree I could avoid the problem. However there is one minor inconsistency, which I do not understand. As described there have been many generated MC files, which have been processed without any problem, although containing NO_INTERACTION files. Thanks for looking at it. Best Hans =============================== Hi dataclass, I am currently experiencing some troubles with Tableio (rootwriter) in the context of generated nugen files (5000 frames each). The problem does not occur when running on level2 processed files (i.e. IC79 datasets 6457, 6005). The generated datasets I tried: IC79: http://internal.icecube.wisc.edu/simulation/dataset/7747 IC80+DC6: http://internal.icecube.wisc.edu/simulation/dataset/2476 some files are processed by rootwriter smoothly, others produce the following error message: /home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost' /home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was never properly closed. This is a BUG! This message appears i.e. when running on file /data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000700.i3.gz but not on i.e. /data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000800.i3.gz Same story for dataset 7747: /home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino- generator/7747/00000-00999/nugen_nue_IC79.007747.000080.i3.gz /home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost' /home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was never properly closed. This is a BUG! whereas i.e. the files 90-99 work fine: /home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino- generator/7747/00000-00999/nugen_nue_IC79.007747.000099.i3.gz I3Tray finishing... writer: 50000 calls to physics 225.93s user 0.65s system You can have a look at my bare tableio scripts at /home/hmniederhausen/scripts/2476_rootwriter.py /home/hmniederhausen/scripts/7747_generated_rootwriter.py -f insert_integer_in_range(0,10) // -f 8 gives error, -f 9 works fine Am I doing something wrong? I am wondering what makes the difference between generated and level2 processed files when it comes down to tableio. However I need to run some root scripts on the generated 7747 to debug some problems. I would be glad if somebody has an idea. Thanks in advance. Cheers, Hans <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/386 , reported by hans.niederhausen and owned by olivas_</summary> <p> ```json { "status": "closed", "changetime": "2015-03-06T20:02:40", "description": "Hi Alex,\n\nyou asked me to write a ticket about my problem with tableio and generated nugen files. \n\nI understand that I am not supposed to submit tickets as user \"icecube\", but it seems as if I do not have the needed permissions. I apologize for any inconvenience in this context.\n\nYou find a copy of the description I have sent to the dataclass mailing list below. I just have to add, that the suggestion of Jakob works. By simply filtering NO_INTERACTION events by using the neutrino_event_type of the I3MCTree I could avoid the problem.\n\nHowever there is one minor inconsistency, which I do not understand. As described there have been many generated MC files, which have been processed without any problem, although containing NO_INTERACTION files.\n\nThanks for looking at it.\n\nBest\nHans\n\n===============================\n\nHi dataclass,\n\nI am currently experiencing some troubles with Tableio (rootwriter) in the context of generated nugen files (5000 frames each). The problem does not occur when running on level2 processed files (i.e. IC79 datasets 6457, 6005).\n\nThe generated datasets I tried:\n IC79: http://internal.icecube.wisc.edu/simulation/dataset/7747\n IC80+DC6: http://internal.icecube.wisc.edu/simulation/dataset/2476\n\nsome files are processed by rootwriter smoothly, others produce the following error message: \n\t/home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost'\n\t/home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was \tnever properly closed. This is a BUG! \n\nThis message appears i.e. when running on file \n\t/data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000700.i3.gz \n\nbut not on i.e.\n\t/data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000800.i3.gz \n\nSame story for dataset 7747:\n\t/home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino-\t\t\t \n\tgenerator/7747/00000-00999/nugen_nue_IC79.007747.000080.i3.gz\n\t/home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost'\n\t/home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was \tnever properly closed. This is a BUG!\n\nwhereas i.e. the files 90-99 work fine:\n\t/home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino-\t\n\tgenerator/7747/00000-00999/nugen_nue_IC79.007747.000099.i3.gz\n\tI3Tray finishing...\n writer: 50000 calls to physics 225.93s user 0.65s system\n\nYou can have a look at my bare tableio scripts at \n/home/hmniederhausen/scripts/2476_rootwriter.py \n/home/hmniederhausen/scripts/7747_generated_rootwriter.py -f insert_integer_in_range(0,10) // -f 8 gives error, -f 9 works fine\n\n\nAm I doing something wrong? I am wondering what makes the difference between generated and level2 processed files when it comes down to tableio. However I need to run some root scripts on the generated 7747 to debug some problems. I would be glad if somebody has an idea.\n\nThanks in advance.\n\nCheers,\nHans", "reporter": "hans.niederhausen", "cc": "", "resolution": "wontfix", "_ts": "1425672160951089", "component": "other", "summary": "tableio and generated nugen files", "priority": "minor", "keywords": "tableio generated", "time": "2012-03-28T02:06:12", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
tableio and generated nugen files trac hi alex you asked me to write a ticket about my problem with tableio and generated nugen files i understand that i am not supposed to submit tickets as user icecube but it seems as if i do not have the needed permissions i apologize for any inconvenience in this context you find a copy of the description i have sent to the dataclass mailing list below i just have to add that the suggestion of jakob works by simply filtering no interaction events by using the neutrino event type of the i could avoid the problem however there is one minor inconsistency which i do not understand as described there have been many generated mc files which have been processed without any problem although containing no interaction files thanks for looking at it best hans hi dataclass i am currently experiencing some troubles with tableio rootwriter in the context of generated nugen files frames each the problem does not occur when running on processed files i e datasets the generated datasets i tried some files are processed by rootwriter smoothly others produce the following error message home hmniederhausen icerec src tableio public tableio h fatal tried to set unknown column energylost home hmniederhausen icerec src tableio private tableio cxx fatal table is still connected which means that the output file was never properly closed this is a bug this message appears i e when running on file data sim icecube generated neutrino generator nugen nue gz but not on i e data sim icecube generated neutrino generator nugen nue gz same story for dataset home hmniederhausen icerec src dataio private dataio cxx info opened file data sim icecube generated neutrino generator nugen nue gz home hmniederhausen icerec src tableio public tableio h fatal tried to set unknown column energylost home hmniederhausen icerec src tableio private tableio cxx fatal table is still connected which means that the output file was never properly closed this is a bug whereas i e the files work fine home hmniederhausen icerec src dataio private dataio cxx info opened file data sim icecube generated neutrino generator nugen nue gz finishing writer calls to physics user system you can have a look at my bare tableio scripts at home hmniederhausen scripts rootwriter py home hmniederhausen scripts generated rootwriter py f insert integer in range f gives error f works fine am i doing something wrong i am wondering what makes the difference between generated and processed files when it comes down to tableio however i need to run some root scripts on the generated to debug some problems i would be glad if somebody has an idea thanks in advance cheers hans migrated from reported by hans niederhausen and owned by olivas json status closed changetime description hi alex n nyou asked me to write a ticket about my problem with tableio and generated nugen files n ni understand that i am not supposed to submit tickets as user icecube but it seems as if i do not have the needed permissions i apologize for any inconvenience in this context n nyou find a copy of the description i have sent to the dataclass mailing list below i just have to add that the suggestion of jakob works by simply filtering no interaction events by using the neutrino event type of the i could avoid the problem n nhowever there is one minor inconsistency which i do not understand as described there have been many generated mc files which have been processed without any problem although containing no interaction files n nthanks for looking at it n nbest nhans n n n nhi dataclass n ni am currently experiencing some troubles with tableio rootwriter in the context of generated nugen files frames each the problem does not occur when running on processed files i e datasets n nthe generated datasets i tried n files are processed by rootwriter smoothly others produce the following error message n t home hmniederhausen icerec src tableio public tableio h fatal tried to set unknown column energylost n t home hmniederhausen icerec src tableio private tableio cxx fatal table is still connected which means that the output file was tnever properly closed this is a bug n nthis message appears i e when running on file n t data sim icecube generated neutrino generator nugen nue gz n nbut not on i e n t data sim icecube generated neutrino generator nugen nue gz n nsame story for dataset n t home hmniederhausen icerec src dataio private dataio cxx info opened file data sim icecube generated neutrino t t t n tgenerator nugen nue gz n t home hmniederhausen icerec src tableio public tableio h fatal tried to set unknown column energylost n t home hmniederhausen icerec src tableio private tableio cxx fatal table is still connected which means that the output file was tnever properly closed this is a bug n nwhereas i e the files work fine n t home hmniederhausen icerec src dataio private dataio cxx info opened file data sim icecube generated neutrino t n tgenerator nugen nue gz n finishing n writer calls to physics user system n nyou can have a look at my bare tableio scripts at n home hmniederhausen scripts rootwriter py n home hmniederhausen scripts generated rootwriter py f insert integer in range f gives error f works fine n n nam i doing something wrong i am wondering what makes the difference between generated and processed files when it comes down to tableio however i need to run some root scripts on the generated to debug some problems i would be glad if somebody has an idea n nthanks in advance n ncheers nhans reporter hans niederhausen cc resolution wontfix ts component other summary tableio and generated nugen files priority minor keywords tableio generated time milestone owner olivas type defect
1
40,570
10,054,793,349
IssuesEvent
2019-07-22 03:19:26
line/armeria
https://api.github.com/repos/line/armeria
closed
`RequestLogAvailabilityException: REQUEST_FIRST_BYTES_TRANSFERRED` when `RequestLogAvailability` is `COMPLETE`
defect
Reported by @adriancole: ``` WARN 88577 --- [orageComponent}] c.l.a.c.l.RequestLogListenerInvoker : onRequestLog() failed with an exception:com.linecorp.armeria.common.logging.RequestLogAvailabilityException: REQUEST_FIRST_BYTES_TRANSFERRED at com.linecorp.armeria.common.logging.RequestLog.ensureAvailability(RequestLog.java:120) ~[armeria-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.DefaultRequestLog.requestFirstBytesTransferredTimeNanos(DefaultRequestLog.java:446) ~[armeria-0.88.0.jar!/:?] at com.linecorp.armeria.client.brave.BraveClient.lambda$execute$2(BraveClient.java:126) ~[armeria-brave-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.RequestLogListenerInvoker.invokeOnRequestLog(RequestLogListenerInvoker.java:38) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.DefaultRequestLog.notifyListeners(DefaultRequestLog.java:1036) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.DefaultRequestLog.updateAvailability(DefaultRequestLog.java:993) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.DefaultRequestLog.endResponse0(DefaultRequestLog.java:970) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.DefaultRequestLog.endResponse0(DefaultRequestLog.java:949) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.DefaultRequestLog.endResponse(DefaultRequestLog.java:935) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.internal.ClientUtil.executeWithFallback(ClientUtil.java:50) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.client.UserClient.execute(UserClient.java:159) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.client.DefaultHttpClient.execute(DefaultHttpClient.java:61) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.client.DefaultHttpClient.execute(DefaultHttpClient.java:78) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.client.DefaultHttpClient.execute(DefaultHttpClient.java:74) [armeria-0.88.0.jar!/:?] at zipkin2.elasticsearch.internal.client.HttpCall.sendRequest(HttpCall.java:145) [zipkin-storage-elasticsearch-2.15.1-SNAPSHOT.jar!/:?] at zipkin2.elasticsearch.internal.client.HttpCall.doExecute(HttpCall.java:101) [zipkin-storage-elasticsearch-2.15.1-SNAPSHOT.jar!/:?] at zipkin2.Call$Base.execute(Call.java:379) [zipkin-2.15.1-SNAPSHOT.jar!/:?] at zipkin2.reporter.AsyncReporter$BoundedAsyncReporter.flush(AsyncReporter.java:286) [zipkin-reporter-2.8.4.jar!/:?] at zipkin2.reporter.AsyncReporter$Builder$1.run(AsyncReporter.java:190) [zipkin-reporter-2.8.4.jar!/:?] ``` In BraveClient.java: ```java ctx.log().addListener(log -> { SpanTags.logWireSend(span, log.requestFirstBytesTransferredTimeNanos(), log); // If the client timed-out the request, we will have never received any response data at all. if (log.isAvailable(RequestLogAvailability.RESPONSE_FIRST_BYTES_TRANSFERRED)) { SpanTags.logWireReceive(span, log.responseFirstBytesTransferredTimeNanos(), log); } finishSpan(span, log); }, RequestLogAvailability.COMPLETE); ``` /cc @anuraaga
1.0
`RequestLogAvailabilityException: REQUEST_FIRST_BYTES_TRANSFERRED` when `RequestLogAvailability` is `COMPLETE` - Reported by @adriancole: ``` WARN 88577 --- [orageComponent}] c.l.a.c.l.RequestLogListenerInvoker : onRequestLog() failed with an exception:com.linecorp.armeria.common.logging.RequestLogAvailabilityException: REQUEST_FIRST_BYTES_TRANSFERRED at com.linecorp.armeria.common.logging.RequestLog.ensureAvailability(RequestLog.java:120) ~[armeria-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.DefaultRequestLog.requestFirstBytesTransferredTimeNanos(DefaultRequestLog.java:446) ~[armeria-0.88.0.jar!/:?] at com.linecorp.armeria.client.brave.BraveClient.lambda$execute$2(BraveClient.java:126) ~[armeria-brave-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.RequestLogListenerInvoker.invokeOnRequestLog(RequestLogListenerInvoker.java:38) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.DefaultRequestLog.notifyListeners(DefaultRequestLog.java:1036) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.DefaultRequestLog.updateAvailability(DefaultRequestLog.java:993) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.DefaultRequestLog.endResponse0(DefaultRequestLog.java:970) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.DefaultRequestLog.endResponse0(DefaultRequestLog.java:949) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.common.logging.DefaultRequestLog.endResponse(DefaultRequestLog.java:935) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.internal.ClientUtil.executeWithFallback(ClientUtil.java:50) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.client.UserClient.execute(UserClient.java:159) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.client.DefaultHttpClient.execute(DefaultHttpClient.java:61) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.client.DefaultHttpClient.execute(DefaultHttpClient.java:78) [armeria-0.88.0.jar!/:?] at com.linecorp.armeria.client.DefaultHttpClient.execute(DefaultHttpClient.java:74) [armeria-0.88.0.jar!/:?] at zipkin2.elasticsearch.internal.client.HttpCall.sendRequest(HttpCall.java:145) [zipkin-storage-elasticsearch-2.15.1-SNAPSHOT.jar!/:?] at zipkin2.elasticsearch.internal.client.HttpCall.doExecute(HttpCall.java:101) [zipkin-storage-elasticsearch-2.15.1-SNAPSHOT.jar!/:?] at zipkin2.Call$Base.execute(Call.java:379) [zipkin-2.15.1-SNAPSHOT.jar!/:?] at zipkin2.reporter.AsyncReporter$BoundedAsyncReporter.flush(AsyncReporter.java:286) [zipkin-reporter-2.8.4.jar!/:?] at zipkin2.reporter.AsyncReporter$Builder$1.run(AsyncReporter.java:190) [zipkin-reporter-2.8.4.jar!/:?] ``` In BraveClient.java: ```java ctx.log().addListener(log -> { SpanTags.logWireSend(span, log.requestFirstBytesTransferredTimeNanos(), log); // If the client timed-out the request, we will have never received any response data at all. if (log.isAvailable(RequestLogAvailability.RESPONSE_FIRST_BYTES_TRANSFERRED)) { SpanTags.logWireReceive(span, log.responseFirstBytesTransferredTimeNanos(), log); } finishSpan(span, log); }, RequestLogAvailability.COMPLETE); ``` /cc @anuraaga
defect
requestlogavailabilityexception request first bytes transferred when requestlogavailability is complete reported by adriancole warn c l a c l requestloglistenerinvoker onrequestlog failed with an exception com linecorp armeria common logging requestlogavailabilityexception request first bytes transferred at com linecorp armeria common logging requestlog ensureavailability requestlog java at com linecorp armeria common logging defaultrequestlog requestfirstbytestransferredtimenanos defaultrequestlog java at com linecorp armeria client brave braveclient lambda execute braveclient java at com linecorp armeria common logging requestloglistenerinvoker invokeonrequestlog requestloglistenerinvoker java at com linecorp armeria common logging defaultrequestlog notifylisteners defaultrequestlog java at com linecorp armeria common logging defaultrequestlog updateavailability defaultrequestlog java at com linecorp armeria common logging defaultrequestlog defaultrequestlog java at com linecorp armeria common logging defaultrequestlog defaultrequestlog java at com linecorp armeria common logging defaultrequestlog endresponse defaultrequestlog java at com linecorp armeria internal clientutil executewithfallback clientutil java at com linecorp armeria client userclient execute userclient java at com linecorp armeria client defaulthttpclient execute defaulthttpclient java at com linecorp armeria client defaulthttpclient execute defaulthttpclient java at com linecorp armeria client defaulthttpclient execute defaulthttpclient java at elasticsearch internal client httpcall sendrequest httpcall java at elasticsearch internal client httpcall doexecute httpcall java at call base execute call java at reporter asyncreporter boundedasyncreporter flush asyncreporter java at reporter asyncreporter builder run asyncreporter java in braveclient java java ctx log addlistener log spantags logwiresend span log requestfirstbytestransferredtimenanos log if the client timed out the request we will have never received any response data at all if log isavailable requestlogavailability response first bytes transferred spantags logwirereceive span log responsefirstbytestransferredtimenanos log finishspan span log requestlogavailability complete cc anuraaga
1
70,967
23,391,993,353
IssuesEvent
2022-08-11 18:47:54
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
App Layout: Bottom Sheet
T-Defect X-Needs-Design Z-AppLayout
### Steps to reproduce We had several bugs and issues raised during internal testing regarding the Spaces bottom sheet. ### Outcome #### What did you expect? #### What happened instead? ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store App Layout APK ### Homeserver _No response_ ### Will you send logs? No ### Are you willing to provide a PR? No
1.0
App Layout: Bottom Sheet - ### Steps to reproduce We had several bugs and issues raised during internal testing regarding the Spaces bottom sheet. ### Outcome #### What did you expect? #### What happened instead? ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store App Layout APK ### Homeserver _No response_ ### Will you send logs? No ### Are you willing to provide a PR? No
defect
app layout bottom sheet steps to reproduce we had several bugs and issues raised during internal testing regarding the spaces bottom sheet outcome what did you expect what happened instead your phone model no response operating system version no response application version and app store app layout apk homeserver no response will you send logs no are you willing to provide a pr no
1
65,423
7,878,740,311
IssuesEvent
2018-06-26 11:14:45
openebs/openebs
https://api.github.com/repos/openebs/openebs
closed
Document the steps to run GRPC application
contributor-docs kind/design summerhack
We need to create a document on how we're going to place the GRPC related stubs in `maya` project. It will be helpful for community.
1.0
Document the steps to run GRPC application - We need to create a document on how we're going to place the GRPC related stubs in `maya` project. It will be helpful for community.
non_defect
document the steps to run grpc application we need to create a document on how we re going to place the grpc related stubs in maya project it will be helpful for community
0
52,472
13,224,768,742
IssuesEvent
2020-08-17 19:48:33
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
dataio - more ncurses issues blocking "dataio-shovel" (Trac #2268)
Incomplete Migration Migrated from Trac analysis defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2268">https://code.icecube.wisc.edu/projects/icecube/ticket/2268</a>, reported by negaand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-09-04T12:46:55", "_ts": "1567601215767531", "description": "After updating `ncurses` detection, building `dataio-shovel` dies with:\n\n{{{\n[ 62%] Building CXX object dataio/CMakeFiles/dataio-shovel.dir/private/shovel/View.cxx.o\n/build/buildslave/chernobog/SuSE/source/dataio/private/shovel/View.cxx: In function \u2018void resize(int)\u2019:\n/build/buildslave/chernobog/SuSE/source/dataio/private/shovel/View.cxx:99:25: error: lvalue required as left operand of assignment\n LINES = win.ws_row;\n ^~~~~~\n/build/buildslave/chernobog/SuSE/source/dataio/private/shovel/View.cxx:102:24: error: lvalue required as left operand of assignment\n COLS = win.ws_col;\n ^~~~~~\nmake[2]: *** [dataio/CMakeFiles/dataio-shovel.dir/build.make:111: dataio/CMakeFiles/dataio-shovel.dir/private/shovel/View.cxx.o] Error 1\nmake[1]: *** [CMakeFiles/Makefile2:5679: dataio/CMakeFiles/dataio-shovel.dir/all] Error 2\nmake: *** [Makefile:141: all] Error 2\n}}}\n\nMaybe a `TIOCGSIZE`/`TIOCGWINSZ` issue. \n\nSUSE is such a super minor OS for us, kicking it down the road. ", "reporter": "nega", "cc": "", "resolution": "worksforme", "time": "2019-04-03T18:55:54", "component": "analysis", "summary": "dataio - more ncurses issues blocking \"dataio-shovel\"", "priority": "normal", "keywords": "dataio dataio-shovel curses ncurses janitors", "milestone": "Autumnal Equinox 2019", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
dataio - more ncurses issues blocking "dataio-shovel" (Trac #2268) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2268">https://code.icecube.wisc.edu/projects/icecube/ticket/2268</a>, reported by negaand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-09-04T12:46:55", "_ts": "1567601215767531", "description": "After updating `ncurses` detection, building `dataio-shovel` dies with:\n\n{{{\n[ 62%] Building CXX object dataio/CMakeFiles/dataio-shovel.dir/private/shovel/View.cxx.o\n/build/buildslave/chernobog/SuSE/source/dataio/private/shovel/View.cxx: In function \u2018void resize(int)\u2019:\n/build/buildslave/chernobog/SuSE/source/dataio/private/shovel/View.cxx:99:25: error: lvalue required as left operand of assignment\n LINES = win.ws_row;\n ^~~~~~\n/build/buildslave/chernobog/SuSE/source/dataio/private/shovel/View.cxx:102:24: error: lvalue required as left operand of assignment\n COLS = win.ws_col;\n ^~~~~~\nmake[2]: *** [dataio/CMakeFiles/dataio-shovel.dir/build.make:111: dataio/CMakeFiles/dataio-shovel.dir/private/shovel/View.cxx.o] Error 1\nmake[1]: *** [CMakeFiles/Makefile2:5679: dataio/CMakeFiles/dataio-shovel.dir/all] Error 2\nmake: *** [Makefile:141: all] Error 2\n}}}\n\nMaybe a `TIOCGSIZE`/`TIOCGWINSZ` issue. \n\nSUSE is such a super minor OS for us, kicking it down the road. ", "reporter": "nega", "cc": "", "resolution": "worksforme", "time": "2019-04-03T18:55:54", "component": "analysis", "summary": "dataio - more ncurses issues blocking \"dataio-shovel\"", "priority": "normal", "keywords": "dataio dataio-shovel curses ncurses janitors", "milestone": "Autumnal Equinox 2019", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
dataio more ncurses issues blocking dataio shovel trac migrated from json status closed changetime ts description after updating ncurses detection building dataio shovel dies with n n n building cxx object dataio cmakefiles dataio shovel dir private shovel view cxx o n build buildslave chernobog suse source dataio private shovel view cxx in function resize int n build buildslave chernobog suse source dataio private shovel view cxx error lvalue required as left operand of assignment n lines win ws row n n build buildslave chernobog suse source dataio private shovel view cxx error lvalue required as left operand of assignment n cols win ws col n nmake error nmake error nmake error n n nmaybe a tiocgsize tiocgwinsz issue n nsuse is such a super minor os for us kicking it down the road reporter nega cc resolution worksforme time component analysis summary dataio more ncurses issues blocking dataio shovel priority normal keywords dataio dataio shovel curses ncurses janitors milestone autumnal equinox owner nega type defect
1
42,736
11,224,112,035
IssuesEvent
2020-01-08 01:03:21
codepotent/Update-Manager
https://api.github.com/repos/codepotent/Update-Manager
closed
Footer version number is inconsistent with current version
Status: Completed Type: Defect
# Description The version noted in the footer of the Update Manager plugin is '1.0.0' while the actual version is '1.0.0-rc2'. Not critical, but, not accurate.
1.0
Footer version number is inconsistent with current version - # Description The version noted in the footer of the Update Manager plugin is '1.0.0' while the actual version is '1.0.0-rc2'. Not critical, but, not accurate.
defect
footer version number is inconsistent with current version description the version noted in the footer of the update manager plugin is while the actual version is not critical but not accurate
1
48,653
13,184,711,378
IssuesEvent
2020-08-12 19:57:17
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
dartboard on icecode (Trac #10)
Incomplete Migration Migrated from Trac defect infrastructure
<details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/10 , reported by troy and owned by cgils_</summary> <p> ```json { "status": "closed", "changetime": "2007-07-06T14:14:06", "description": "needs startup scripts, bring up autobuild boxes.\n\n", "reporter": "troy", "cc": "", "resolution": "fixed", "_ts": "1183731246000000", "component": "infrastructure", "summary": "dartboard on icecode", "priority": "normal", "keywords": "", "time": "2007-06-03T16:29:58", "milestone": "", "owner": "cgils", "type": "defect" } ``` </p> </details>
1.0
dartboard on icecode (Trac #10) - <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/10 , reported by troy and owned by cgils_</summary> <p> ```json { "status": "closed", "changetime": "2007-07-06T14:14:06", "description": "needs startup scripts, bring up autobuild boxes.\n\n", "reporter": "troy", "cc": "", "resolution": "fixed", "_ts": "1183731246000000", "component": "infrastructure", "summary": "dartboard on icecode", "priority": "normal", "keywords": "", "time": "2007-06-03T16:29:58", "milestone": "", "owner": "cgils", "type": "defect" } ``` </p> </details>
defect
dartboard on icecode trac migrated from reported by troy and owned by cgils json status closed changetime description needs startup scripts bring up autobuild boxes n n reporter troy cc resolution fixed ts component infrastructure summary dartboard on icecode priority normal keywords time milestone owner cgils type defect
1
72,110
23,947,553,230
IssuesEvent
2022-09-12 08:43:13
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
I did a thread and it went wrong for Travis (maybe other folks too?)
T-Defect S-Minor O-Uncommon A-Threads
### Steps to reproduce 1. Where are you starting? What can you see? I made a thread and typed lots of messages in it ### Outcome The thread looked good for me, but for Travis it looked like a giant pile of replies. ### Operating system Windows 10 ### Application version Element Nightly version: 0.0.1-nightly.2022090801 Olm version: 3.2.12 ### How did you install the app? website ### Homeserver lant.uk ### Will you send logs? Yes
1.0
I did a thread and it went wrong for Travis (maybe other folks too?) - ### Steps to reproduce 1. Where are you starting? What can you see? I made a thread and typed lots of messages in it ### Outcome The thread looked good for me, but for Travis it looked like a giant pile of replies. ### Operating system Windows 10 ### Application version Element Nightly version: 0.0.1-nightly.2022090801 Olm version: 3.2.12 ### How did you install the app? website ### Homeserver lant.uk ### Will you send logs? Yes
defect
i did a thread and it went wrong for travis maybe other folks too steps to reproduce where are you starting what can you see i made a thread and typed lots of messages in it outcome the thread looked good for me but for travis it looked like a giant pile of replies operating system windows application version element nightly version nightly olm version how did you install the app website homeserver lant uk will you send logs yes
1
36,085
4,713,602,782
IssuesEvent
2016-10-14 20:40:07
WordPress/twentyseventeen
https://api.github.com/repos/WordPress/twentyseventeen
closed
Explore viewport units for responsive typography
design enhancement question
Placeholder issue for now. I'll come back later and explain what I'm thinking.
1.0
Explore viewport units for responsive typography - Placeholder issue for now. I'll come back later and explain what I'm thinking.
non_defect
explore viewport units for responsive typography placeholder issue for now i ll come back later and explain what i m thinking
0
8,539
2,611,516,986
IssuesEvent
2015-02-27 05:51:50
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
hang while connecting to custom server via hwplay scheme if you have empty nickname
auto-migrated Priority-Low Type-Defect
``` What steps will reproduce the problem? 1. Associate Hedgewars file extension on Linux 2. Click on (eg) hwplay://m8y.org 3. Wait indefinitely What is the expected output? What do you see instead? Instead of hanging in the 'connecting' page it should connect to server! Same code, only happens on Linux (why?). It's also possible to cause this by calling $(hedgewars hwplay://m8y.org) from the command line ``` Original issue reported on code.google.com by `vittorio...@gmail.com` on 25 Mar 2013 at 11:21 * Blocking: #580
1.0
hang while connecting to custom server via hwplay scheme if you have empty nickname - ``` What steps will reproduce the problem? 1. Associate Hedgewars file extension on Linux 2. Click on (eg) hwplay://m8y.org 3. Wait indefinitely What is the expected output? What do you see instead? Instead of hanging in the 'connecting' page it should connect to server! Same code, only happens on Linux (why?). It's also possible to cause this by calling $(hedgewars hwplay://m8y.org) from the command line ``` Original issue reported on code.google.com by `vittorio...@gmail.com` on 25 Mar 2013 at 11:21 * Blocking: #580
defect
hang while connecting to custom server via hwplay scheme if you have empty nickname what steps will reproduce the problem associate hedgewars file extension on linux click on eg hwplay org wait indefinitely what is the expected output what do you see instead instead of hanging in the connecting page it should connect to server same code only happens on linux why it s also possible to cause this by calling hedgewars hwplay org from the command line original issue reported on code google com by vittorio gmail com on mar at blocking
1
51,217
13,207,396,112
IssuesEvent
2020-08-14 22:56:44
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
I3DOMCalibration class needs a suck-ectomy (Trac #54)
Incomplete Migration Migrated from Trac defect offline-software
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/54">https://code.icecube.wisc.edu/projects/icecube/ticket/54</a>, reported by blaufussand owned by blaufuss</em></summary> <p> ```json { "status": "closed", "changetime": "2007-11-11T03:51:18", "_ts": "1194753078000000", "description": "There's gotta be some performance to be gained by redoing the internals of\nthis class.\n\n typedef map<unsigned, map<unsigned, LinearFit> > ATWDBinParam_t;\n\nCan't be the optimal way.", "reporter": "blaufuss", "cc": "", "resolution": "duplicate", "time": "2007-06-08T17:16:45", "component": "offline-software", "summary": "I3DOMCalibration class needs a suck-ectomy", "priority": "normal", "keywords": "", "milestone": "", "owner": "blaufuss", "type": "defect" } ``` </p> </details>
1.0
I3DOMCalibration class needs a suck-ectomy (Trac #54) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/54">https://code.icecube.wisc.edu/projects/icecube/ticket/54</a>, reported by blaufussand owned by blaufuss</em></summary> <p> ```json { "status": "closed", "changetime": "2007-11-11T03:51:18", "_ts": "1194753078000000", "description": "There's gotta be some performance to be gained by redoing the internals of\nthis class.\n\n typedef map<unsigned, map<unsigned, LinearFit> > ATWDBinParam_t;\n\nCan't be the optimal way.", "reporter": "blaufuss", "cc": "", "resolution": "duplicate", "time": "2007-06-08T17:16:45", "component": "offline-software", "summary": "I3DOMCalibration class needs a suck-ectomy", "priority": "normal", "keywords": "", "milestone": "", "owner": "blaufuss", "type": "defect" } ``` </p> </details>
defect
class needs a suck ectomy trac migrated from json status closed changetime ts description there s gotta be some performance to be gained by redoing the internals of nthis class n n typedef map atwdbinparam t n ncan t be the optimal way reporter blaufuss cc resolution duplicate time component offline software summary class needs a suck ectomy priority normal keywords milestone owner blaufuss type defect
1
525,516
15,255,262,986
IssuesEvent
2021-02-20 15:24:58
google/oboe
https://api.github.com/repos/google/oboe
closed
LiveEffect sample will crash when earphone plug is connected after plug-out earphone
bug high priority
Android version(s): 8.0.0 Android device(s): Garaxy S8 Oboe version: 1.5.1 (1.5.0 ) App name used for testing: LiveEffect sample **Short description** Try to disconnect and reconnect earphone plug, on Garaxy S8, app will crash after reconnect earphone plug. **Steps to reproduce** Start Engine → Plug-out earphone → wait few seconds → plug-in earphone without tapping stop engine button →Crash happens **Expected behavior** not crash **Actual behavior** crash error has occured ``` 2021-02-17 00:09:27.246 23407-23419/com.google.oboe.samples.liveeffect A/libc: Fatal signal 11 (SIGSEGV), code 1, fault addr 0xffffffffffffffe8 in tid 23419 (Binder:23407_1) 2021-02-17 00:09:27.387 23534-23534/? A/DEBUG: *** *** *** *** *** *** *** *** *** *** *** *** *** *** *** *** 2021-02-17 00:09:27.387 23534-23534/? A/DEBUG: Build fingerprint: 'samsung/dreamltexx/dreamlte:8.0.0/R16NW/G950FXXU4CRK1:user/release-keys' 2021-02-17 00:09:27.387 23534-23534/? A/DEBUG: Revision: '10' 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: ABI: 'arm64' 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: pid: 23407, tid: 23419, name: Binder:23407_1 >>> com.google.oboe.samples.liveeffect <<< 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: signal 11 (SIGSEGV), code 1 (SEGV_MAPERR), fault addr 0xffffffffffffffe8 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x0 00000076778891c0 x1 00000076778cf580 x2 0000000000000001 x3 0000000000000000 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x4 0000000000000000 x5 0000000000000000 x6 0000000000000000 x7 000000765c24ee38 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x8 00000076590c0508 x9 0000000000000000 x10 0000000000000008 x11 0000000000000000 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x12 0000000000000000 x13 0000000000000004 x14 000000000000001e x15 0000000000000008 2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x16 00000076780776e0 x17 000000767bc9fa70 x18 0000000000000020 x19 00000076778cf6c0 2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x20 0000000000000001 x21 0000000000000000 x22 0000000000000000 x23 000000767bc9c3c0 2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x24 000000765c24f570 x25 000000765c153000 x26 00000076650770c0 x27 0000000000000000 2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x28 0000000000000000 x29 000000765c24f340 x30 000000767802c204 2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: sp 000000765c24f320 pc 000000767802c220 pstate 0000000060000000 2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: backtrace: 2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: #00 pc 0000000000063220 /system/lib64/libbinder.so (_ZN7android14IPCThreadState20processPendingDerefsEv+128) 2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: #01 pc 0000000000063294 /system/lib64/libbinder.so (_ZN7android14IPCThreadState14joinThreadPoolEb+52) 2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: #02 pc 0000000000085408 /system/lib64/libbinder.so (_ZN7android10PoolThread10threadLoopEv+24) 2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #03 pc 0000000000011be8 /system/lib64/libutils.so (_ZN7android6Thread11_threadLoopEPv+280) 2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #04 pc 00000000000fd6c8 /system/lib64/libandroid_runtime.so (_ZN7android14AndroidRuntime15javaThreadShellEPv+136) 2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #05 pc 000000000006de3c /system/lib64/libc.so (_ZL15__pthread_startPv+220) 2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #06 pc 000000000001f9c4 /system/lib64/libc.so (__start_thread+68) ``` **Device** **Any additional context** On Pixel3, crash doesn't happen, but we can't hear audio playback until restart oboe engine. ( we need to stop , and restart engine to enable audio playback after plug-out earphone when engine running). I don't have any idea to prevent this crash on c++ code...
1.0
LiveEffect sample will crash when earphone plug is connected after plug-out earphone - Android version(s): 8.0.0 Android device(s): Garaxy S8 Oboe version: 1.5.1 (1.5.0 ) App name used for testing: LiveEffect sample **Short description** Try to disconnect and reconnect earphone plug, on Garaxy S8, app will crash after reconnect earphone plug. **Steps to reproduce** Start Engine → Plug-out earphone → wait few seconds → plug-in earphone without tapping stop engine button →Crash happens **Expected behavior** not crash **Actual behavior** crash error has occured ``` 2021-02-17 00:09:27.246 23407-23419/com.google.oboe.samples.liveeffect A/libc: Fatal signal 11 (SIGSEGV), code 1, fault addr 0xffffffffffffffe8 in tid 23419 (Binder:23407_1) 2021-02-17 00:09:27.387 23534-23534/? A/DEBUG: *** *** *** *** *** *** *** *** *** *** *** *** *** *** *** *** 2021-02-17 00:09:27.387 23534-23534/? A/DEBUG: Build fingerprint: 'samsung/dreamltexx/dreamlte:8.0.0/R16NW/G950FXXU4CRK1:user/release-keys' 2021-02-17 00:09:27.387 23534-23534/? A/DEBUG: Revision: '10' 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: ABI: 'arm64' 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: pid: 23407, tid: 23419, name: Binder:23407_1 >>> com.google.oboe.samples.liveeffect <<< 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: signal 11 (SIGSEGV), code 1 (SEGV_MAPERR), fault addr 0xffffffffffffffe8 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x0 00000076778891c0 x1 00000076778cf580 x2 0000000000000001 x3 0000000000000000 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x4 0000000000000000 x5 0000000000000000 x6 0000000000000000 x7 000000765c24ee38 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x8 00000076590c0508 x9 0000000000000000 x10 0000000000000008 x11 0000000000000000 2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x12 0000000000000000 x13 0000000000000004 x14 000000000000001e x15 0000000000000008 2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x16 00000076780776e0 x17 000000767bc9fa70 x18 0000000000000020 x19 00000076778cf6c0 2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x20 0000000000000001 x21 0000000000000000 x22 0000000000000000 x23 000000767bc9c3c0 2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x24 000000765c24f570 x25 000000765c153000 x26 00000076650770c0 x27 0000000000000000 2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x28 0000000000000000 x29 000000765c24f340 x30 000000767802c204 2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: sp 000000765c24f320 pc 000000767802c220 pstate 0000000060000000 2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: backtrace: 2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: #00 pc 0000000000063220 /system/lib64/libbinder.so (_ZN7android14IPCThreadState20processPendingDerefsEv+128) 2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: #01 pc 0000000000063294 /system/lib64/libbinder.so (_ZN7android14IPCThreadState14joinThreadPoolEb+52) 2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: #02 pc 0000000000085408 /system/lib64/libbinder.so (_ZN7android10PoolThread10threadLoopEv+24) 2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #03 pc 0000000000011be8 /system/lib64/libutils.so (_ZN7android6Thread11_threadLoopEPv+280) 2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #04 pc 00000000000fd6c8 /system/lib64/libandroid_runtime.so (_ZN7android14AndroidRuntime15javaThreadShellEPv+136) 2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #05 pc 000000000006de3c /system/lib64/libc.so (_ZL15__pthread_startPv+220) 2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #06 pc 000000000001f9c4 /system/lib64/libc.so (__start_thread+68) ``` **Device** **Any additional context** On Pixel3, crash doesn't happen, but we can't hear audio playback until restart oboe engine. ( we need to stop , and restart engine to enable audio playback after plug-out earphone when engine running). I don't have any idea to prevent this crash on c++ code...
non_defect
liveeffect sample will crash when earphone plug is connected after plug out earphone android version s android device s garaxy oboe version app name used for testing liveeffect sample short description try to disconnect and reconnect earphone plug on garaxy app will crash after reconnect earphone plug steps to reproduce start engine → plug out earphone → wait few seconds → plug in earphone without tapping stop engine button →crash happens expected behavior not crash actual behavior crash error has occured com google oboe samples liveeffect a libc fatal signal sigsegv code fault addr in tid binder a debug a debug build fingerprint samsung dreamltexx dreamlte user release keys a debug revision a debug abi a debug pid tid name binder com google oboe samples liveeffect a debug signal sigsegv code segv maperr fault addr a debug a debug a debug a debug a debug a debug a debug a debug a debug sp pc pstate a debug backtrace a debug pc system libbinder so a debug pc system libbinder so a debug pc system libbinder so a debug pc system libutils so threadloopepv a debug pc system libandroid runtime so a debug pc system libc so pthread startpv a debug pc system libc so start thread device any additional context on crash doesn t happen but we can t hear audio playback until restart oboe engine we need to stop and restart engine to enable audio playback after plug out earphone when engine running i don t have any idea to prevent this crash on c code
0
47,006
13,056,013,581
IssuesEvent
2020-07-30 03:23:32
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
[icetray] I3TestModuleFactory doesn't work in python3 (Trac #2256)
Incomplete Migration Migrated from Trac combo core defect
Migrated from https://code.icecube.wisc.edu/ticket/2256 ```json { "status": "closed", "changetime": "2020-06-24T12:31:42", "description": "It looks like the `unittest` works differently in python 2 and 3. In 2 you can keep running the same unit test over and over again, but in 3 after you run it once it gets deleted and becomes a `None`\n{{{\n$ python CoincSuite/resources/test/AfterpulseTest.py \nNOTICE (I3Tray): I3Tray finishing... (I3Tray.cxx:497 in void I3Tray::Execute(unsigned int))\nNOTICE (TesterModule): \nRecombination Decisions by AfterpulseTester:\n YES : 1\n NO : 0\n UNDECIDED : 0\n (TesterModule.cxx:45 in virtual void TesterModule::Finish())\nCoincSuite/resources/test/AfterpulseTest.py:142: DeprecationWarning: Please use assertTrue instead.\n self.assert_(self.frame.Has(SplitName+\"RecombAttempts\"))\nCoincSuite/resources/test/AfterpulseTest.py:144: DeprecationWarning: Please use assertTrue instead.\n self.assert_(ra[0] == \"AfterpulseTester\")\n.\n----------------------------------------------------------------------\nRan 1 test in 0.001s\n\nOK\nCoincSuite/resources/test/AfterpulseTest.py:156: DeprecationWarning: Please use assertTrue instead.\n self.assert_(eh.sub_event_stream==\"split\" and eh.sub_event_id==0)\n.\n----------------------------------------------------------------------\nRan 1 test in 0.000s\n\nOK\nERROR (I3Module): TestPhysicsSequence: Exception thrown (I3Module.cxx:127 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"CoincSuite/resources/test/AfterpulseTest.py\", line 194, in <module>\n tray.Execute()\n File \"/Users/kmeagher/icecube/combo/release/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\n File \"/Users/kmeagher/icecube/combo/release/lib/icecube/icetray/test_module.py\", line 83, in RunTests\n test_result = self.test_runner.run(self.test_suites)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/runner.py\", line 176, in run\n test(result)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py\", line 84, in __call__\n return self.run(*args, **kwds)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py\", line 122, in run\n test(result)\nTypeError: 'NoneType' object is not callable\n\n}}}\n", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "_ts": "1593001902142004", "component": "combo core", "summary": "[icetray] I3TestModuleFactory doesn't work in python3", "priority": "blocker", "keywords": "", "time": "2019-03-18T19:19:29", "milestone": "Autumnal Equinox 2020", "owner": "olivas", "type": "defect" } ```
1.0
[icetray] I3TestModuleFactory doesn't work in python3 (Trac #2256) - Migrated from https://code.icecube.wisc.edu/ticket/2256 ```json { "status": "closed", "changetime": "2020-06-24T12:31:42", "description": "It looks like the `unittest` works differently in python 2 and 3. In 2 you can keep running the same unit test over and over again, but in 3 after you run it once it gets deleted and becomes a `None`\n{{{\n$ python CoincSuite/resources/test/AfterpulseTest.py \nNOTICE (I3Tray): I3Tray finishing... (I3Tray.cxx:497 in void I3Tray::Execute(unsigned int))\nNOTICE (TesterModule): \nRecombination Decisions by AfterpulseTester:\n YES : 1\n NO : 0\n UNDECIDED : 0\n (TesterModule.cxx:45 in virtual void TesterModule::Finish())\nCoincSuite/resources/test/AfterpulseTest.py:142: DeprecationWarning: Please use assertTrue instead.\n self.assert_(self.frame.Has(SplitName+\"RecombAttempts\"))\nCoincSuite/resources/test/AfterpulseTest.py:144: DeprecationWarning: Please use assertTrue instead.\n self.assert_(ra[0] == \"AfterpulseTester\")\n.\n----------------------------------------------------------------------\nRan 1 test in 0.001s\n\nOK\nCoincSuite/resources/test/AfterpulseTest.py:156: DeprecationWarning: Please use assertTrue instead.\n self.assert_(eh.sub_event_stream==\"split\" and eh.sub_event_id==0)\n.\n----------------------------------------------------------------------\nRan 1 test in 0.000s\n\nOK\nERROR (I3Module): TestPhysicsSequence: Exception thrown (I3Module.cxx:127 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"CoincSuite/resources/test/AfterpulseTest.py\", line 194, in <module>\n tray.Execute()\n File \"/Users/kmeagher/icecube/combo/release/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\n File \"/Users/kmeagher/icecube/combo/release/lib/icecube/icetray/test_module.py\", line 83, in RunTests\n test_result = self.test_runner.run(self.test_suites)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/runner.py\", line 176, in run\n test(result)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py\", line 84, in __call__\n return self.run(*args, **kwds)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py\", line 122, in run\n test(result)\nTypeError: 'NoneType' object is not callable\n\n}}}\n", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "_ts": "1593001902142004", "component": "combo core", "summary": "[icetray] I3TestModuleFactory doesn't work in python3", "priority": "blocker", "keywords": "", "time": "2019-03-18T19:19:29", "milestone": "Autumnal Equinox 2020", "owner": "olivas", "type": "defect" } ```
defect
doesn t work in trac migrated from json status closed changetime description it looks like the unittest works differently in python and in you can keep running the same unit test over and over again but in after you run it once it gets deleted and becomes a none n n python coincsuite resources test afterpulsetest py nnotice finishing cxx in void execute unsigned int nnotice testermodule nrecombination decisions by afterpulsetester n yes n no n undecided n testermodule cxx in virtual void testermodule finish ncoincsuite resources test afterpulsetest py deprecationwarning please use asserttrue instead n self assert self frame has splitname recombattempts ncoincsuite resources test afterpulsetest py deprecationwarning please use asserttrue instead n self assert ra afterpulsetester n n nran test in n nok ncoincsuite resources test afterpulsetest py deprecationwarning please use asserttrue instead n self assert eh sub event stream split and eh sub event id n n nran test in n nok nerror testphysicssequence exception thrown cxx in void do void ntraceback most recent call last n file coincsuite resources test afterpulsetest py line in n tray execute n file users kmeagher icecube combo release lib py line in execute n super self execute n file users kmeagher icecube combo release lib icecube icetray test module py line in runtests n test result self test runner run self test suites n file usr local cellar python frameworks python framework versions lib unittest runner py line in run n test result n file usr local cellar python frameworks python framework versions lib unittest suite py line in call n return self run args kwds n file usr local cellar python frameworks python framework versions lib unittest suite py line in run n test result ntypeerror nonetype object is not callable n n n reporter kjmeagher cc resolution fixed ts component combo core summary doesn t work in priority blocker keywords time milestone autumnal equinox owner olivas type defect
1
28,843
5,390,079,462
IssuesEvent
2017-02-25 10:10:57
patmun/pynetdicom
https://api.github.com/repos/patmun/pynetdicom
closed
Missing self
auto-migrated Priority-Medium Type-Defect
``` https://code.google.com/p/pynetdicom/source/browse/source/netdicom/SOPclass.py?s pec=svn2ebb51548127bf4cc6de867e550f44102e7c4cbf&r=2ebb51548127bf4cc6de867e550f44 102e7c4cbf#351 should be self.CannotUnderstand ``` Original issue reported on code.google.com by `agrothberg` on 22 Apr 2015 at 8:06
1.0
Missing self - ``` https://code.google.com/p/pynetdicom/source/browse/source/netdicom/SOPclass.py?s pec=svn2ebb51548127bf4cc6de867e550f44102e7c4cbf&r=2ebb51548127bf4cc6de867e550f44 102e7c4cbf#351 should be self.CannotUnderstand ``` Original issue reported on code.google.com by `agrothberg` on 22 Apr 2015 at 8:06
defect
missing self pec r should be self cannotunderstand original issue reported on code google com by agrothberg on apr at
1
11,797
13,911,122,977
IssuesEvent
2020-10-20 16:57:37
elementor/elementor
https://api.github.com/repos/elementor/elementor
closed
Advanced Editor Tools (previously TinyMCE Advanced) fails to load in Elementor
compatibility status/merged
https://github.com/elementor/elementor/blob/bb96e7a2cde78e1000997319accb2bc193c48fa4/core/editor/editor.php#L584 When the _Advanced Editor Tools (previously TinyMCE Advanced)_ plugin is activated and a text widget is selected in Elementor, the Elementor panel is empty. The author of the plugin has tracked down the issue but says that it is something that needs to be fixed in Elementor. He provides the solution here: https://wordpress.org/support/topic/editor-fails-to-load-in-elementor/ A temporary solution is to uncheck the "Other instances of the Classic (TinyMCE) editor in wp-admin" option in the plugin's settings (Note: I don't know anything about GitHub so I hope I've done this right!) <img width="1084" alt="elementor-panel" src="https://user-images.githubusercontent.com/26119379/95342674-98282980-08af-11eb-9d13-7e7dc04ef21c.png">
True
Advanced Editor Tools (previously TinyMCE Advanced) fails to load in Elementor - https://github.com/elementor/elementor/blob/bb96e7a2cde78e1000997319accb2bc193c48fa4/core/editor/editor.php#L584 When the _Advanced Editor Tools (previously TinyMCE Advanced)_ plugin is activated and a text widget is selected in Elementor, the Elementor panel is empty. The author of the plugin has tracked down the issue but says that it is something that needs to be fixed in Elementor. He provides the solution here: https://wordpress.org/support/topic/editor-fails-to-load-in-elementor/ A temporary solution is to uncheck the "Other instances of the Classic (TinyMCE) editor in wp-admin" option in the plugin's settings (Note: I don't know anything about GitHub so I hope I've done this right!) <img width="1084" alt="elementor-panel" src="https://user-images.githubusercontent.com/26119379/95342674-98282980-08af-11eb-9d13-7e7dc04ef21c.png">
non_defect
advanced editor tools previously tinymce advanced fails to load in elementor when the advanced editor tools previously tinymce advanced plugin is activated and a text widget is selected in elementor the elementor panel is empty the author of the plugin has tracked down the issue but says that it is something that needs to be fixed in elementor he provides the solution here a temporary solution is to uncheck the other instances of the classic tinymce editor in wp admin option in the plugin s settings note i don t know anything about github so i hope i ve done this right img width alt elementor panel src
0
77,196
26,834,736,287
IssuesEvent
2023-02-02 18:33:38
Software-Hardware-Codesign/jme-alloc
https://api.github.com/repos/Software-Hardware-Codesign/jme-alloc
closed
[Core-Bug] Writing on a buffer after destruction leads to a jvm monitor crash
invalid core defect
Writing on a buffer after destroying it will lead to a jvm crash with the following logs, despite the ability to print the buffer capacity, limit and the current position. The jvm crash log: ``` --------------- P R O C E S S --------------- VM state: at safepoint (shutting down) VM Mutex/Monitor currently owned by a thread: ([mutex/lock_event]) [0x00007ff5180117d0] Threads_lock - owner thread: 0x00007ff518125c60 Heap address: 0x0000000746e00000, size: 2962 MB, Compressed Oops mode: Zero based, Oop shift amount: 3 CDS archive(s) mapped at: [0x0000000800000000-0x0000000800be3000-0x0000000800be3000), size 12464128, SharedBaseAddress: 0x0000000800000000, ArchiveRelocationMode: 0. Compressed class space mapped at: 0x0000000800c00000-0x0000000840c00000, reserved size: 1073741824 Narrow klass base: 0x0000000800000000, Narrow klass shift: 0, Narrow klass range: 0x100000000 GC Precious Log: <Skipped> Heap: garbage-first heap total 194560K, used 5065K [0x0000000746e00000, 0x0000000800000000) region size 2048K, 2 young (4096K), 0 survivors (0K) Metaspace used 841K, committed 1024K, reserved 1056768K class space used 78K, committed 192K, reserved 1048576K ``` A workaround is to nullify the `java.nio.ByteBuffer` reference internally from jni via the invocation api after destroying its memory; because this memory is now owned by another allocated object, so accessing it using this reference and writing on it might be dangerous if it is not thread protected.
1.0
[Core-Bug] Writing on a buffer after destruction leads to a jvm monitor crash - Writing on a buffer after destroying it will lead to a jvm crash with the following logs, despite the ability to print the buffer capacity, limit and the current position. The jvm crash log: ``` --------------- P R O C E S S --------------- VM state: at safepoint (shutting down) VM Mutex/Monitor currently owned by a thread: ([mutex/lock_event]) [0x00007ff5180117d0] Threads_lock - owner thread: 0x00007ff518125c60 Heap address: 0x0000000746e00000, size: 2962 MB, Compressed Oops mode: Zero based, Oop shift amount: 3 CDS archive(s) mapped at: [0x0000000800000000-0x0000000800be3000-0x0000000800be3000), size 12464128, SharedBaseAddress: 0x0000000800000000, ArchiveRelocationMode: 0. Compressed class space mapped at: 0x0000000800c00000-0x0000000840c00000, reserved size: 1073741824 Narrow klass base: 0x0000000800000000, Narrow klass shift: 0, Narrow klass range: 0x100000000 GC Precious Log: <Skipped> Heap: garbage-first heap total 194560K, used 5065K [0x0000000746e00000, 0x0000000800000000) region size 2048K, 2 young (4096K), 0 survivors (0K) Metaspace used 841K, committed 1024K, reserved 1056768K class space used 78K, committed 192K, reserved 1048576K ``` A workaround is to nullify the `java.nio.ByteBuffer` reference internally from jni via the invocation api after destroying its memory; because this memory is now owned by another allocated object, so accessing it using this reference and writing on it might be dangerous if it is not thread protected.
defect
writing on a buffer after destruction leads to a jvm monitor crash writing on a buffer after destroying it will lead to a jvm crash with the following logs despite the ability to print the buffer capacity limit and the current position the jvm crash log p r o c e s s vm state at safepoint shutting down vm mutex monitor currently owned by a thread threads lock owner thread heap address size mb compressed oops mode zero based oop shift amount cds archive s mapped at size sharedbaseaddress archiverelocationmode compressed class space mapped at reserved size narrow klass base narrow klass shift narrow klass range gc precious log heap garbage first heap total used region size young survivors metaspace used committed reserved class space used committed reserved a workaround is to nullify the java nio bytebuffer reference internally from jni via the invocation api after destroying its memory because this memory is now owned by another allocated object so accessing it using this reference and writing on it might be dangerous if it is not thread protected
1
1,851
2,603,972,471
IssuesEvent
2015-02-24 19:00:36
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳病毒疣是否传染
auto-migrated Priority-Medium Type-Defect
``` 沈阳病毒疣是否传染〓沈陽軍區政治部醫院性病〓TEL:024-3102 3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位� ��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:00
1.0
沈阳病毒疣是否传染 - ``` 沈阳病毒疣是否传染〓沈陽軍區政治部醫院性病〓TEL:024-3102 3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位� ��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:00
defect
沈阳病毒疣是否传染 沈阳病毒疣是否传染〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位� �� 。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 original issue reported on code google com by gmail com on jun at
1
409,343
27,734,355,434
IssuesEvent
2023-03-15 10:12:27
pires/go-proxyproto
https://api.github.com/repos/pires/go-proxyproto
closed
Set-up automated releases
help wanted documentation good first issue
Rely on Github Actions, such as https://github.com/softprops/action-gh-release, to automatically create releases based on tags.
1.0
Set-up automated releases - Rely on Github Actions, such as https://github.com/softprops/action-gh-release, to automatically create releases based on tags.
non_defect
set up automated releases rely on github actions such as to automatically create releases based on tags
0