Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
181,389
6,659,273,214
IssuesEvent
2017-10-01 09:03:53
OperationCode/operationcode_frontend
https://api.github.com/repos/OperationCode/operationcode_frontend
opened
Broken link in CONTRIBUTING.md
beginner friendly hacktoberfest Priority: Medium Status: Available
# Bug Report ## What is the current behavior? `CONTRIBUTING.md` contains: `- [Running the development Server](#running-the-development-server)` on line 29. ## What is the expected behavior? It should contain: - [Running the Development Environment](#running-the-development-environment)
1.0
Broken link in CONTRIBUTING.md - # Bug Report ## What is the current behavior? `CONTRIBUTING.md` contains: `- [Running the development Server](#running-the-development-server)` on line 29. ## What is the expected behavior? It should contain: - [Running the Development Environment](#running-the-development-environment)
non_defect
broken link in contributing md bug report what is the current behavior contributing md contains running the development server on line what is the expected behavior it should contain running the development environment
0
167
2,640,878,704
IssuesEvent
2015-03-11 14:56:06
ploneintranet/ploneintranet.theme
https://api.github.com/repos/ploneintranet/ploneintranet.theme
closed
Scroll-bars on texareas new-post/comment/repost areas in IE
browser compatibility in progress
New post area: ![screen shot 2015-03-11 at 14 16 07](https://cloud.githubusercontent.com/assets/1485452/6598160/9eb66200-c7f9-11e4-8830-bc9ebb091e50.png) Comment areas in the stream: ![screen shot 2015-03-11 at 14 16 57](https://cloud.githubusercontent.com/assets/1485452/6598166/a3ae6b4a-c7f9-11e4-8b4f-685adeb7eaa0.png) Repost text area ![screen shot 2015-03-11 at 09 12 25](https://cloud.githubusercontent.com/assets/1485452/6598173/af5cbc62-c7f9-11e4-8846-f0984c0bd545.png)
True
Scroll-bars on texareas new-post/comment/repost areas in IE - New post area: ![screen shot 2015-03-11 at 14 16 07](https://cloud.githubusercontent.com/assets/1485452/6598160/9eb66200-c7f9-11e4-8830-bc9ebb091e50.png) Comment areas in the stream: ![screen shot 2015-03-11 at 14 16 57](https://cloud.githubusercontent.com/assets/1485452/6598166/a3ae6b4a-c7f9-11e4-8b4f-685adeb7eaa0.png) Repost text area ![screen shot 2015-03-11 at 09 12 25](https://cloud.githubusercontent.com/assets/1485452/6598173/af5cbc62-c7f9-11e4-8846-f0984c0bd545.png)
non_defect
scroll bars on texareas new post comment repost areas in ie new post area comment areas in the stream repost text area
0
115,286
4,662,351,494
IssuesEvent
2016-10-05 03:14:00
communityshare/communityshare
https://api.github.com/repos/communityshare/communityshare
closed
Alert text shown briefly on page load
[Effort] Good first issues [Priority] High [Skills] Angular [Type] Bug [Type] UX
When loading the app home page, there is an alert that briefly shows before the rest of the HTML is rendered. It's empty and looks like a bug when viewing the page. This alert should not appear. Likely resolved inside of `static/index.html` ![alerttext](https://cloud.githubusercontent.com/assets/5431237/17349160/78dce4b4-58d0-11e6-8157-8d387012ce1d.gif) This may have been introduced in in [this commit](https://github.com/communityshare/communityshare/commit/d085fb091cdfeaeffd682c73906f9bc9f8e4a68e).
1.0
Alert text shown briefly on page load - When loading the app home page, there is an alert that briefly shows before the rest of the HTML is rendered. It's empty and looks like a bug when viewing the page. This alert should not appear. Likely resolved inside of `static/index.html` ![alerttext](https://cloud.githubusercontent.com/assets/5431237/17349160/78dce4b4-58d0-11e6-8157-8d387012ce1d.gif) This may have been introduced in in [this commit](https://github.com/communityshare/communityshare/commit/d085fb091cdfeaeffd682c73906f9bc9f8e4a68e).
non_defect
alert text shown briefly on page load when loading the app home page there is an alert that briefly shows before the rest of the html is rendered it s empty and looks like a bug when viewing the page this alert should not appear likely resolved inside of static index html this may have been introduced in in
0
171,937
13,254,437,731
IssuesEvent
2020-08-20 09:15:41
google/go-safeweb
https://api.github.com/repos/google/go-safeweb
opened
The test harness for testing net/http parsing is very slow, again
invalid testing
As of Go 1.15 #59 has been become a problem again. The solution we found in #63 doesn't work. But at least we know what is causing the slowdown.`net/http.Server.Shutdown`s polling interval for checking if all connections are finished is 0.5 s and we don't have time to shut everything down before the first sleep starts. Therefore we get punished by 0.5 s every time although closing our connections takes way less than 0.5 s.
1.0
The test harness for testing net/http parsing is very slow, again - As of Go 1.15 #59 has been become a problem again. The solution we found in #63 doesn't work. But at least we know what is causing the slowdown.`net/http.Server.Shutdown`s polling interval for checking if all connections are finished is 0.5 s and we don't have time to shut everything down before the first sleep starts. Therefore we get punished by 0.5 s every time although closing our connections takes way less than 0.5 s.
non_defect
the test harness for testing net http parsing is very slow again as of go has been become a problem again the solution we found in doesn t work but at least we know what is causing the slowdown net http server shutdown s polling interval for checking if all connections are finished is s and we don t have time to shut everything down before the first sleep starts therefore we get punished by s every time although closing our connections takes way less than s
0
304,430
23,065,639,033
IssuesEvent
2022-07-25 13:43:06
zeroback-song/zero-service
https://api.github.com/repos/zeroback-song/zero-service
closed
제로백 rest api 정의
documentation
제로백 rest api 정의 * [x] 힙합 rest api 정의 * [x] 발라드 rest api 정의 * [x] 팝송 rest api 정의 * [x] 재즈 rest api 정의
1.0
제로백 rest api 정의 - 제로백 rest api 정의 * [x] 힙합 rest api 정의 * [x] 발라드 rest api 정의 * [x] 팝송 rest api 정의 * [x] 재즈 rest api 정의
non_defect
제로백 rest api 정의 제로백 rest api 정의 힙합 rest api 정의 발라드 rest api 정의 팝송 rest api 정의 재즈 rest api 정의
0
2,985
3,275,549,269
IssuesEvent
2015-10-26 15:58:01
CellProfiler/CellProfiler
https://api.github.com/repos/CellProfiler/CellProfiler
opened
Saving Preferences should warn user to restart Cellprofiler
Feature request Layout and usability
Python/CellProfiler menu -> Preferences don't get applied until the next restart, no? Or maybe for some settings, like 'Max memory for Java'? The user should be warned about this (e.g. a modal dialog after Save that says "Please restart CellProfiler for these settings to be applied")
True
Saving Preferences should warn user to restart Cellprofiler - Python/CellProfiler menu -> Preferences don't get applied until the next restart, no? Or maybe for some settings, like 'Max memory for Java'? The user should be warned about this (e.g. a modal dialog after Save that says "Please restart CellProfiler for these settings to be applied")
non_defect
saving preferences should warn user to restart cellprofiler python cellprofiler menu preferences don t get applied until the next restart no or maybe for some settings like max memory for java the user should be warned about this e g a modal dialog after save that says please restart cellprofiler for these settings to be applied
0
19,810
10,532,671,798
IssuesEvent
2019-10-01 11:19:28
woocommerce/woocommerce-gutenberg-products-block
https://api.github.com/repos/woocommerce/woocommerce-gutenberg-products-block
opened
Performance of Block Settings on every page load without caching
type: performance
This was reported in https://github.com/woocommerce/woocommerce/issues/24590 Blocks requires certain settings/data to function, and this needs to be present anywhere a block may be used (so admin and frontend, regardless of user type). This is output to JSON inline on every page load. This can be problematic if doing large queries such as getting terms, and counting post types. Currently, the JSON is generated in Assets.php here: https://github.com/woocommerce/woocommerce-gutenberg-products-block/blob/master/src/Assets.php#L86-L123 Most of this is harmless, with the exception of: 1. get_terms - this is querying all categories in the store and generating links for each. 2. wp_count_posts - less intensive but could still be cached. We should look at either caching this data, using APIs instead as needed instead of grabbing data that may not be used, or making some of this conditonal.
True
Performance of Block Settings on every page load without caching - This was reported in https://github.com/woocommerce/woocommerce/issues/24590 Blocks requires certain settings/data to function, and this needs to be present anywhere a block may be used (so admin and frontend, regardless of user type). This is output to JSON inline on every page load. This can be problematic if doing large queries such as getting terms, and counting post types. Currently, the JSON is generated in Assets.php here: https://github.com/woocommerce/woocommerce-gutenberg-products-block/blob/master/src/Assets.php#L86-L123 Most of this is harmless, with the exception of: 1. get_terms - this is querying all categories in the store and generating links for each. 2. wp_count_posts - less intensive but could still be cached. We should look at either caching this data, using APIs instead as needed instead of grabbing data that may not be used, or making some of this conditonal.
non_defect
performance of block settings on every page load without caching this was reported in blocks requires certain settings data to function and this needs to be present anywhere a block may be used so admin and frontend regardless of user type this is output to json inline on every page load this can be problematic if doing large queries such as getting terms and counting post types currently the json is generated in assets php here most of this is harmless with the exception of get terms this is querying all categories in the store and generating links for each wp count posts less intensive but could still be cached we should look at either caching this data using apis instead as needed instead of grabbing data that may not be used or making some of this conditonal
0
278,588
24,162,102,620
IssuesEvent
2022-09-22 12:30:15
Tencent/bk-job
https://api.github.com/repos/Tencent/bk-job
closed
bugfix: 已上线的脚本,为啥还提示未保存的内容
kind/bug stage/test accepted
**Version / Branch / tag** ≥3.5.1 **出了什么问题?(What Happened?)** ![image](https://user-images.githubusercontent.com/3852595/190993406-d4520f95-cbb9-4087-bd3d-f3bf2bc669fa.png) 已上线的脚本,为啥还提示未保存的内容,操作方式是 新建版本,选了一个版本,保存,上线,再点右上角的叉时,提示的这个 **如何复现?(How to reproduce?)** 新建版本,选了一个版本,保存,上线,再点右上角的叉时,提示的这个 **预期结果(What you expect?)** 上线后,因为已经是只读状态而非编辑状态了,这时离开不应该还要提示未保存信息的弹窗
1.0
bugfix: 已上线的脚本,为啥还提示未保存的内容 - **Version / Branch / tag** ≥3.5.1 **出了什么问题?(What Happened?)** ![image](https://user-images.githubusercontent.com/3852595/190993406-d4520f95-cbb9-4087-bd3d-f3bf2bc669fa.png) 已上线的脚本,为啥还提示未保存的内容,操作方式是 新建版本,选了一个版本,保存,上线,再点右上角的叉时,提示的这个 **如何复现?(How to reproduce?)** 新建版本,选了一个版本,保存,上线,再点右上角的叉时,提示的这个 **预期结果(What you expect?)** 上线后,因为已经是只读状态而非编辑状态了,这时离开不应该还要提示未保存信息的弹窗
non_defect
bugfix 已上线的脚本,为啥还提示未保存的内容 version branch tag ≥ 出了什么问题? what happened 已上线的脚本,为啥还提示未保存的内容,操作方式是 新建版本,选了一个版本,保存,上线,再点右上角的叉时,提示的这个 如何复现? how to reproduce 新建版本,选了一个版本,保存,上线,再点右上角的叉时,提示的这个 预期结果 what you expect 上线后,因为已经是只读状态而非编辑状态了,这时离开不应该还要提示未保存信息的弹窗
0
655
2,823,166,139
IssuesEvent
2015-05-21 06:53:04
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Consider open sourcing ApiCompat.
enhancement Infrastructure
Internally we have some tooling to ensure that implementation assemblies are compatible with contracts. It would be a good idea to open source that and use it as part of the build or CI system or something. I'm not sure if this will happen naturally when we actually start generating contracts in the open or if we'll need to take extra steps.
1.0
Consider open sourcing ApiCompat. - Internally we have some tooling to ensure that implementation assemblies are compatible with contracts. It would be a good idea to open source that and use it as part of the build or CI system or something. I'm not sure if this will happen naturally when we actually start generating contracts in the open or if we'll need to take extra steps.
non_defect
consider open sourcing apicompat internally we have some tooling to ensure that implementation assemblies are compatible with contracts it would be a good idea to open source that and use it as part of the build or ci system or something i m not sure if this will happen naturally when we actually start generating contracts in the open or if we ll need to take extra steps
0
93,501
3,901,182,294
IssuesEvent
2016-04-18 09:46:27
natsys/tempesta
https://api.github.com/repos/natsys/tempesta
opened
Kernel warning when calling __alloc_skb from the soft-IRQ
enhancement low priority
Kernel warning was caught by me on the laptop: ~~~ [ 3.129833] ------------[ cut here ]------------ [ 3.129839] WARNING: CPU: 1 PID: 0 at kernel/softirq.c:150 __local_bh_enable_ip+0x72/0xa0() [ 3.129841] Modules linked in: media btusb btbcm btintel bluetooth cdc_mbim intel_rapl iosf_mbi x86_pkg_temp_thermal intel_powerclamp coretemp cdc_ncm usbnet kvm_intel mii cdc_wdm cdc_acm arc4 kvm iwldvm iTCO_wdt iTCO_vendor_support crct10dif_pclmul crc32_pclmul crc32c_intel mac80211 aesni_intel aes_x86_64 iwlwifi lrw gf128mul glue_helper snd_hda_codec_hdmi ablk_helper cryptd psmouse serio_raw pcspkr cfg80211 snd_hda_codec_realtek snd_hda_codec_generic snd_hda_intel thermal snd_hda_controller snd_hda_codec wmi snd_hda_core snd_hwdep thinkpad_acpi snd_pcm snd_timer nvram ac rfkill tpm_tis mei_me tpm battery snd mei evdev i2c_i801 lpc_ich soundcore shpchp mac_hid processor sch_fq_codel vboxnetadp(O) vboxdrv(O) ip_tables x_tables ext4 crc16 mbcache jbd2 sd_mod atkbd libps2 sdhci_pci ahci libahci libata [ 3.129883] scsi_mod xhci_pci ehci_pci sdhci ehci_hcd xhci_hcd led_class mmc_core usbcore usb_common i8042 serio i915 button intel_gtt i2c_algo_bit video drm_kms_helper drm [ 3.129894] CPU: 1 PID: 0 Comm: swapper/1 Tainted: G O 4.1.21-1-tfw #1 [ 3.129897] Hardware name: LENOVO 34442KU/34442KU, BIOS G6ETB1WW (2.71 ) 04/23/2015 [ 3.129899] 0000000000000086 3f00b632d218473c ffff88011e243c98 ffffffff81588da2 [ 3.129902] 0000000000000000 ffffffff817129f0 ffff88011e243cd8 ffffffff8107be4a [ 3.129904] ffff88011e2107e0 0000000000000200 0000000000000020 000077ff80000000 [ 3.129907] Call Trace: [ 3.129909] <IRQ> [<ffffffff81588da2>] dump_stack+0x63/0x81 [ 3.129915] [<ffffffff8107be4a>] warn_slowpath_common+0x8a/0xc0 [ 3.129918] [<ffffffff8107bf7a>] warn_slowpath_null+0x1a/0x20 [ 3.129921] [<ffffffff8107fd12>] __local_bh_enable_ip+0x72/0xa0 [ 3.129925] [<ffffffff8147456a>] __pg_skb_alloc+0x2fa/0x3e0 [ 3.129928] [<ffffffff8147469e>] __alloc_skb+0x4e/0x250 [ 3.129932] [<ffffffffa07a9506>] btusb_recv_intr+0x126/0x170 [btusb] [ 3.129935] [<ffffffffa07a9600>] btusb_intr_complete+0xb0/0x140 [btusb] [ 3.129943] [<ffffffffa01fd2b5>] __usb_hcd_giveback_urb+0x85/0x140 [usbcore] [ 3.129950] [<ffffffffa01fd416>] usb_giveback_urb_bh+0xa6/0x100 [usbcore] [ 3.129953] [<ffffffff81080677>] tasklet_hi_action+0xb7/0xd0 [ 3.129955] [<ffffffff8107fe51>] __do_softirq+0x111/0x2d0 [ 3.129958] [<ffffffff8108018d>] irq_exit+0xad/0xb0 [ 3.129961] [<ffffffff81590d48>] do_IRQ+0x58/0xf0 [ 3.129965] [<ffffffff8158ed2e>] common_interrupt+0x6e/0x6e [ 3.129966] <EOI> [<ffffffff8101eb0d>] ? native_sched_clock+0x2d/0x90 [ 3.129972] [<ffffffff81442b63>] ? cpuidle_enter_state+0x93/0x240 [ 3.129975] [<ffffffff81442b36>] ? cpuidle_enter_state+0x66/0x240 [ 3.129977] [<ffffffff81442d47>] cpuidle_enter+0x17/0x20 [ 3.129981] [<ffffffff810bdf7e>] cpu_startup_entry+0x31e/0x430 [ 3.129984] [<ffffffff8104ea6e>] start_secondary+0x18e/0x1d0 [ 3.129986] ---[ end trace df240af6b2a08d04 ]--- ~~~ [__local_bh_enable_ip](http://lxr.free-electrons.com/source/kernel/softirq.c?v=4.1#L148) ~~~ 148 void __local_bh_enable_ip(unsigned long ip, unsigned int cnt) 149 { 150 WARN_ON_ONCE(in_irq() || irqs_disabled()); ~~~ The reason is [btusb_recv_intr](http://lxr.free-electrons.com/source/drivers/bluetooth/btusb.c?v=4.1#L396) which runs under tasklet and calls `alloc_skb` which calls `__pg_skb_alloc`.
1.0
Kernel warning when calling __alloc_skb from the soft-IRQ - Kernel warning was caught by me on the laptop: ~~~ [ 3.129833] ------------[ cut here ]------------ [ 3.129839] WARNING: CPU: 1 PID: 0 at kernel/softirq.c:150 __local_bh_enable_ip+0x72/0xa0() [ 3.129841] Modules linked in: media btusb btbcm btintel bluetooth cdc_mbim intel_rapl iosf_mbi x86_pkg_temp_thermal intel_powerclamp coretemp cdc_ncm usbnet kvm_intel mii cdc_wdm cdc_acm arc4 kvm iwldvm iTCO_wdt iTCO_vendor_support crct10dif_pclmul crc32_pclmul crc32c_intel mac80211 aesni_intel aes_x86_64 iwlwifi lrw gf128mul glue_helper snd_hda_codec_hdmi ablk_helper cryptd psmouse serio_raw pcspkr cfg80211 snd_hda_codec_realtek snd_hda_codec_generic snd_hda_intel thermal snd_hda_controller snd_hda_codec wmi snd_hda_core snd_hwdep thinkpad_acpi snd_pcm snd_timer nvram ac rfkill tpm_tis mei_me tpm battery snd mei evdev i2c_i801 lpc_ich soundcore shpchp mac_hid processor sch_fq_codel vboxnetadp(O) vboxdrv(O) ip_tables x_tables ext4 crc16 mbcache jbd2 sd_mod atkbd libps2 sdhci_pci ahci libahci libata [ 3.129883] scsi_mod xhci_pci ehci_pci sdhci ehci_hcd xhci_hcd led_class mmc_core usbcore usb_common i8042 serio i915 button intel_gtt i2c_algo_bit video drm_kms_helper drm [ 3.129894] CPU: 1 PID: 0 Comm: swapper/1 Tainted: G O 4.1.21-1-tfw #1 [ 3.129897] Hardware name: LENOVO 34442KU/34442KU, BIOS G6ETB1WW (2.71 ) 04/23/2015 [ 3.129899] 0000000000000086 3f00b632d218473c ffff88011e243c98 ffffffff81588da2 [ 3.129902] 0000000000000000 ffffffff817129f0 ffff88011e243cd8 ffffffff8107be4a [ 3.129904] ffff88011e2107e0 0000000000000200 0000000000000020 000077ff80000000 [ 3.129907] Call Trace: [ 3.129909] <IRQ> [<ffffffff81588da2>] dump_stack+0x63/0x81 [ 3.129915] [<ffffffff8107be4a>] warn_slowpath_common+0x8a/0xc0 [ 3.129918] [<ffffffff8107bf7a>] warn_slowpath_null+0x1a/0x20 [ 3.129921] [<ffffffff8107fd12>] __local_bh_enable_ip+0x72/0xa0 [ 3.129925] [<ffffffff8147456a>] __pg_skb_alloc+0x2fa/0x3e0 [ 3.129928] [<ffffffff8147469e>] __alloc_skb+0x4e/0x250 [ 3.129932] [<ffffffffa07a9506>] btusb_recv_intr+0x126/0x170 [btusb] [ 3.129935] [<ffffffffa07a9600>] btusb_intr_complete+0xb0/0x140 [btusb] [ 3.129943] [<ffffffffa01fd2b5>] __usb_hcd_giveback_urb+0x85/0x140 [usbcore] [ 3.129950] [<ffffffffa01fd416>] usb_giveback_urb_bh+0xa6/0x100 [usbcore] [ 3.129953] [<ffffffff81080677>] tasklet_hi_action+0xb7/0xd0 [ 3.129955] [<ffffffff8107fe51>] __do_softirq+0x111/0x2d0 [ 3.129958] [<ffffffff8108018d>] irq_exit+0xad/0xb0 [ 3.129961] [<ffffffff81590d48>] do_IRQ+0x58/0xf0 [ 3.129965] [<ffffffff8158ed2e>] common_interrupt+0x6e/0x6e [ 3.129966] <EOI> [<ffffffff8101eb0d>] ? native_sched_clock+0x2d/0x90 [ 3.129972] [<ffffffff81442b63>] ? cpuidle_enter_state+0x93/0x240 [ 3.129975] [<ffffffff81442b36>] ? cpuidle_enter_state+0x66/0x240 [ 3.129977] [<ffffffff81442d47>] cpuidle_enter+0x17/0x20 [ 3.129981] [<ffffffff810bdf7e>] cpu_startup_entry+0x31e/0x430 [ 3.129984] [<ffffffff8104ea6e>] start_secondary+0x18e/0x1d0 [ 3.129986] ---[ end trace df240af6b2a08d04 ]--- ~~~ [__local_bh_enable_ip](http://lxr.free-electrons.com/source/kernel/softirq.c?v=4.1#L148) ~~~ 148 void __local_bh_enable_ip(unsigned long ip, unsigned int cnt) 149 { 150 WARN_ON_ONCE(in_irq() || irqs_disabled()); ~~~ The reason is [btusb_recv_intr](http://lxr.free-electrons.com/source/drivers/bluetooth/btusb.c?v=4.1#L396) which runs under tasklet and calls `alloc_skb` which calls `__pg_skb_alloc`.
non_defect
kernel warning when calling alloc skb from the soft irq kernel warning was caught by me on the laptop warning cpu pid at kernel softirq c local bh enable ip modules linked in media btusb btbcm btintel bluetooth cdc mbim intel rapl iosf mbi pkg temp thermal intel powerclamp coretemp cdc ncm usbnet kvm intel mii cdc wdm cdc acm kvm iwldvm itco wdt itco vendor support pclmul pclmul intel aesni intel aes iwlwifi lrw glue helper snd hda codec hdmi ablk helper cryptd psmouse serio raw pcspkr snd hda codec realtek snd hda codec generic snd hda intel thermal snd hda controller snd hda codec wmi snd hda core snd hwdep thinkpad acpi snd pcm snd timer nvram ac rfkill tpm tis mei me tpm battery snd mei evdev lpc ich soundcore shpchp mac hid processor sch fq codel vboxnetadp o vboxdrv o ip tables x tables mbcache sd mod atkbd sdhci pci ahci libahci libata scsi mod xhci pci ehci pci sdhci ehci hcd xhci hcd led class mmc core usbcore usb common serio button intel gtt algo bit video drm kms helper drm cpu pid comm swapper tainted g o tfw hardware name lenovo bios call trace dump stack warn slowpath common warn slowpath null local bh enable ip pg skb alloc alloc skb btusb recv intr btusb intr complete usb hcd giveback urb usb giveback urb bh tasklet hi action do softirq irq exit do irq common interrupt native sched clock cpuidle enter state cpuidle enter state cpuidle enter cpu startup entry start secondary void local bh enable ip unsigned long ip unsigned int cnt warn on once in irq irqs disabled the reason is which runs under tasklet and calls alloc skb which calls pg skb alloc
0
13,964
2,789,803,377
IssuesEvent
2015-05-08 21:35:35
google/google-visualization-api-issues
https://api.github.com/repos/google/google-visualization-api-issues
opened
LineChart select event to open URL does not work on Internet Explorer
Priority-Medium Type-Defect
Original [issue 157](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=157) created by orwant on 2009-12-29T22:33:16.000Z: <b>What steps will reproduce the problem? Please provide a link to a</b> <b>demonstration page if at all possible, or attach code.</b> 1. Code is attached. Hover over data points to see the pop up tooltip, and when clicked, it will go to a URL (in this case, www.google.com). It seems to work fine in Firefox, Safari, and Chrome but in Internet Explorer, the clicking of the point to go to the URL doesn't work. You can see the highlighting of the point in the other browsers, but not in IE. This same URL functionality works fine in the PieChart component on Internet Explorer. <b>What component is this issue related to (PieChart, LineChart, DataTable,</b> <b>Query, etc)?</b> LineChart <b>Are you using the test environment (version 1.1)?</b> No. <b>What operating system and browser are you using?</b> Windows XP &amp; Internet Explorer 7 &amp; 8. <b>*********************************************************</b> <b>For developers viewing this issue: please click the 'star' icon to be</b> <b>notified of future changes, and to let us know how many of you are</b> <b>interested in seeing it resolved.</b> <b>*********************************************************</b>
1.0
LineChart select event to open URL does not work on Internet Explorer - Original [issue 157](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=157) created by orwant on 2009-12-29T22:33:16.000Z: <b>What steps will reproduce the problem? Please provide a link to a</b> <b>demonstration page if at all possible, or attach code.</b> 1. Code is attached. Hover over data points to see the pop up tooltip, and when clicked, it will go to a URL (in this case, www.google.com). It seems to work fine in Firefox, Safari, and Chrome but in Internet Explorer, the clicking of the point to go to the URL doesn't work. You can see the highlighting of the point in the other browsers, but not in IE. This same URL functionality works fine in the PieChart component on Internet Explorer. <b>What component is this issue related to (PieChart, LineChart, DataTable,</b> <b>Query, etc)?</b> LineChart <b>Are you using the test environment (version 1.1)?</b> No. <b>What operating system and browser are you using?</b> Windows XP &amp; Internet Explorer 7 &amp; 8. <b>*********************************************************</b> <b>For developers viewing this issue: please click the 'star' icon to be</b> <b>notified of future changes, and to let us know how many of you are</b> <b>interested in seeing it resolved.</b> <b>*********************************************************</b>
defect
linechart select event to open url does not work on internet explorer original created by orwant on what steps will reproduce the problem please provide a link to a demonstration page if at all possible or attach code code is attached hover over data points to see the pop up tooltip and when clicked it will go to a url in this case it seems to work fine in firefox safari and chrome but in internet explorer the clicking of the point to go to the url doesn t work you can see the highlighting of the point in the other browsers but not in ie this same url functionality works fine in the piechart component on internet explorer what component is this issue related to piechart linechart datatable query etc linechart are you using the test environment version no what operating system and browser are you using windows xp amp internet explorer amp for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved
1
48,498
13,100,884,246
IssuesEvent
2020-08-04 01:59:51
FoldingAtHome/fah-issues
https://api.github.com/repos/FoldingAtHome/fah-issues
closed
Incomplete fix for Core_Outdated with two slots
1.Type - Defect 3.Component - FAHClient 4.OS - All
Core_outdated means FAHClient needs to download a new core. I have two slots using that core. One was running and I unpaused the other. Naturally, the file was is use so the update couldn't proceed. Pausing the active slot gives me two slots with status Update_Core. That.s correct ... except that now the download needs to proceed but it doesn't Windows 7.6.9 Core had been deleted looking for a fresh copy. After a long wait, I paused both slots and then unpaused them and something worked. I'm not sure if that unjammed it or if it just took ~10 minutes to decide to proceed.
1.0
Incomplete fix for Core_Outdated with two slots - Core_outdated means FAHClient needs to download a new core. I have two slots using that core. One was running and I unpaused the other. Naturally, the file was is use so the update couldn't proceed. Pausing the active slot gives me two slots with status Update_Core. That.s correct ... except that now the download needs to proceed but it doesn't Windows 7.6.9 Core had been deleted looking for a fresh copy. After a long wait, I paused both slots and then unpaused them and something worked. I'm not sure if that unjammed it or if it just took ~10 minutes to decide to proceed.
defect
incomplete fix for core outdated with two slots core outdated means fahclient needs to download a new core i have two slots using that core one was running and i unpaused the other naturally the file was is use so the update couldn t proceed pausing the active slot gives me two slots with status update core that s correct except that now the download needs to proceed but it doesn t windows core had been deleted looking for a fresh copy after a long wait i paused both slots and then unpaused them and something worked i m not sure if that unjammed it or if it just took minutes to decide to proceed
1
10,636
2,622,178,205
IssuesEvent
2015-03-04 00:17:37
byzhang/leveldb
https://api.github.com/repos/byzhang/leveldb
closed
Add version info in include/c.h
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. There isn't any version info in b binding file include/c.h What is the expected output? What do you see instead? I can get version info in c interface What version of the product are you using? On what operating system? ANY I'm working on a php binding using c binding, but I can't get version info, If it was supplied as a macro will be much appreciated. ``` Original issue reported on code.google.com by `reeze....@gmail.com` on 13 Jun 2012 at 3:42
1.0
Add version info in include/c.h - ``` What steps will reproduce the problem? 1. There isn't any version info in b binding file include/c.h What is the expected output? What do you see instead? I can get version info in c interface What version of the product are you using? On what operating system? ANY I'm working on a php binding using c binding, but I can't get version info, If it was supplied as a macro will be much appreciated. ``` Original issue reported on code.google.com by `reeze....@gmail.com` on 13 Jun 2012 at 3:42
defect
add version info in include c h what steps will reproduce the problem there isn t any version info in b binding file include c h what is the expected output what do you see instead i can get version info in c interface what version of the product are you using on what operating system any i m working on a php binding using c binding but i can t get version info if it was supplied as a macro will be much appreciated original issue reported on code google com by reeze gmail com on jun at
1
58,466
16,546,038,260
IssuesEvent
2021-05-28 00:13:12
idaholab/moose
https://api.github.com/repos/idaholab/moose
closed
Fixed PETSc alt update script
P: normal T: defect
## Bug Description <!--A clear and concise description of the problem (Note: A missing feature is not a bug).--> ``` ❯ cd ~/projects/moose/ ❯ rm -rf petsc ❯ cd scripts ❯ ./update_and_rebuild_petsc_alt.sh *** WARNING *** scripts/update_and_rebuild_petsc_alt.sh is intended for internal use only. Please use scripts/update_and_rebuild_petsc.sh instead. /Users/milljm/projects/moose/scripts ./update_and_rebuild_petsc_alt.sh: line 30: cd: /Users/milljm/projects/moose/scripts/../petsc: No such file or directory error: pathspec 'v3.11.4' did not match any file(s) known to git git checkout command failed, are your proxy settings correct? ``` ## Fix <!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)--> We need to add "cd $SCRIPT_DIR/.." So that we are always in the right directory ## Impact <!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
1.0
Fixed PETSc alt update script - ## Bug Description <!--A clear and concise description of the problem (Note: A missing feature is not a bug).--> ``` ❯ cd ~/projects/moose/ ❯ rm -rf petsc ❯ cd scripts ❯ ./update_and_rebuild_petsc_alt.sh *** WARNING *** scripts/update_and_rebuild_petsc_alt.sh is intended for internal use only. Please use scripts/update_and_rebuild_petsc.sh instead. /Users/milljm/projects/moose/scripts ./update_and_rebuild_petsc_alt.sh: line 30: cd: /Users/milljm/projects/moose/scripts/../petsc: No such file or directory error: pathspec 'v3.11.4' did not match any file(s) known to git git checkout command failed, are your proxy settings correct? ``` ## Fix <!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)--> We need to add "cd $SCRIPT_DIR/.." So that we are always in the right directory ## Impact <!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
defect
fixed petsc alt update script bug description ❯ cd projects moose ❯ rm rf petsc ❯ cd scripts ❯ update and rebuild petsc alt sh warning scripts update and rebuild petsc alt sh is intended for internal use only please use scripts update and rebuild petsc sh instead users milljm projects moose scripts update and rebuild petsc alt sh line cd users milljm projects moose scripts petsc no such file or directory error pathspec did not match any file s known to git git checkout command failed are your proxy settings correct fix we need to add cd script dir so that we are always in the right directory impact
1
451,118
32,008,908,622
IssuesEvent
2023-09-21 16:34:08
NOAA-EMC/NCEPLIBS-ip
https://api.github.com/repos/NOAA-EMC/NCEPLIBS-ip
closed
Missing documentation for iplib and ip2lib header files.
documentation
In ip2lib_4.h we have: ``` /** @file * @brief C interface to gdswzd() and gdswzd_grib1() functions for '4' * library build. * @author NOAA Programmer */ #ifndef IPLIB #define IPLIB /** GDSWZD in C. @param igdtnum @param igdtmpl @param igdtlen @param iopt @param npts @param fill @param *xpts @param *ypts @param *rlon @param *rlat @param *nret @param *crot @param *srot @param *xlon @param *xlat @param *ylon @param *ylat @param *area */ void gdswzd(int igdtnum, int *igdtmpl, int igdtlen, int iopt, int npts, float fill, float *xpts, float *ypts, float *rlon, float *rlat, int *nret, float *crot, float *srot, float *xlon, float *xlat, float *ylon, float *ylat, float *area); /** GDSWZD_grib1 in C. @param kgds @param iopt @param npts @param *fill @param *xpts @param *ypts @param *rlon @param *rlat @param nret @param *crot @param *srot @param *xlon @param *xlat @param *ylon @param *ylat @param *area */ void gdswzd_grib1(int kgds, int iopt, int npts, float *fill, float *xpts, float *ypts, float *rlon, float *rlat, int nret, float *crot, float *srot, float *xlon, float *xlat, float *ylon, float *ylat, float *area); #endif ``` We need much better documentation than this. Let's start by figuring out who the author is. @GeorgeGayno-NOAA are you the author of this code?
1.0
Missing documentation for iplib and ip2lib header files. - In ip2lib_4.h we have: ``` /** @file * @brief C interface to gdswzd() and gdswzd_grib1() functions for '4' * library build. * @author NOAA Programmer */ #ifndef IPLIB #define IPLIB /** GDSWZD in C. @param igdtnum @param igdtmpl @param igdtlen @param iopt @param npts @param fill @param *xpts @param *ypts @param *rlon @param *rlat @param *nret @param *crot @param *srot @param *xlon @param *xlat @param *ylon @param *ylat @param *area */ void gdswzd(int igdtnum, int *igdtmpl, int igdtlen, int iopt, int npts, float fill, float *xpts, float *ypts, float *rlon, float *rlat, int *nret, float *crot, float *srot, float *xlon, float *xlat, float *ylon, float *ylat, float *area); /** GDSWZD_grib1 in C. @param kgds @param iopt @param npts @param *fill @param *xpts @param *ypts @param *rlon @param *rlat @param nret @param *crot @param *srot @param *xlon @param *xlat @param *ylon @param *ylat @param *area */ void gdswzd_grib1(int kgds, int iopt, int npts, float *fill, float *xpts, float *ypts, float *rlon, float *rlat, int nret, float *crot, float *srot, float *xlon, float *xlat, float *ylon, float *ylat, float *area); #endif ``` We need much better documentation than this. Let's start by figuring out who the author is. @GeorgeGayno-NOAA are you the author of this code?
non_defect
missing documentation for iplib and header files in h we have file brief c interface to gdswzd and gdswzd functions for library build author noaa programmer ifndef iplib define iplib gdswzd in c param igdtnum param igdtmpl param igdtlen param iopt param npts param fill param xpts param ypts param rlon param rlat param nret param crot param srot param xlon param xlat param ylon param ylat param area void gdswzd int igdtnum int igdtmpl int igdtlen int iopt int npts float fill float xpts float ypts float rlon float rlat int nret float crot float srot float xlon float xlat float ylon float ylat float area gdswzd in c param kgds param iopt param npts param fill param xpts param ypts param rlon param rlat param nret param crot param srot param xlon param xlat param ylon param ylat param area void gdswzd int kgds int iopt int npts float fill float xpts float ypts float rlon float rlat int nret float crot float srot float xlon float xlat float ylon float ylat float area endif we need much better documentation than this let s start by figuring out who the author is georgegayno noaa are you the author of this code
0
534,879
15,651,035,094
IssuesEvent
2021-03-23 09:43:22
leihs/leihs
https://api.github.com/repos/leihs/leihs
reopened
show a message when the user can login but has no access right
low priority
- [ ] add page in `my` - [ ] redirect from legacy
1.0
show a message when the user can login but has no access right - - [ ] add page in `my` - [ ] redirect from legacy
non_defect
show a message when the user can login but has no access right add page in my redirect from legacy
0
354,749
25,174,839,999
IssuesEvent
2022-11-11 08:17:48
TZL0/pe
https://api.github.com/repos/TZL0/pe
opened
Incorrect destination of PDF link
severity.Low type.DocumentationBug
Pressing the 'Command Line Interface' takes the user to Graphical User Interface ![image.png](https://raw.githubusercontent.com/TZL0/pe/main/files/319a9015-1f20-434b-ac35-bd2b9f70e024.png) ![image.png](https://raw.githubusercontent.com/TZL0/pe/main/files/67bef17e-a3dd-4acd-a0cf-ae430c9f5476.png) <!--session: 1668154547514-b0cf5391-b882-4c7a-8aeb-5d77963c2458--> <!--Version: Web v3.4.4-->
1.0
Incorrect destination of PDF link - Pressing the 'Command Line Interface' takes the user to Graphical User Interface ![image.png](https://raw.githubusercontent.com/TZL0/pe/main/files/319a9015-1f20-434b-ac35-bd2b9f70e024.png) ![image.png](https://raw.githubusercontent.com/TZL0/pe/main/files/67bef17e-a3dd-4acd-a0cf-ae430c9f5476.png) <!--session: 1668154547514-b0cf5391-b882-4c7a-8aeb-5d77963c2458--> <!--Version: Web v3.4.4-->
non_defect
incorrect destination of pdf link pressing the command line interface takes the user to graphical user interface
0
68,305
21,628,211,685
IssuesEvent
2022-05-05 06:43:16
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
SQL Translation Error
T: Defect P: Medium E: All Editions C: Translator
### SQL translator incorrectly translates this query: ```sql CREATE PROCEDURE unluckyEmployees() BEGIN SET @rn =0
1.0
SQL Translation Error - ### SQL translator incorrectly translates this query: ```sql CREATE PROCEDURE unluckyEmployees() BEGIN SET @rn =0
defect
sql translation error sql translator incorrectly translates this query sql create procedure unluckyemployees begin set rn
1
48,347
13,068,455,172
IssuesEvent
2020-07-31 03:38:01
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
[clsim] GPU detection broken for RTX-series cards (Trac #2212)
Migrated from Trac combo simulation defect
Nvidia decided that ray-tracing was a thing, and renamed their GTX series "RTX", and e.g. the 2080 Ti advertises itself as "GeForce RTX 2080 Ti". clsim.traysegments.common.configureOpenCLDevices, however, special-cases cards named "GTX" to set the number of work items to something sane and enable native math. Is there any reason not to do away with the special cases, and simply scale the number of work items to be proportional to the global memory size, as well as enabling native math by default? Migrated from https://code.icecube.wisc.edu/ticket/2212 ```json { "status": "closed", "changetime": "2019-03-06T20:58:01", "description": "Nvidia decided that ray-tracing was a thing, and renamed their GTX series \"RTX\", and e.g. the 2080 Ti advertises itself as \"GeForce RTX 2080 Ti\". clsim.traysegments.common.configureOpenCLDevices, however, special-cases cards named \"GTX\" to set the number of work items to something sane and enable native math. Is there any reason not to do away with the special cases, and simply scale the number of work items to be proportional to the global memory size, as well as enabling native math by default?", "reporter": "jvansanten", "cc": "", "resolution": "fixed", "_ts": "1551905881354849", "component": "combo simulation", "summary": "[clsim] GPU detection broken for RTX-series cards", "priority": "normal", "keywords": "", "time": "2018-11-29T14:14:01", "milestone": "", "owner": "jvansanten", "type": "defect" } ```
1.0
[clsim] GPU detection broken for RTX-series cards (Trac #2212) - Nvidia decided that ray-tracing was a thing, and renamed their GTX series "RTX", and e.g. the 2080 Ti advertises itself as "GeForce RTX 2080 Ti". clsim.traysegments.common.configureOpenCLDevices, however, special-cases cards named "GTX" to set the number of work items to something sane and enable native math. Is there any reason not to do away with the special cases, and simply scale the number of work items to be proportional to the global memory size, as well as enabling native math by default? Migrated from https://code.icecube.wisc.edu/ticket/2212 ```json { "status": "closed", "changetime": "2019-03-06T20:58:01", "description": "Nvidia decided that ray-tracing was a thing, and renamed their GTX series \"RTX\", and e.g. the 2080 Ti advertises itself as \"GeForce RTX 2080 Ti\". clsim.traysegments.common.configureOpenCLDevices, however, special-cases cards named \"GTX\" to set the number of work items to something sane and enable native math. Is there any reason not to do away with the special cases, and simply scale the number of work items to be proportional to the global memory size, as well as enabling native math by default?", "reporter": "jvansanten", "cc": "", "resolution": "fixed", "_ts": "1551905881354849", "component": "combo simulation", "summary": "[clsim] GPU detection broken for RTX-series cards", "priority": "normal", "keywords": "", "time": "2018-11-29T14:14:01", "milestone": "", "owner": "jvansanten", "type": "defect" } ```
defect
gpu detection broken for rtx series cards trac nvidia decided that ray tracing was a thing and renamed their gtx series rtx and e g the ti advertises itself as geforce rtx ti clsim traysegments common configureopencldevices however special cases cards named gtx to set the number of work items to something sane and enable native math is there any reason not to do away with the special cases and simply scale the number of work items to be proportional to the global memory size as well as enabling native math by default migrated from json status closed changetime description nvidia decided that ray tracing was a thing and renamed their gtx series rtx and e g the ti advertises itself as geforce rtx ti clsim traysegments common configureopencldevices however special cases cards named gtx to set the number of work items to something sane and enable native math is there any reason not to do away with the special cases and simply scale the number of work items to be proportional to the global memory size as well as enabling native math by default reporter jvansanten cc resolution fixed ts component combo simulation summary gpu detection broken for rtx series cards priority normal keywords time milestone owner jvansanten type defect
1
232,095
25,564,962,440
IssuesEvent
2022-11-30 13:40:30
billmcchesney1/concord
https://api.github.com/repos/billmcchesney1/concord
opened
CVE-2022-38900 (High) detected in decode-uri-component-0.2.0.tgz
security vulnerability
## CVE-2022-38900 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>decode-uri-component-0.2.0.tgz</b></p></summary> <p>A better decodeURIComponent</p> <p>Library home page: <a href="https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz">https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz</a></p> <p>Path to dependency file: /console2/package.json</p> <p>Path to vulnerable library: /console2/node_modules/decode-uri-component/package.json</p> <p> Dependency Hierarchy: - query-string-6.13.5.tgz (Root Library) - :x: **decode-uri-component-0.2.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> decode-uri-component 0.2.0 is vulnerable to Improper Input Validation resulting in DoS. <p>Publish Date: 2022-11-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38900>CVE-2022-38900</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
True
CVE-2022-38900 (High) detected in decode-uri-component-0.2.0.tgz - ## CVE-2022-38900 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>decode-uri-component-0.2.0.tgz</b></p></summary> <p>A better decodeURIComponent</p> <p>Library home page: <a href="https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz">https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz</a></p> <p>Path to dependency file: /console2/package.json</p> <p>Path to vulnerable library: /console2/node_modules/decode-uri-component/package.json</p> <p> Dependency Hierarchy: - query-string-6.13.5.tgz (Root Library) - :x: **decode-uri-component-0.2.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> decode-uri-component 0.2.0 is vulnerable to Improper Input Validation resulting in DoS. <p>Publish Date: 2022-11-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38900>CVE-2022-38900</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
non_defect
cve high detected in decode uri component tgz cve high severity vulnerability vulnerable library decode uri component tgz a better decodeuricomponent library home page a href path to dependency file package json path to vulnerable library node modules decode uri component package json dependency hierarchy query string tgz root library x decode uri component tgz vulnerable library found in base branch master vulnerability details decode uri component is vulnerable to improper input validation resulting in dos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href
0
59,945
17,023,296,356
IssuesEvent
2021-07-03 01:17:22
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Handling of notConnectedSameTag don't work with juction with dirrerent ways
Component: osmarender Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 1.56pm, Wednesday, 17th September 2008]** Take folowing test case: a rount about junction which is part of a primary road (so that way forming the round about itself is also tagged with same name a ways of primary road to an from the junction). In this case the round about is not labelled as expected due to notConnectedSameTag on corresponding rule (see correct.svg). But if I connect an other road with a different name to the round about, the labels reappears (see notcorrect.svg). The attached path corrects this issue.
1.0
Handling of notConnectedSameTag don't work with juction with dirrerent ways - **[Submitted to the original trac issue database at 1.56pm, Wednesday, 17th September 2008]** Take folowing test case: a rount about junction which is part of a primary road (so that way forming the round about itself is also tagged with same name a ways of primary road to an from the junction). In this case the round about is not labelled as expected due to notConnectedSameTag on corresponding rule (see correct.svg). But if I connect an other road with a different name to the round about, the labels reappears (see notcorrect.svg). The attached path corrects this issue.
defect
handling of notconnectedsametag don t work with juction with dirrerent ways take folowing test case a rount about junction which is part of a primary road so that way forming the round about itself is also tagged with same name a ways of primary road to an from the junction in this case the round about is not labelled as expected due to notconnectedsametag on corresponding rule see correct svg but if i connect an other road with a different name to the round about the labels reappears see notcorrect svg the attached path corrects this issue
1
592,952
17,934,668,281
IssuesEvent
2021-09-10 13:56:50
mozilla/addons-server
https://api.github.com/repos/mozilla/addons-server
closed
Uploading an add-on with a GUID in non string format triggers a server error
component: devhub priority: p3
### Describe the problem and steps to reproduce it: 1. Submit an add-on with a GUID in non string format, for example: ``` "browser_specific_settings": { "gecko": { "id": 12345 ``` 2. Check the validation results ### What happened? "There was a problem contacting the server" error message is received with a 500 response ![image](https://user-images.githubusercontent.com/31961530/132530234-33e445ea-4fec-4d10-affc-2a51bb77ab53.png) ### What did you expect to happen? The error message should mention that the GUID needs to be in string format and no server error should be triggered ### Anything else we should know? - reproduced on all AMO servers - the linter already handles this exception and has an error message for it: ``` Code Message Description File Line Column MANIFEST_FIELD_INVALID "/browser_specific_s… See https://mzl.la/1ZOhoEN (MDN Docs) for manifest.json ettings/gecko/id" more information. should be string ```
1.0
Uploading an add-on with a GUID in non string format triggers a server error - ### Describe the problem and steps to reproduce it: 1. Submit an add-on with a GUID in non string format, for example: ``` "browser_specific_settings": { "gecko": { "id": 12345 ``` 2. Check the validation results ### What happened? "There was a problem contacting the server" error message is received with a 500 response ![image](https://user-images.githubusercontent.com/31961530/132530234-33e445ea-4fec-4d10-affc-2a51bb77ab53.png) ### What did you expect to happen? The error message should mention that the GUID needs to be in string format and no server error should be triggered ### Anything else we should know? - reproduced on all AMO servers - the linter already handles this exception and has an error message for it: ``` Code Message Description File Line Column MANIFEST_FIELD_INVALID "/browser_specific_s… See https://mzl.la/1ZOhoEN (MDN Docs) for manifest.json ettings/gecko/id" more information. should be string ```
non_defect
uploading an add on with a guid in non string format triggers a server error describe the problem and steps to reproduce it submit an add on with a guid in non string format for example browser specific settings gecko id check the validation results what happened there was a problem contacting the server error message is received with a response what did you expect to happen the error message should mention that the guid needs to be in string format and no server error should be triggered anything else we should know reproduced on all amo servers the linter already handles this exception and has an error message for it code message description file line column manifest field invalid browser specific s… see mdn docs for manifest json ettings gecko id more information should be string
0
61,157
17,023,620,285
IssuesEvent
2021-07-03 02:57:49
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
oneway arrows not visible
Component: mapnik Priority: major Resolution: invalid Type: defect
**[Submitted to the original trac issue database at 10.07pm, Wednesday, 28th July 2010]** It would seem when the road name just barely fits the length of the road it can completely hide the oneway arrows making navigation rather exciting. When you don't see the road names at least you know to zoom in to see it, but when you see oneway arrows all around it's not so obvious. [http://www.openstreetmap.org/?lat=51.49191&lon=-0.19612&zoom=17&layers=M For example here] Nevern Square is oneway all the way; the north-south stretch is split to nortbound and southbound oneways, but it looks very much like two-way street.
1.0
oneway arrows not visible - **[Submitted to the original trac issue database at 10.07pm, Wednesday, 28th July 2010]** It would seem when the road name just barely fits the length of the road it can completely hide the oneway arrows making navigation rather exciting. When you don't see the road names at least you know to zoom in to see it, but when you see oneway arrows all around it's not so obvious. [http://www.openstreetmap.org/?lat=51.49191&lon=-0.19612&zoom=17&layers=M For example here] Nevern Square is oneway all the way; the north-south stretch is split to nortbound and southbound oneways, but it looks very much like two-way street.
defect
oneway arrows not visible it would seem when the road name just barely fits the length of the road it can completely hide the oneway arrows making navigation rather exciting when you don t see the road names at least you know to zoom in to see it but when you see oneway arrows all around it s not so obvious nevern square is oneway all the way the north south stretch is split to nortbound and southbound oneways but it looks very much like two way street
1
26,474
4,726,256,857
IssuesEvent
2016-10-18 09:37:52
PowerDNS/pdns
https://api.github.com/repos/PowerDNS/pdns
closed
API zone info returns an url without a starting /
auth defect web
When querying the API for zones with: `curl -H 'X-API-Key: secret' http://127.0.0.1:8081/api/v1/servers/localhost/zones` the returned json contains an url to the api zone info. This url does not start with a `/` whereas the docs specify that it should. (Other url fields from the api do start with a `/`). The 3.4.x api does not seem to have this issue Powerdns version 4.0.1 Distribution: Gentoo Reproduce: 1. enable api 2. curl the api along the lines of `curl -H 'X-API-Key: secret' http://127.0.0.1:8081/api/v1/servers/localhost/zones` expected behaviour: URL field starts with a `/`, for example: `"url": "/api/v1/servers/localhost/zones/test.example.org."` Observed behavior URL does not start with a slash, for example: "url": "api/v1/servers/localhost/zones/test.example.org." Proposed fix: Add a `/` at the relevant place in getZoneInfo in ws-auth.cc
1.0
API zone info returns an url without a starting / - When querying the API for zones with: `curl -H 'X-API-Key: secret' http://127.0.0.1:8081/api/v1/servers/localhost/zones` the returned json contains an url to the api zone info. This url does not start with a `/` whereas the docs specify that it should. (Other url fields from the api do start with a `/`). The 3.4.x api does not seem to have this issue Powerdns version 4.0.1 Distribution: Gentoo Reproduce: 1. enable api 2. curl the api along the lines of `curl -H 'X-API-Key: secret' http://127.0.0.1:8081/api/v1/servers/localhost/zones` expected behaviour: URL field starts with a `/`, for example: `"url": "/api/v1/servers/localhost/zones/test.example.org."` Observed behavior URL does not start with a slash, for example: "url": "api/v1/servers/localhost/zones/test.example.org." Proposed fix: Add a `/` at the relevant place in getZoneInfo in ws-auth.cc
defect
api zone info returns an url without a starting when querying the api for zones with curl h x api key secret the returned json contains an url to the api zone info this url does not start with a whereas the docs specify that it should other url fields from the api do start with a the x api does not seem to have this issue powerdns version distribution gentoo reproduce enable api curl the api along the lines of curl h x api key secret expected behaviour url field starts with a for example url api servers localhost zones test example org observed behavior url does not start with a slash for example url api servers localhost zones test example org proposed fix add a at the relevant place in getzoneinfo in ws auth cc
1
16,228
2,878,875,574
IssuesEvent
2015-06-10 06:00:50
msg4real/pygooglevoice
https://api.github.com/repos/msg4real/pygooglevoice
closed
The comment in settings.py about forwardingNumber is wrong
auto-migrated Priority-Medium Type-Defect
``` # Number to place calls from (eg, your google voice number) forwardingNumber= /snip The forwarding number is *NOT* your google voice number. It is the number you want GV to connect the call to. This confuses people trying to setup gvoice to make call by running 'gvoice call 8005551212' ``` Original issue reported on code.google.com by `prog...@gmail.com` on 22 Sep 2010 at 9:35
1.0
The comment in settings.py about forwardingNumber is wrong - ``` # Number to place calls from (eg, your google voice number) forwardingNumber= /snip The forwarding number is *NOT* your google voice number. It is the number you want GV to connect the call to. This confuses people trying to setup gvoice to make call by running 'gvoice call 8005551212' ``` Original issue reported on code.google.com by `prog...@gmail.com` on 22 Sep 2010 at 9:35
defect
the comment in settings py about forwardingnumber is wrong number to place calls from eg your google voice number forwardingnumber snip the forwarding number is not your google voice number it is the number you want gv to connect the call to this confuses people trying to setup gvoice to make call by running gvoice call original issue reported on code google com by prog gmail com on sep at
1
39,680
5,116,571,411
IssuesEvent
2017-01-07 05:18:02
caseyg/knutepunkt2017
https://api.github.com/repos/caseyg/knutepunkt2017
opened
8. Pettersson, Hamlets, Vampires, and the Italian Alps
design
Paper. Has images that we want to make bigger.
1.0
8. Pettersson, Hamlets, Vampires, and the Italian Alps - Paper. Has images that we want to make bigger.
non_defect
pettersson hamlets vampires and the italian alps paper has images that we want to make bigger
0
76,178
26,276,535,376
IssuesEvent
2023-01-06 22:48:58
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
Linux Kernel 6.2 renamed get_acl to get_inode_acl
Type: Defect
### System information Type | Version/Name --- | --- Distribution Name | Gentoo Distribution Version | - Kernel Version | `next-20221220` Architecture | LoongArch OpenZFS Version | 2.1.99-1641_gc935fe2e9 ### Describe the problem you're observing ``` checking whether iops->get_acl() exists... configure: error: *** None of the expected "iops->get_acl()" interfaces were detected. *** This may be because your kernel version is newer than what is *** supported, or you are using a patched custom kernel with *** incompatible modifications. *** *** ZFS Version: zfs-2.1.99-1641_gc935fe2e9 *** Compatible Kernels: 3.10 - 6.0 ``` ⚠️ ⚠️ **Kernel renamed `get_acl` to `get_inode_acl`: https://github.com/torvalds/linux/commit/cac2f8b8d8b50ef32b3e34f6dcbbf08937e4f616** ⚠️ ⚠️ ### Describe how to reproduce the problem Build oepnzfs master with a post-6.1 kernel. ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` --> ``` # make modules -C /usr/src/linux M=/var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu make: Entering directory '/usr/src/linux-loongson' CC [M] /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.o /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.c:53:36: error: initialization of ‘struct posix_acl * (*)(struct user_namespace *, struct dentry *, int)’ from incompatible pointer type ‘struct posix_acl * (*)(struct inode *, int, bool)’ {aka ‘struct posix_acl * (*)(struct inode *, int, _Bool)’} [-Werror=incompatible-pointer-types] 53 | .get_acl = get_acl_fn, | ^~~~~~~~~~ /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.c:53:36: note: (near initialization for ‘iops.get_acl’) cc1: all warnings being treated as errors make[1]: *** [scripts/Makefile.build:250: /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.o] Error 1 make: *** [Makefile:1994: /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu] Error 2 ```
1.0
Linux Kernel 6.2 renamed get_acl to get_inode_acl - ### System information Type | Version/Name --- | --- Distribution Name | Gentoo Distribution Version | - Kernel Version | `next-20221220` Architecture | LoongArch OpenZFS Version | 2.1.99-1641_gc935fe2e9 ### Describe the problem you're observing ``` checking whether iops->get_acl() exists... configure: error: *** None of the expected "iops->get_acl()" interfaces were detected. *** This may be because your kernel version is newer than what is *** supported, or you are using a patched custom kernel with *** incompatible modifications. *** *** ZFS Version: zfs-2.1.99-1641_gc935fe2e9 *** Compatible Kernels: 3.10 - 6.0 ``` ⚠️ ⚠️ **Kernel renamed `get_acl` to `get_inode_acl`: https://github.com/torvalds/linux/commit/cac2f8b8d8b50ef32b3e34f6dcbbf08937e4f616** ⚠️ ⚠️ ### Describe how to reproduce the problem Build oepnzfs master with a post-6.1 kernel. ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` --> ``` # make modules -C /usr/src/linux M=/var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu make: Entering directory '/usr/src/linux-loongson' CC [M] /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.o /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.c:53:36: error: initialization of ‘struct posix_acl * (*)(struct user_namespace *, struct dentry *, int)’ from incompatible pointer type ‘struct posix_acl * (*)(struct inode *, int, bool)’ {aka ‘struct posix_acl * (*)(struct inode *, int, _Bool)’} [-Werror=incompatible-pointer-types] 53 | .get_acl = get_acl_fn, | ^~~~~~~~~~ /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.c:53:36: note: (near initialization for ‘iops.get_acl’) cc1: all warnings being treated as errors make[1]: *** [scripts/Makefile.build:250: /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.o] Error 1 make: *** [Makefile:1994: /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu] Error 2 ```
defect
linux kernel renamed get acl to get inode acl system information type version name distribution name gentoo distribution version kernel version next architecture loongarch openzfs version describe the problem you re observing checking whether iops get acl exists configure error none of the expected iops get acl interfaces were detected this may be because your kernel version is newer than what is supported or you are using a patched custom kernel with incompatible modifications zfs version zfs compatible kernels ⚠️ ⚠️ kernel renamed get acl to get inode acl ⚠️ ⚠️ describe how to reproduce the problem build oepnzfs master with a post kernel include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with make modules c usr src linux m var tmp portage sys fs zfs loong kmod work zfs loong kmod build inode operations get acl rcu make entering directory usr src linux loongson cc var tmp portage sys fs zfs loong kmod work zfs loong kmod build inode operations get acl rcu inode operations get acl rcu o var tmp portage sys fs zfs loong kmod work zfs loong kmod build inode operations get acl rcu inode operations get acl rcu c error initialization of ‘struct posix acl struct user namespace struct dentry int ’ from incompatible pointer type ‘struct posix acl struct inode int bool ’ aka ‘struct posix acl struct inode int bool ’ get acl get acl fn var tmp portage sys fs zfs loong kmod work zfs loong kmod build inode operations get acl rcu inode operations get acl rcu c note near initialization for ‘iops get acl’ all warnings being treated as errors make error make error
1
19,006
10,312,367,584
IssuesEvent
2019-08-29 19:41:16
jamijam/WebGoat-Legacy
https://api.github.com/repos/jamijam/WebGoat-Legacy
opened
CVE-2019-12384 (Medium) detected in jackson-databind-2.0.4.jar
security vulnerability
## CVE-2019-12384 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.0.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Path to dependency file: /WebGoat-Legacy/pom.xml</p> <p>Path to vulnerable library: 2/repository/com/fasterxml/jackson/core/jackson-databind/2.0.4/jackson-databind-2.0.4.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.0.4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jamijam/WebGoat-Legacy/commit/47e282c80c0e0057166440fc7140b77c4f06feb6">47e282c80c0e0057166440fc7140b77c4f06feb6</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.9.1 might allow attackers to have a variety of impacts by leveraging failure to block the logback-core class from polymorphic deserialization. Depending on the classpath content, remote code execution may be possible. <p>Publish Date: 2019-06-24 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384>CVE-2019-12384</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384</a></p> <p>Release Date: 2019-08-12</p> <p>Fix Resolution: 2.9.9.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-12384 (Medium) detected in jackson-databind-2.0.4.jar - ## CVE-2019-12384 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.0.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Path to dependency file: /WebGoat-Legacy/pom.xml</p> <p>Path to vulnerable library: 2/repository/com/fasterxml/jackson/core/jackson-databind/2.0.4/jackson-databind-2.0.4.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.0.4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jamijam/WebGoat-Legacy/commit/47e282c80c0e0057166440fc7140b77c4f06feb6">47e282c80c0e0057166440fc7140b77c4f06feb6</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.9.1 might allow attackers to have a variety of impacts by leveraging failure to block the logback-core class from polymorphic deserialization. Depending on the classpath content, remote code execution may be possible. <p>Publish Date: 2019-06-24 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384>CVE-2019-12384</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384</a></p> <p>Release Date: 2019-08-12</p> <p>Fix Resolution: 2.9.9.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to dependency file webgoat legacy pom xml path to vulnerable library repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before might allow attackers to have a variety of impacts by leveraging failure to block the logback core class from polymorphic deserialization depending on the classpath content remote code execution may be possible publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
680,817
23,286,793,111
IssuesEvent
2022-08-05 17:25:08
mm-ninja-turtles/bulbasaur-express
https://api.github.com/repos/mm-ninja-turtles/bulbasaur-express
closed
[bug]: remove `response` object from resolver function args.
type: bug priority: high task: done
## Problem `resolver()` function currently provide express's `response` object and it should be removed since it will give the user the freedom to response as they wanted and it'll create the problem of response data is not matching with response schema. ## Solution 1. remove `response` object from args 1. #16 will solve the necessity of `response` object inside resolver function
1.0
[bug]: remove `response` object from resolver function args. - ## Problem `resolver()` function currently provide express's `response` object and it should be removed since it will give the user the freedom to response as they wanted and it'll create the problem of response data is not matching with response schema. ## Solution 1. remove `response` object from args 1. #16 will solve the necessity of `response` object inside resolver function
non_defect
remove response object from resolver function args problem resolver function currently provide express s response object and it should be removed since it will give the user the freedom to response as they wanted and it ll create the problem of response data is not matching with response schema solution remove response object from args will solve the necessity of response object inside resolver function
0
231,045
18,735,238,071
IssuesEvent
2021-11-04 06:15:17
kartoza/django-bims
https://api.github.com/repos/kartoza/django-bims
closed
Remove date when uploading unpublished datasets
testing SIZE 2 Checked by FRC bug-report Source references Repeat offender - previous ticket
Please remove date when uploading unpublished data and database. It is creating extra work for us having to go and combine unpublished datasets I fixed all of these source reference a few weeks ago and now they are all wrong again :(
1.0
Remove date when uploading unpublished datasets - Please remove date when uploading unpublished data and database. It is creating extra work for us having to go and combine unpublished datasets I fixed all of these source reference a few weeks ago and now they are all wrong again :(
non_defect
remove date when uploading unpublished datasets please remove date when uploading unpublished data and database it is creating extra work for us having to go and combine unpublished datasets i fixed all of these source reference a few weeks ago and now they are all wrong again
0
79,503
22,782,906,821
IssuesEvent
2022-07-08 22:33:22
chaotic-aur/packages
https://api.github.com/repos/chaotic-aur/packages
closed
[Outdated] ffmpeg-full
request:new-pkg request:rebuild-pkg priority:medium bug:PKGBUILD
### If available, link to the latest build [ffmpeg-full.log](https://builds.garudalinux.org/repos/chaotic-aur/logs/ffmpeg-full.log) ### Package name `ffmpeg-full` ### Latest build `None` ### Latest version available `5.0.1-2` ### Have you tested if the package builds in a clean chroot? - [ ] Yes ### More information Adding missing deps in https://github.com/chaotic-aur/packages/commit/e2072dd61dca0b0b8c9732e76803278bba06ee27 but `python-pocketsphinx-git` still won't build: [python-pocketsphinx-git.log](https://builds.garudalinux.org/repos/chaotic-aur/logs/python-pocketsphinx-git.log)
2.0
[Outdated] ffmpeg-full - ### If available, link to the latest build [ffmpeg-full.log](https://builds.garudalinux.org/repos/chaotic-aur/logs/ffmpeg-full.log) ### Package name `ffmpeg-full` ### Latest build `None` ### Latest version available `5.0.1-2` ### Have you tested if the package builds in a clean chroot? - [ ] Yes ### More information Adding missing deps in https://github.com/chaotic-aur/packages/commit/e2072dd61dca0b0b8c9732e76803278bba06ee27 but `python-pocketsphinx-git` still won't build: [python-pocketsphinx-git.log](https://builds.garudalinux.org/repos/chaotic-aur/logs/python-pocketsphinx-git.log)
non_defect
ffmpeg full if available link to the latest build package name ffmpeg full latest build none latest version available have you tested if the package builds in a clean chroot yes more information adding missing deps in but python pocketsphinx git still won t build
0
9,137
2,615,133,741
IssuesEvent
2015-03-01 06:04:46
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
opened
With null accountName, get infinity loop
auto-migrated Priority-Medium Type-Defect
``` Version of google-api-java-client (e.g. 1.15.0-rc)? 1.17.0-rc Java environment (e.g. Java 6, Android 2.3, App Engine)? Android 4.2.2 Describe the problem. from source GoogleAccountCredential.java: public String getToken() throws IOException, GoogleAuthException { if (backOff != null) { backOff.reset(); } while (true) { try { return GoogleAuthUtil.getToken(context, accountName, scope); } catch (IOException e) { // network or server error, so retry using back-off policy try { if (backOff == null || !BackOffUtils.next(sleeper, backOff)) { throw e; } } catch (InterruptedException e2) { // ignore } } } If accountName = "" (something wrong, yes), GoogleAuthUtil.getToken runs forever. No exception or something. How would you expect it to be fixed? Exception? ``` Original issue reported on code.google.com by `rusxa...@gmail.com` on 14 Oct 2013 at 1:01
1.0
With null accountName, get infinity loop - ``` Version of google-api-java-client (e.g. 1.15.0-rc)? 1.17.0-rc Java environment (e.g. Java 6, Android 2.3, App Engine)? Android 4.2.2 Describe the problem. from source GoogleAccountCredential.java: public String getToken() throws IOException, GoogleAuthException { if (backOff != null) { backOff.reset(); } while (true) { try { return GoogleAuthUtil.getToken(context, accountName, scope); } catch (IOException e) { // network or server error, so retry using back-off policy try { if (backOff == null || !BackOffUtils.next(sleeper, backOff)) { throw e; } } catch (InterruptedException e2) { // ignore } } } If accountName = "" (something wrong, yes), GoogleAuthUtil.getToken runs forever. No exception or something. How would you expect it to be fixed? Exception? ``` Original issue reported on code.google.com by `rusxa...@gmail.com` on 14 Oct 2013 at 1:01
defect
with null accountname get infinity loop version of google api java client e g rc rc java environment e g java android app engine android describe the problem from source googleaccountcredential java public string gettoken throws ioexception googleauthexception if backoff null backoff reset while true try return googleauthutil gettoken context accountname scope catch ioexception e network or server error so retry using back off policy try if backoff null backoffutils next sleeper backoff throw e catch interruptedexception ignore if accountname something wrong yes googleauthutil gettoken runs forever no exception or something how would you expect it to be fixed exception original issue reported on code google com by rusxa gmail com on oct at
1
36,422
7,928,753,966
IssuesEvent
2018-07-06 12:53:41
otros-systems/otroslogviewer
https://api.github.com/repos/otros-systems/otroslogviewer
closed
Listening to udp not possible
Priority-Medium Type-Defect
``` When starting a socket listener, it always starts listening to tcp messages. However, it would be great if also listening to udp messages was supported. For example the log4judp logging backend of log4cplus could be used with the otroslogviewer then. ``` Original issue reported on code.google.com by `jan.rue...@dacuda.com` on 25 Sep 2013 at 11:45
1.0
Listening to udp not possible - ``` When starting a socket listener, it always starts listening to tcp messages. However, it would be great if also listening to udp messages was supported. For example the log4judp logging backend of log4cplus could be used with the otroslogviewer then. ``` Original issue reported on code.google.com by `jan.rue...@dacuda.com` on 25 Sep 2013 at 11:45
defect
listening to udp not possible when starting a socket listener it always starts listening to tcp messages however it would be great if also listening to udp messages was supported for example the logging backend of could be used with the otroslogviewer then original issue reported on code google com by jan rue dacuda com on sep at
1
21,470
3,511,527,923
IssuesEvent
2016-01-10 10:29:53
nielsAD/lape
https://api.github.com/repos/nielsAD/lape
closed
It's impossible with certain boolean comparisons.
auto-migrated Priority-High Type-Defect
``` What will reproduce the problem? if (SomeBool and True) then What is the expected output? What do you see instead? I get "It's impossible!" (Line 706 of lpvartypes_ord.pas) Which version are you using? c96f612e0066 Please provide any additional information below. Seems to not be an issue if True is first. ``` Original issue reported on code.google.com by `Dgby...@gmail.com` on 7 Sep 2014 at 12:33
1.0
It's impossible with certain boolean comparisons. - ``` What will reproduce the problem? if (SomeBool and True) then What is the expected output? What do you see instead? I get "It's impossible!" (Line 706 of lpvartypes_ord.pas) Which version are you using? c96f612e0066 Please provide any additional information below. Seems to not be an issue if True is first. ``` Original issue reported on code.google.com by `Dgby...@gmail.com` on 7 Sep 2014 at 12:33
defect
it s impossible with certain boolean comparisons what will reproduce the problem if somebool and true then what is the expected output what do you see instead i get it s impossible line of lpvartypes ord pas which version are you using please provide any additional information below seems to not be an issue if true is first original issue reported on code google com by dgby gmail com on sep at
1
28,079
5,184,504,936
IssuesEvent
2017-01-20 06:31:44
GarageGames/Torque3D
https://api.github.com/repos/GarageGames/Torque3D
closed
Terrain materials are double sided
Defect
I noticed when I fly under the terrain, it is still visible, this means the terrain materials are all double sided and this cannot be changed, since terrain materials, do not have the value to change if they are double sided or not. Since double sided materials cost more to render, the terrain should not be double sided. I also checked one version before and the terrain materials were all single sided like it should be, so they got broken in 3.9 update.
1.0
Terrain materials are double sided - I noticed when I fly under the terrain, it is still visible, this means the terrain materials are all double sided and this cannot be changed, since terrain materials, do not have the value to change if they are double sided or not. Since double sided materials cost more to render, the terrain should not be double sided. I also checked one version before and the terrain materials were all single sided like it should be, so they got broken in 3.9 update.
defect
terrain materials are double sided i noticed when i fly under the terrain it is still visible this means the terrain materials are all double sided and this cannot be changed since terrain materials do not have the value to change if they are double sided or not since double sided materials cost more to render the terrain should not be double sided i also checked one version before and the terrain materials were all single sided like it should be so they got broken in update
1
34,979
7,497,166,764
IssuesEvent
2018-04-08 17:00:56
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
[Bug?] Example of Bland's Rule for optimize.linprog (simplex) cycling/preventing termination
defect scipy.optimize
Forcing Bland's Rule with the "bland":True options seems to prevent the code example below from terminating, while the purpose of Bland's Rule is to make sure it terminates. The problem is solved quickly when not forcing Bland's Rule, giving an optimal value of -6.044533469014448. By "not terminating" i mean that it's still running after running for a couple of hours on a MacBook Pro, while my own implementation of Bland's Rule finds it in seconds. ### Reproducing code example: ``` import numpy as np import scipy.optimize as opt import sys np.random.seed(4) m = int(np.round(10 * np.exp(np.log(50) * np.random.rand()))) n = int(np.round(10 * np.exp(np.log(50) * np.random.rand()))) c, A, b = np.round(10*np.random.randn(n)),np.round(10*np.random.randn(m,n)),np.round(10*np.abs(np.random.randn(m))) opt.linprog(-1*c, A, b, options={"maxiter": sys.maxsize, "bland": True}) ``` ### Scipy/Numpy/Python version information: 1.0.0 1.14.0 sys.version_info(major=3, minor=6, micro=4, releaselevel='final', serial=0)
1.0
[Bug?] Example of Bland's Rule for optimize.linprog (simplex) cycling/preventing termination - Forcing Bland's Rule with the "bland":True options seems to prevent the code example below from terminating, while the purpose of Bland's Rule is to make sure it terminates. The problem is solved quickly when not forcing Bland's Rule, giving an optimal value of -6.044533469014448. By "not terminating" i mean that it's still running after running for a couple of hours on a MacBook Pro, while my own implementation of Bland's Rule finds it in seconds. ### Reproducing code example: ``` import numpy as np import scipy.optimize as opt import sys np.random.seed(4) m = int(np.round(10 * np.exp(np.log(50) * np.random.rand()))) n = int(np.round(10 * np.exp(np.log(50) * np.random.rand()))) c, A, b = np.round(10*np.random.randn(n)),np.round(10*np.random.randn(m,n)),np.round(10*np.abs(np.random.randn(m))) opt.linprog(-1*c, A, b, options={"maxiter": sys.maxsize, "bland": True}) ``` ### Scipy/Numpy/Python version information: 1.0.0 1.14.0 sys.version_info(major=3, minor=6, micro=4, releaselevel='final', serial=0)
defect
example of bland s rule for optimize linprog simplex cycling preventing termination forcing bland s rule with the bland true options seems to prevent the code example below from terminating while the purpose of bland s rule is to make sure it terminates the problem is solved quickly when not forcing bland s rule giving an optimal value of by not terminating i mean that it s still running after running for a couple of hours on a macbook pro while my own implementation of bland s rule finds it in seconds reproducing code example import numpy as np import scipy optimize as opt import sys np random seed m int np round np exp np log np random rand n int np round np exp np log np random rand c a b np round np random randn n np round np random randn m n np round np abs np random randn m opt linprog c a b options maxiter sys maxsize bland true scipy numpy python version information sys version info major minor micro releaselevel final serial
1
12,444
2,700,140,850
IssuesEvent
2015-04-03 22:54:05
netty/netty
https://api.github.com/repos/netty/netty
closed
HTTP/2 RST_STREAM frame for an IDLE stream should result in connection error
defect
http://http2.github.io/http2-spec/index.html#RST_STREAM > RST_STREAM frames MUST NOT be sent for a stream in the "idle" state. If a RST_STREAM frame identifying an idle stream is received, the recipient MUST treat this as a connection error (Section 5.4.1) of type PROTOCOL_ERROR.
1.0
HTTP/2 RST_STREAM frame for an IDLE stream should result in connection error - http://http2.github.io/http2-spec/index.html#RST_STREAM > RST_STREAM frames MUST NOT be sent for a stream in the "idle" state. If a RST_STREAM frame identifying an idle stream is received, the recipient MUST treat this as a connection error (Section 5.4.1) of type PROTOCOL_ERROR.
defect
http rst stream frame for an idle stream should result in connection error rst stream frames must not be sent for a stream in the idle state if a rst stream frame identifying an idle stream is received the recipient must treat this as a connection error section of type protocol error
1
39,390
6,741,336,636
IssuesEvent
2017-10-20 00:03:09
scrabill/how-many-days-until-halloween
https://api.github.com/repos/scrabill/how-many-days-until-halloween
closed
Documentation: Update CONTRIBUTE.md
documentation Hacktoberfest help wanted
Include the following - How to contribute - What is currently there / how it works - Links to resources
1.0
Documentation: Update CONTRIBUTE.md - Include the following - How to contribute - What is currently there / how it works - Links to resources
non_defect
documentation update contribute md include the following how to contribute what is currently there how it works links to resources
0
51,772
13,211,304,546
IssuesEvent
2020-08-15 22:10:39
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
Link to PROPOSAL project (Trac #1019)
Incomplete Migration Migrated from Trac cmake defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1019">https://code.icecube.wisc.edu/projects/icecube/ticket/1019</a>, reported by icecube</summary> <p> ```json { "status": "closed", "changetime": "2015-06-11T18:38:27", "_ts": "1434047907705473", "description": "Hi,\n\nCould you please provide a link to the paper describing how the PROPOSAL icesim meta-project work?\n\nThanks.", "reporter": "icecube", "cc": "jtatar@lbl.gov", "resolution": "wontfix", "time": "2015-06-11T17:37:23", "component": "cmake", "summary": "Link to PROPOSAL project", "priority": "normal", "keywords": "", "milestone": "", "owner": "", "type": "defect" } ``` </p> </details>
1.0
Link to PROPOSAL project (Trac #1019) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1019">https://code.icecube.wisc.edu/projects/icecube/ticket/1019</a>, reported by icecube</summary> <p> ```json { "status": "closed", "changetime": "2015-06-11T18:38:27", "_ts": "1434047907705473", "description": "Hi,\n\nCould you please provide a link to the paper describing how the PROPOSAL icesim meta-project work?\n\nThanks.", "reporter": "icecube", "cc": "jtatar@lbl.gov", "resolution": "wontfix", "time": "2015-06-11T17:37:23", "component": "cmake", "summary": "Link to PROPOSAL project", "priority": "normal", "keywords": "", "milestone": "", "owner": "", "type": "defect" } ``` </p> </details>
defect
link to proposal project trac migrated from json status closed changetime ts description hi n ncould you please provide a link to the paper describing how the proposal icesim meta project work n nthanks reporter icecube cc jtatar lbl gov resolution wontfix time component cmake summary link to proposal project priority normal keywords milestone owner type defect
1
338,745
24,597,597,255
IssuesEvent
2022-10-14 09:37:56
Thaza-Kun/sarjana
https://api.github.com/repos/Thaza-Kun/sarjana
opened
Check The Unit for $a$ and $K$
documentation
The values of $a$ should be the inverse of $K$ yet in `thesis\literature.qmd`, the units are not entirely the same: $a=4.148 806 4239(11) \text{GH}^2\text{pc}^{-1}\text{cm}^3\text{ms}$ $K=241.033 1786(66) \text{GH}^{-2}\text{pc}\text{cm}^{-3}\text{s}$ Note: In $a$, it is ...ms while in $K$, it is ...s <!-- Edit the body of your new issue then click the ✓ "Create Issue" button in the top right of the editor. The first line will be the issue title. Assignees and Labels follow after a blank line. Leave an empty line before beginning the body of the issue. -->
1.0
Check The Unit for $a$ and $K$ - The values of $a$ should be the inverse of $K$ yet in `thesis\literature.qmd`, the units are not entirely the same: $a=4.148 806 4239(11) \text{GH}^2\text{pc}^{-1}\text{cm}^3\text{ms}$ $K=241.033 1786(66) \text{GH}^{-2}\text{pc}\text{cm}^{-3}\text{s}$ Note: In $a$, it is ...ms while in $K$, it is ...s <!-- Edit the body of your new issue then click the ✓ "Create Issue" button in the top right of the editor. The first line will be the issue title. Assignees and Labels follow after a blank line. Leave an empty line before beginning the body of the issue. -->
non_defect
check the unit for a and k the values of a should be the inverse of k yet in thesis literature qmd the units are not entirely the same a text gh text pc text cm text ms k text gh text pc text cm text s note in a it is ms while in k it is s
0
21,992
3,587,833,741
IssuesEvent
2016-01-30 16:15:46
dkfans/keeperfx
https://api.github.com/repos/dkfans/keeperfx
reopened
Level Kari-Mar automatic victory due to Computer Player suicide
Component-CompPlayer Component-Maps Priority-Low Status-Duplicate Type-Defect
Originally reported on Google Code with ID 213 ``` Free Play level Kari-Mar on version 0.4.4. Early in the game you suddenly win, without ever seeing an enemy. Most likely map invites Computer Player to dig to strong heroes. Savegame attached. ``` Reported by `Loobinex` on 2014-01-15 15:40:39 <hr> * *Attachment: [fx1g0004.sav](https://storage.googleapis.com/google-code-attachments/keeperfx/issue-213/comment-0/fx1g0004.sav)*
1.0
Level Kari-Mar automatic victory due to Computer Player suicide - Originally reported on Google Code with ID 213 ``` Free Play level Kari-Mar on version 0.4.4. Early in the game you suddenly win, without ever seeing an enemy. Most likely map invites Computer Player to dig to strong heroes. Savegame attached. ``` Reported by `Loobinex` on 2014-01-15 15:40:39 <hr> * *Attachment: [fx1g0004.sav](https://storage.googleapis.com/google-code-attachments/keeperfx/issue-213/comment-0/fx1g0004.sav)*
defect
level kari mar automatic victory due to computer player suicide originally reported on google code with id free play level kari mar on version early in the game you suddenly win without ever seeing an enemy most likely map invites computer player to dig to strong heroes savegame attached reported by loobinex on attachment
1
10,473
2,622,165,432
IssuesEvent
2015-03-04 00:12:05
byzhang/graphchi
https://api.github.com/repos/byzhang/graphchi
opened
Cannot make sharder_basic - fixed by adding an extra include
auto-migrated Priority-Medium Type-Defect
``` I tried to build the sharder_basic program with the latest download and it failed with this error: g++ -g -O3 -I/usr/local/include/ -I./src/ -fopenmp -Wall -Wno-strict-aliasing src/preprocessing/sharder_basic.cpp -o bin/sharder_basic In file included from ./src/preprocessing/conversions.hpp:36, from src/preprocessing/sharder_basic.cpp:35: ./src/preprocessing/sharder.hpp: In member function ‘virtual void graphchi::sharder<EdgeDataType>::write_shards()’: ./src/preprocessing/sharder.hpp:465: error: ‘degree’ was not declared in this scope ./src/preprocessing/sharder.hpp:467: error: ‘degrees’ was not declared in this scope ./src/preprocessing/sharder.hpp:469: error: expected primary-expression before ‘)’ token ./src/preprocessing/sharder.hpp:469: error: expected `;' before ‘calloc’ make: *** [sharder_basic] Error 1 Including the degree_data header seemed to fix this and I was able to build and run basic_sharder after this #include "engine/auxdata/degree_data.hpp" ``` Original issue reported on code.google.com by `Jennifer...@gmail.com` on 13 Jan 2013 at 3:16
1.0
Cannot make sharder_basic - fixed by adding an extra include - ``` I tried to build the sharder_basic program with the latest download and it failed with this error: g++ -g -O3 -I/usr/local/include/ -I./src/ -fopenmp -Wall -Wno-strict-aliasing src/preprocessing/sharder_basic.cpp -o bin/sharder_basic In file included from ./src/preprocessing/conversions.hpp:36, from src/preprocessing/sharder_basic.cpp:35: ./src/preprocessing/sharder.hpp: In member function ‘virtual void graphchi::sharder<EdgeDataType>::write_shards()’: ./src/preprocessing/sharder.hpp:465: error: ‘degree’ was not declared in this scope ./src/preprocessing/sharder.hpp:467: error: ‘degrees’ was not declared in this scope ./src/preprocessing/sharder.hpp:469: error: expected primary-expression before ‘)’ token ./src/preprocessing/sharder.hpp:469: error: expected `;' before ‘calloc’ make: *** [sharder_basic] Error 1 Including the degree_data header seemed to fix this and I was able to build and run basic_sharder after this #include "engine/auxdata/degree_data.hpp" ``` Original issue reported on code.google.com by `Jennifer...@gmail.com` on 13 Jan 2013 at 3:16
defect
cannot make sharder basic fixed by adding an extra include i tried to build the sharder basic program with the latest download and it failed with this error g g i usr local include i src fopenmp wall wno strict aliasing src preprocessing sharder basic cpp o bin sharder basic in file included from src preprocessing conversions hpp from src preprocessing sharder basic cpp src preprocessing sharder hpp in member function ‘virtual void graphchi sharder write shards ’ src preprocessing sharder hpp error ‘degree’ was not declared in this scope src preprocessing sharder hpp error ‘degrees’ was not declared in this scope src preprocessing sharder hpp error expected primary expression before ‘ ’ token src preprocessing sharder hpp error expected before ‘calloc’ make error including the degree data header seemed to fix this and i was able to build and run basic sharder after this include engine auxdata degree data hpp original issue reported on code google com by jennifer gmail com on jan at
1
20,097
3,295,315,224
IssuesEvent
2015-10-31 20:48:32
chief-atx/bcmon
https://api.github.com/repos/chief-atx/bcmon
closed
Galaxy s 4
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? Dont finding anything. What version of the product are you using? On what operating system? galaxy s 4 cm 12.1 5.1.1 Please provide any additional information below. Maybe u can help me to connect tl-wn722 cos i need mon mode on galaxy s 4 ``` Original issue reported on code.google.com by `nshasher...@gmail.com` on 23 Aug 2015 at 12:23
1.0
Galaxy s 4 - ``` What steps will reproduce the problem? Dont finding anything. What version of the product are you using? On what operating system? galaxy s 4 cm 12.1 5.1.1 Please provide any additional information below. Maybe u can help me to connect tl-wn722 cos i need mon mode on galaxy s 4 ``` Original issue reported on code.google.com by `nshasher...@gmail.com` on 23 Aug 2015 at 12:23
defect
galaxy s what steps will reproduce the problem dont finding anything what version of the product are you using on what operating system galaxy s cm please provide any additional information below maybe u can help me to connect tl cos i need mon mode on galaxy s original issue reported on code google com by nshasher gmail com on aug at
1
63,728
17,872,009,895
IssuesEvent
2021-09-06 17:09:04
martinrotter/rssguard
https://api.github.com/repos/martinrotter/rssguard
closed
Linking fails when version 3.9.2 is installed
Type-Defect Status-Not-Enough-Data
When building version 4.0.1 while version 3.9.2 is installed, linking fails, being unable to resolve `parseCmdArgumentsFromOtherInstance(const QString& message)` symbol. Removing or renaming `/usr/local/lib/librssguard.so` helps, but the following small change to `src/rssguard/rssguard.pro` would make the build more robust: ```patch -unix: LIBS += -L$$OUT_PWD/../librssguard/ -lrssguard +unix: LIBS += $$OUT_PWD/../librssguard/librssguard.so ```
1.0
Linking fails when version 3.9.2 is installed - When building version 4.0.1 while version 3.9.2 is installed, linking fails, being unable to resolve `parseCmdArgumentsFromOtherInstance(const QString& message)` symbol. Removing or renaming `/usr/local/lib/librssguard.so` helps, but the following small change to `src/rssguard/rssguard.pro` would make the build more robust: ```patch -unix: LIBS += -L$$OUT_PWD/../librssguard/ -lrssguard +unix: LIBS += $$OUT_PWD/../librssguard/librssguard.so ```
defect
linking fails when version is installed when building version while version is installed linking fails being unable to resolve parsecmdargumentsfromotherinstance const qstring message symbol removing or renaming usr local lib librssguard so helps but the following small change to src rssguard rssguard pro would make the build more robust patch unix libs l out pwd librssguard lrssguard unix libs out pwd librssguard librssguard so
1
17,624
3,012,774,652
IssuesEvent
2015-07-29 02:27:30
yawlfoundation/yawl
https://api.github.com/repos/yawlfoundation/yawl
closed
Importing adds to specification in case specification is erroneous
auto-migrated Category-Component-Editor Priority-Medium Type-Defect
``` When I *open* specification new60.ywl I get an error and the editor states that the load file will be discarded. However, the specification is shown and when I subsequently *import* another specification (e.g. new61.xml), two nets are shown. This is not possible with specifications which load normally as the import option is then disabled. ``` Original issue reported on code.google.com by `arthurte...@gmail.com` on 20 Aug 2008 at 7:49 Attachments: * [net60.ywl](https://storage.googleapis.com/google-code-attachments/yawl/issue-108/comment-0/net60.ywl) * [new61.xml](https://storage.googleapis.com/google-code-attachments/yawl/issue-108/comment-0/new61.xml)
1.0
Importing adds to specification in case specification is erroneous - ``` When I *open* specification new60.ywl I get an error and the editor states that the load file will be discarded. However, the specification is shown and when I subsequently *import* another specification (e.g. new61.xml), two nets are shown. This is not possible with specifications which load normally as the import option is then disabled. ``` Original issue reported on code.google.com by `arthurte...@gmail.com` on 20 Aug 2008 at 7:49 Attachments: * [net60.ywl](https://storage.googleapis.com/google-code-attachments/yawl/issue-108/comment-0/net60.ywl) * [new61.xml](https://storage.googleapis.com/google-code-attachments/yawl/issue-108/comment-0/new61.xml)
defect
importing adds to specification in case specification is erroneous when i open specification ywl i get an error and the editor states that the load file will be discarded however the specification is shown and when i subsequently import another specification e g xml two nets are shown this is not possible with specifications which load normally as the import option is then disabled original issue reported on code google com by arthurte gmail com on aug at attachments
1
38,995
5,207,021,621
IssuesEvent
2017-01-24 22:14:56
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
We should support Load Balancer or remove from drop-down list
area/webhooks kind/bug status/resolved status/to-test
**Rancher Versions:** master 1/17 **Steps to Reproduce:** 1. Add a load balancer 2. Go to Add webhook and click on drop-down **Results:** Load balancer is in list **Expected:** We should support it or remove it from drop-down.
1.0
We should support Load Balancer or remove from drop-down list - **Rancher Versions:** master 1/17 **Steps to Reproduce:** 1. Add a load balancer 2. Go to Add webhook and click on drop-down **Results:** Load balancer is in list **Expected:** We should support it or remove it from drop-down.
non_defect
we should support load balancer or remove from drop down list rancher versions master steps to reproduce add a load balancer go to add webhook and click on drop down results load balancer is in list expected we should support it or remove it from drop down
0
796,903
28,131,175,889
IssuesEvent
2023-03-31 23:24:59
integrations/terraform-provider-github
https://api.github.com/repos/integrations/terraform-provider-github
closed
The repository_file data source throws an error when trying to access a non-existent file
Type: Bug Priority: Normal
### Affected Resource(s) Please list the resources as a list, for example: - `data.github_repository_file` ### Expected Behavior The `repository_file` data source allows checking if a specific file exists in a repository. The same way that `branch`data source, for example, does (see: https://github.com/integrations/terraform-provider-github/blob/main/github/data_source_github_branch.go#L52). ### Actual Behavior The `repository_file` data source throws an error when trying to get a file which does not exist (see: https://github.com/integrations/terraform-provider-github/blob/main/github/data_source_github_repository_file.go#L82). ### References - https://github.com/integrations/terraform-provider-github/pull/1129
1.0
The repository_file data source throws an error when trying to access a non-existent file - ### Affected Resource(s) Please list the resources as a list, for example: - `data.github_repository_file` ### Expected Behavior The `repository_file` data source allows checking if a specific file exists in a repository. The same way that `branch`data source, for example, does (see: https://github.com/integrations/terraform-provider-github/blob/main/github/data_source_github_branch.go#L52). ### Actual Behavior The `repository_file` data source throws an error when trying to get a file which does not exist (see: https://github.com/integrations/terraform-provider-github/blob/main/github/data_source_github_repository_file.go#L82). ### References - https://github.com/integrations/terraform-provider-github/pull/1129
non_defect
the repository file data source throws an error when trying to access a non existent file affected resource s please list the resources as a list for example data github repository file expected behavior the repository file data source allows checking if a specific file exists in a repository the same way that branch data source for example does see actual behavior the repository file data source throws an error when trying to get a file which does not exist see references
0
75,505
25,888,415,428
IssuesEvent
2022-12-14 16:06:14
DependencyTrack/dependency-track
https://api.github.com/repos/DependencyTrack/dependency-track
opened
After Component update, only CVEs related to updated CPE should be available
defect in triage
### Current Behavior If we update the version and CPE of an existing component, all old CVEs not related to the new CPE and their audit histories are still available in Vulnerability Audit. ### Steps to Reproduce 1. Edit any exist component with outdated version and CPE > Note available CVEs 2. Change version number and CPE to new one 3. After some time check for new CVEs > Old and new CVEs are observed ### Expected Behavior After updating the version and CPE of an existing component, only new CVEs and previously found CVEs related to the new CPE and their audit histories should be available in Vulnerability Audit. ### Dependency-Track Version 4.6.2 ### Dependency-Track Distribution Container Image ### Database Server PostgreSQL ### Database Server Version _No response_ ### Browser Mozilla Firefox ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
1.0
After Component update, only CVEs related to updated CPE should be available - ### Current Behavior If we update the version and CPE of an existing component, all old CVEs not related to the new CPE and their audit histories are still available in Vulnerability Audit. ### Steps to Reproduce 1. Edit any exist component with outdated version and CPE > Note available CVEs 2. Change version number and CPE to new one 3. After some time check for new CVEs > Old and new CVEs are observed ### Expected Behavior After updating the version and CPE of an existing component, only new CVEs and previously found CVEs related to the new CPE and their audit histories should be available in Vulnerability Audit. ### Dependency-Track Version 4.6.2 ### Dependency-Track Distribution Container Image ### Database Server PostgreSQL ### Database Server Version _No response_ ### Browser Mozilla Firefox ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
defect
after component update only cves related to updated cpe should be available current behavior if we update the version and cpe of an existing component all old cves not related to the new cpe and their audit histories are still available in vulnerability audit steps to reproduce edit any exist component with outdated version and cpe note available cves change version number and cpe to new one after some time check for new cves old and new cves are observed expected behavior after updating the version and cpe of an existing component only new cves and previously found cves related to the new cpe and their audit histories should be available in vulnerability audit dependency track version dependency track distribution container image database server postgresql database server version no response browser mozilla firefox checklist i have read and understand the i have checked the for whether this defect was already reported
1
16,739
2,941,305,424
IssuesEvent
2015-07-02 06:49:26
tnt944445/reaver-wps
https://api.github.com/repos/tnt944445/reaver-wps
closed
Try multiple BSSIDs, round robin
auto-migrated Priority-Triage Type-Defect
``` I found myself sitting an entire evening switching between two routers I was testing. The first (Thomson) locks for five minutes after five attempts. The second (dlink) reboots after a few minutes of testing. Try one until it locks, swith to the other until it reboots, switch back to the first, etc ad nauseam. It would be great if reaver could take arguments like -b thefirst,thesecond,thethird -c 1,6,11 and do this switching by itself. Switch to the next if it detects a lock or progress stops. ``` Original issue reported on code.google.com by `stur...@gmail.com` on 27 Feb 2013 at 9:05
1.0
Try multiple BSSIDs, round robin - ``` I found myself sitting an entire evening switching between two routers I was testing. The first (Thomson) locks for five minutes after five attempts. The second (dlink) reboots after a few minutes of testing. Try one until it locks, swith to the other until it reboots, switch back to the first, etc ad nauseam. It would be great if reaver could take arguments like -b thefirst,thesecond,thethird -c 1,6,11 and do this switching by itself. Switch to the next if it detects a lock or progress stops. ``` Original issue reported on code.google.com by `stur...@gmail.com` on 27 Feb 2013 at 9:05
defect
try multiple bssids round robin i found myself sitting an entire evening switching between two routers i was testing the first thomson locks for five minutes after five attempts the second dlink reboots after a few minutes of testing try one until it locks swith to the other until it reboots switch back to the first etc ad nauseam it would be great if reaver could take arguments like b thefirst thesecond thethird c and do this switching by itself switch to the next if it detects a lock or progress stops original issue reported on code google com by stur gmail com on feb at
1
310,441
26,716,616,535
IssuesEvent
2023-01-28 15:54:07
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
Error when trying use trait alias for Fn() in impl
A-traits E-needs-test T-compiler C-bug F-trait_alias
This code produce compiling error: ```rust #![feature(trait_alias)] struct MyStruct {} trait MyFn = Fn(&MyStruct); fn foo (_: impl MyFn) {} fn main () { foo(|_| {}); } ``` ([Playground](https://play.rust-lang.org/?version=nightly&mode=debug&edition=2018&gist=6f1bfe3f9fca8240a2111c94b12a963d)) Errors: ``` Compiling playground v0.0.1 (/playground) error[E0631]: type mismatch in closure arguments --> src/main.rs:9:5 | 9 | foo(|_| {}); | ^^^ --- found signature of `fn(_) -> _` | | | expected signature of `for<'r> fn(&'r MyStruct) -> _` | note: required by `foo` --> src/main.rs:6:1 | 6 | fn foo (_: impl MyFn) {} | ^^^^^^^^^^^^^^^^^^^^^ error[E0271]: type mismatch resolving `for<'r> <[closure@src/main.rs:9:9: 9:15] as std::ops::FnOnce<(&'r MyStruct,)>>::Output == ()` --> src/main.rs:9:5 | 9 | foo(|_| {}); | ^^^ expected bound lifetime parameter, found concrete lifetime | note: required by `foo` --> src/main.rs:6:1 | 6 | fn foo (_: impl MyFn) {} | ^^^^^^^^^^^^^^^^^^^^^ error: aborting due to 2 previous errors For more information about this error, try `rustc --explain E0271`. error: Could not compile `playground`. To learn more, run the command again with --verbose. ``` But when non aliased trait used, then no error occured. That is when we replace `fn foo (_: impl MyFn) {}` with `fn foo (_: impl Fn(&MyStruct)) {}`
1.0
Error when trying use trait alias for Fn() in impl - This code produce compiling error: ```rust #![feature(trait_alias)] struct MyStruct {} trait MyFn = Fn(&MyStruct); fn foo (_: impl MyFn) {} fn main () { foo(|_| {}); } ``` ([Playground](https://play.rust-lang.org/?version=nightly&mode=debug&edition=2018&gist=6f1bfe3f9fca8240a2111c94b12a963d)) Errors: ``` Compiling playground v0.0.1 (/playground) error[E0631]: type mismatch in closure arguments --> src/main.rs:9:5 | 9 | foo(|_| {}); | ^^^ --- found signature of `fn(_) -> _` | | | expected signature of `for<'r> fn(&'r MyStruct) -> _` | note: required by `foo` --> src/main.rs:6:1 | 6 | fn foo (_: impl MyFn) {} | ^^^^^^^^^^^^^^^^^^^^^ error[E0271]: type mismatch resolving `for<'r> <[closure@src/main.rs:9:9: 9:15] as std::ops::FnOnce<(&'r MyStruct,)>>::Output == ()` --> src/main.rs:9:5 | 9 | foo(|_| {}); | ^^^ expected bound lifetime parameter, found concrete lifetime | note: required by `foo` --> src/main.rs:6:1 | 6 | fn foo (_: impl MyFn) {} | ^^^^^^^^^^^^^^^^^^^^^ error: aborting due to 2 previous errors For more information about this error, try `rustc --explain E0271`. error: Could not compile `playground`. To learn more, run the command again with --verbose. ``` But when non aliased trait used, then no error occured. That is when we replace `fn foo (_: impl MyFn) {}` with `fn foo (_: impl Fn(&MyStruct)) {}`
non_defect
error when trying use trait alias for fn in impl this code produce compiling error rust struct mystruct trait myfn fn mystruct fn foo impl myfn fn main foo errors compiling playground playground error type mismatch in closure arguments src main rs foo found signature of fn expected signature of for fn r mystruct note required by foo src main rs fn foo impl myfn error type mismatch resolving for output src main rs foo expected bound lifetime parameter found concrete lifetime note required by foo src main rs fn foo impl myfn error aborting due to previous errors for more information about this error try rustc explain error could not compile playground to learn more run the command again with verbose but when non aliased trait used then no error occured that is when we replace fn foo impl myfn with fn foo impl fn mystruct
0
49,001
13,185,189,493
IssuesEvent
2020-08-12 20:54:06
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
Increase robustness of ROOTSYS detection foo (Trac #587)
Incomplete Migration Migrated from Trac cmake defect
<details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/587 , reported by blaufuss and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2013-06-28T13:01:06", "description": "still often get strangish errors from ROOT voodoo (rootcint sometimes)\n\nOften attached to having non-ports ROOTSYS in path or LD_LIB path or\nenv.\n\nSee, for example, email from Seon-hee:\n\nHi all again,\n\nWhen I build before, I was setting \nROOTSYS=/data1/i3_ports-v3/root-v5.20.00 .\nWhen I do not set this as Henrik suggested,\nthe compilation goes fine.\n\nThanks!\n(and sorry for the spam,)\nSeon-Hee\n\nSeon-Hee Seo wrote:\nHi all again,\n\nI just needed to include dependent projects, simclasses and trigger-sim.\nNow it works for cmake part.\n(Thanks Henrik and RobertF.)\n\nBut my compile fails with the following error.\n\n-------------------------------------\n[ 56%] Generating \n/home/seo/icerec/V3-1-0/build/simclasses/CMakeFiles/simclassesDict.cxx \nwith rootcint\n/tmp/gab3A1_cint.cxx:1:13: error: C: No such file or directory\nError: external preprocessing failed. /usr/include/c++/4.1.3/exception:42:\nError: Too many '}' /usr/include/c++/4.1.3/exception:103:\n/tmp/FsefFB_cint.cxx:1:13: error: C: No such file or directory\nError: external preprocessing failed. /usr/include/c++/4.1.3/typeinfo:42:\nError: Symbol __constchar is not defined in current scope \n/usr/include/gconv.h:118:\nError: Symbol __modname is not defined in current scope \n/usr/include/gconv.h:118:\nError: class,struct,union or type __builtin_va_list not defined \n/usr/lib/gcc/x86_64-linux-gnu/4.1.3/include/stdarg.h:43:\nError: Too many '}' /usr/include/stdio.h:841:\nError: Too many '}' /usr/include/nl_types.h:53:\nSyntax error /usr/include/c++/4.1.3/x86_64-linux-gnu/bits/c++locale.h:55:\nError: unrecognized language construct \n/usr/include/c++/4.1.3/x86_64-linux-gnu/bits/c++locale.h:55:\nError: Symbol __uselocale is not defined in current scope \n/usr/include/c++/4.1.3/x86_64-linux-gnu/bits/c++locale.h:55:\nError: Symbol __constchar is not defined in current scope \n/usr/include/time.h:148:\nError: Symbol tm_zone is not defined in current scope \n/usr/include/time.h:148:\nError: Too many '}' /usr/include/pthread.h:1104:\nError: Too many '}' /usr/include/unistd.h:1069:\nSyntax Error: void* \n/usr/include/c++/4.1.3/x86_64-linux-gnu/bits/gthr-default.h:113:\nError: Too many '}' /usr/include/ctype.h:218:\nError: Too many '}' /usr/include/c++/4.1.3/bits/functexcept.h:91:\nError: class,struct,union or type ... not defined \n/usr/include/c++/4.1.3/bits/cpp_type_traits.h:83:\nInternal error: global function template arg type \n/usr/include/c++/4.1.3/bits/cpp_type_traits.h:83:\nError: Missing closing brace for the block opened around line 44.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 30.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 28.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 217.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 31.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 30.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 42.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nWarning: Error occurred during reading source files\nWarning: Error occurred during dictionary source generation\n!!!Removing \n/home/seo/icerec/V3-1-0/build/simclasses/CMakeFiles/simclassesDict.cxx \n/home/seo/icerec/V3-1-0/build/simclasses/CMakeFiles/simclassesDict.h !!!\nError: rootcint: error loading headers...\nmake[2]: *** [simclasses/CMakeFiles/simclassesDict.cxx] Error 1\nmake[1]: *** [simclasses/CMakeFiles/simclasses.dir/all] Error 2\nmake: *** [all] Error 2\n---------------------------------\n\nCould this be ROOT version issue?\nHow to deal with this?\n\nCheers,\nSeon-Hee\n\nSeon-Hee Seo wrote:\n\nHi all,\n\nI was able to build V03-01-00 icerec.\nBut when I include sim-services (most recent release V00-07-01),\nI get cmake error:\n\n--------------------------------------\n\n....\n-- sim-services\nCMake Error at cmake/project.cmake:82 (message):\n Attempt to use nonexistent project 'simclasses'\nCall Stack (most recent call first):\n cmake/project.cmake:208 (use_projects)\n sim-services/CMakeLists.txt:6 (i3_add_library)\n\n\n-- Configuring incomplete, errors occurred!\nmake: *** [rebuild_cache] Error 1\n----------------------------------------\n\nIn V2 icerec, I did not have this problem.\n\nI need sim-services project to change RunID, EventID\nto be able to use data which contains same RunID, EventID pair.\n(So that my analysis-tree handles the data properly.)\n\nThis data set is corsika set 2267 being used for filter study.\n\nAny suggestion?\n\nThanks,\nSeon-Hee\n_______________________________________________", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "_ts": "1372424466000000", "component": "cmake", "summary": "Increase robustness of ROOTSYS detection foo", "priority": "normal", "keywords": "", "time": "2010-01-19T20:53:52", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
Increase robustness of ROOTSYS detection foo (Trac #587) - <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/587 , reported by blaufuss and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2013-06-28T13:01:06", "description": "still often get strangish errors from ROOT voodoo (rootcint sometimes)\n\nOften attached to having non-ports ROOTSYS in path or LD_LIB path or\nenv.\n\nSee, for example, email from Seon-hee:\n\nHi all again,\n\nWhen I build before, I was setting \nROOTSYS=/data1/i3_ports-v3/root-v5.20.00 .\nWhen I do not set this as Henrik suggested,\nthe compilation goes fine.\n\nThanks!\n(and sorry for the spam,)\nSeon-Hee\n\nSeon-Hee Seo wrote:\nHi all again,\n\nI just needed to include dependent projects, simclasses and trigger-sim.\nNow it works for cmake part.\n(Thanks Henrik and RobertF.)\n\nBut my compile fails with the following error.\n\n-------------------------------------\n[ 56%] Generating \n/home/seo/icerec/V3-1-0/build/simclasses/CMakeFiles/simclassesDict.cxx \nwith rootcint\n/tmp/gab3A1_cint.cxx:1:13: error: C: No such file or directory\nError: external preprocessing failed. /usr/include/c++/4.1.3/exception:42:\nError: Too many '}' /usr/include/c++/4.1.3/exception:103:\n/tmp/FsefFB_cint.cxx:1:13: error: C: No such file or directory\nError: external preprocessing failed. /usr/include/c++/4.1.3/typeinfo:42:\nError: Symbol __constchar is not defined in current scope \n/usr/include/gconv.h:118:\nError: Symbol __modname is not defined in current scope \n/usr/include/gconv.h:118:\nError: class,struct,union or type __builtin_va_list not defined \n/usr/lib/gcc/x86_64-linux-gnu/4.1.3/include/stdarg.h:43:\nError: Too many '}' /usr/include/stdio.h:841:\nError: Too many '}' /usr/include/nl_types.h:53:\nSyntax error /usr/include/c++/4.1.3/x86_64-linux-gnu/bits/c++locale.h:55:\nError: unrecognized language construct \n/usr/include/c++/4.1.3/x86_64-linux-gnu/bits/c++locale.h:55:\nError: Symbol __uselocale is not defined in current scope \n/usr/include/c++/4.1.3/x86_64-linux-gnu/bits/c++locale.h:55:\nError: Symbol __constchar is not defined in current scope \n/usr/include/time.h:148:\nError: Symbol tm_zone is not defined in current scope \n/usr/include/time.h:148:\nError: Too many '}' /usr/include/pthread.h:1104:\nError: Too many '}' /usr/include/unistd.h:1069:\nSyntax Error: void* \n/usr/include/c++/4.1.3/x86_64-linux-gnu/bits/gthr-default.h:113:\nError: Too many '}' /usr/include/ctype.h:218:\nError: Too many '}' /usr/include/c++/4.1.3/bits/functexcept.h:91:\nError: class,struct,union or type ... not defined \n/usr/include/c++/4.1.3/bits/cpp_type_traits.h:83:\nInternal error: global function template arg type \n/usr/include/c++/4.1.3/bits/cpp_type_traits.h:83:\nError: Missing closing brace for the block opened around line 44.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 30.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 28.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 217.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 31.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 30.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 42.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nWarning: Error occurred during reading source files\nWarning: Error occurred during dictionary source generation\n!!!Removing \n/home/seo/icerec/V3-1-0/build/simclasses/CMakeFiles/simclassesDict.cxx \n/home/seo/icerec/V3-1-0/build/simclasses/CMakeFiles/simclassesDict.h !!!\nError: rootcint: error loading headers...\nmake[2]: *** [simclasses/CMakeFiles/simclassesDict.cxx] Error 1\nmake[1]: *** [simclasses/CMakeFiles/simclasses.dir/all] Error 2\nmake: *** [all] Error 2\n---------------------------------\n\nCould this be ROOT version issue?\nHow to deal with this?\n\nCheers,\nSeon-Hee\n\nSeon-Hee Seo wrote:\n\nHi all,\n\nI was able to build V03-01-00 icerec.\nBut when I include sim-services (most recent release V00-07-01),\nI get cmake error:\n\n--------------------------------------\n\n....\n-- sim-services\nCMake Error at cmake/project.cmake:82 (message):\n Attempt to use nonexistent project 'simclasses'\nCall Stack (most recent call first):\n cmake/project.cmake:208 (use_projects)\n sim-services/CMakeLists.txt:6 (i3_add_library)\n\n\n-- Configuring incomplete, errors occurred!\nmake: *** [rebuild_cache] Error 1\n----------------------------------------\n\nIn V2 icerec, I did not have this problem.\n\nI need sim-services project to change RunID, EventID\nto be able to use data which contains same RunID, EventID pair.\n(So that my analysis-tree handles the data properly.)\n\nThis data set is corsika set 2267 being used for filter study.\n\nAny suggestion?\n\nThanks,\nSeon-Hee\n_______________________________________________", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "_ts": "1372424466000000", "component": "cmake", "summary": "Increase robustness of ROOTSYS detection foo", "priority": "normal", "keywords": "", "time": "2010-01-19T20:53:52", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
increase robustness of rootsys detection foo trac migrated from reported by blaufuss and owned by nega json status closed changetime description still often get strangish errors from root voodoo rootcint sometimes n noften attached to having non ports rootsys in path or ld lib path or nenv n nsee for example email from seon hee n nhi all again n nwhen i build before i was setting nrootsys ports root nwhen i do not set this as henrik suggested nthe compilation goes fine n nthanks n and sorry for the spam nseon hee n nseon hee seo wrote nhi all again n ni just needed to include dependent projects simclasses and trigger sim nnow it works for cmake part n thanks henrik and robertf n nbut my compile fails with the following error n n n generating n home seo icerec build simclasses cmakefiles simclassesdict cxx nwith rootcint n tmp cint cxx error c no such file or directory nerror external preprocessing failed usr include c exception nerror too many usr include c exception n tmp fseffb cint cxx error c no such file or directory nerror external preprocessing failed usr include c typeinfo nerror symbol constchar is not defined in current scope n usr include gconv h nerror symbol modname is not defined in current scope n usr include gconv h nerror class struct union or type builtin va list not defined n usr lib gcc linux gnu include stdarg h nerror too many usr include stdio h nerror too many usr include nl types h nsyntax error usr include c linux gnu bits c locale h nerror unrecognized language construct n usr include c linux gnu bits c locale h nerror symbol uselocale is not defined in current scope n usr include c linux gnu bits c locale h nerror symbol constchar is not defined in current scope n usr include time h nerror symbol tm zone is not defined in current scope n usr include time h nerror too many usr include pthread h nerror too many usr include unistd h nsyntax error void n usr include c linux gnu bits gthr default h nerror too many usr include ctype h nerror too many usr include c bits functexcept h nerror class struct union or type not defined n usr include c bits cpp type traits h ninternal error global function template arg type n usr include c bits cpp type traits h nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nwarning error occurred during reading source files nwarning error occurred during dictionary source generation n removing n home seo icerec build simclasses cmakefiles simclassesdict cxx n home seo icerec build simclasses cmakefiles simclassesdict h nerror rootcint error loading headers nmake error nmake error nmake error n n ncould this be root version issue nhow to deal with this n ncheers nseon hee n nseon hee seo wrote n nhi all n ni was able to build icerec nbut when i include sim services most recent release ni get cmake error n n n n n sim services ncmake error at cmake project cmake message n attempt to use nonexistent project simclasses ncall stack most recent call first n cmake project cmake use projects n sim services cmakelists txt add library n n n configuring incomplete errors occurred nmake error n n nin icerec i did not have this problem n ni need sim services project to change runid eventid nto be able to use data which contains same runid eventid pair n so that my analysis tree handles the data properly n nthis data set is corsika set being used for filter study n nany suggestion n nthanks nseon hee n reporter blaufuss cc resolution fixed ts component cmake summary increase robustness of rootsys detection foo priority normal keywords time milestone owner nega type defect
1
15,477
2,856,459,610
IssuesEvent
2015-06-02 15:03:53
svalinn/DAGMC
https://api.github.com/repos/svalinn/DAGMC
closed
Change amalgamated pyne library name
Type: Defect Type: Enhancement
The compiled shared object library name clashes with standard pyne library name, it should be changed to something more appropriate that doesn't clash like `pyne_builtin`, something definitely not called pyne.
1.0
Change amalgamated pyne library name - The compiled shared object library name clashes with standard pyne library name, it should be changed to something more appropriate that doesn't clash like `pyne_builtin`, something definitely not called pyne.
defect
change amalgamated pyne library name the compiled shared object library name clashes with standard pyne library name it should be changed to something more appropriate that doesn t clash like pyne builtin something definitely not called pyne
1
166,390
6,303,924,156
IssuesEvent
2017-07-21 14:48:17
chez-nestor/backoffice
https://api.github.com/repos/chez-nestor/backoffice
closed
Write unit tests for complex Liana getters
[priority] P2 [type] enhancement
- [ ] load Liana collections ourselves from src/models ourselves, - [ ] find a way to build objects that will have those getters
1.0
Write unit tests for complex Liana getters - - [ ] load Liana collections ourselves from src/models ourselves, - [ ] find a way to build objects that will have those getters
non_defect
write unit tests for complex liana getters load liana collections ourselves from src models ourselves find a way to build objects that will have those getters
0
77,773
27,156,303,056
IssuesEvent
2023-02-17 08:09:17
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
opened
DataExporter: Attrribute 'visibleOnly' not documented
:lady_beetle: defect :bangbang: needs-triage
### Describe the bug The [online documentation](https://primefaces.github.io/primefaces/12_0_0/#/components/dataexporter) of component DataExporter has no description for attribute 'visibleOnly'. ### Reproducer _No response_ ### Expected behavior _No response_ ### PrimeFaces edition Community ### PrimeFaces version 12.0.0 ### Theme _No response_ ### JSF implementation Mojarra ### JSF version 2.3 ### Java version 11 ### Browser(s) _No response_
1.0
DataExporter: Attrribute 'visibleOnly' not documented - ### Describe the bug The [online documentation](https://primefaces.github.io/primefaces/12_0_0/#/components/dataexporter) of component DataExporter has no description for attribute 'visibleOnly'. ### Reproducer _No response_ ### Expected behavior _No response_ ### PrimeFaces edition Community ### PrimeFaces version 12.0.0 ### Theme _No response_ ### JSF implementation Mojarra ### JSF version 2.3 ### Java version 11 ### Browser(s) _No response_
defect
dataexporter attrribute visibleonly not documented describe the bug the of component dataexporter has no description for attribute visibleonly reproducer no response expected behavior no response primefaces edition community primefaces version theme no response jsf implementation mojarra jsf version java version browser s no response
1
9,578
2,615,162,942
IssuesEvent
2015-03-01 06:42:23
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
bssid.wpc empty
auto-migrated Priority-Triage Type-Defect
``` System crashed and i need to reboot. Now the bssid.wpc is empty! The file exist, sure reaver ask to restore the session but starts from the beginning. It was about 18% done and Pin 20XXXXXX, it takes me 1 week to get that far. Do i really have to start from the beginning? Or maybe i can write something in the bssid.wpc file to start from pin 20000000? What else do i have to write in this file so that the restore option from reaver will work and doesn´t start from the very beginning 0,0%. Sry my english isn´t the best. regards ``` Original issue reported on code.google.com by `rkutsch...@gmail.com` on 14 Jun 2012 at 11:11
1.0
bssid.wpc empty - ``` System crashed and i need to reboot. Now the bssid.wpc is empty! The file exist, sure reaver ask to restore the session but starts from the beginning. It was about 18% done and Pin 20XXXXXX, it takes me 1 week to get that far. Do i really have to start from the beginning? Or maybe i can write something in the bssid.wpc file to start from pin 20000000? What else do i have to write in this file so that the restore option from reaver will work and doesn´t start from the very beginning 0,0%. Sry my english isn´t the best. regards ``` Original issue reported on code.google.com by `rkutsch...@gmail.com` on 14 Jun 2012 at 11:11
defect
bssid wpc empty system crashed and i need to reboot now the bssid wpc is empty the file exist sure reaver ask to restore the session but starts from the beginning it was about done and pin it takes me week to get that far do i really have to start from the beginning or maybe i can write something in the bssid wpc file to start from pin what else do i have to write in this file so that the restore option from reaver will work and doesn´t start from the very beginning sry my english isn´t the best regards original issue reported on code google com by rkutsch gmail com on jun at
1
36,805
8,139,791,383
IssuesEvent
2018-08-20 18:52:40
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
Null-propagation does not work on nullable value tuples.
defect in-progress premium
### Steps To Reproduce https://deck.net/6574bfcdd0addabfc66d1cc0c3c7498f Related to https://github.com/bridgedotnet/Bridge/issues/3645 ```csharp public class App { public static void Main() { (string Prop1, string Prop2)? val = ("test1", "test2"); Console.WriteLine(val.Value.Prop1); Console.WriteLine(val?.Prop1); } } ``` ### Expected Result ``` > test1 > test1 ``` ### Actual Result ``` > test1 > ```
1.0
Null-propagation does not work on nullable value tuples. - ### Steps To Reproduce https://deck.net/6574bfcdd0addabfc66d1cc0c3c7498f Related to https://github.com/bridgedotnet/Bridge/issues/3645 ```csharp public class App { public static void Main() { (string Prop1, string Prop2)? val = ("test1", "test2"); Console.WriteLine(val.Value.Prop1); Console.WriteLine(val?.Prop1); } } ``` ### Expected Result ``` > test1 > test1 ``` ### Actual Result ``` > test1 > ```
defect
null propagation does not work on nullable value tuples steps to reproduce related to csharp public class app public static void main string string val console writeline val value console writeline val expected result actual result
1
3,900
2,610,083,707
IssuesEvent
2015-02-26 18:25:33
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
深圳宝安哪里祛痘好
auto-migrated Priority-Medium Type-Defect
``` 深圳宝安哪里祛痘好【深圳韩方科颜全国热线400-869-1818,24小 时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国�� �方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩� ��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹” 健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专�� �治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的� ��痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:51
1.0
深圳宝安哪里祛痘好 - ``` 深圳宝安哪里祛痘好【深圳韩方科颜全国热线400-869-1818,24小 时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国�� �方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩� ��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹” 健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专�� �治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的� ��痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:51
defect
深圳宝安哪里祛痘好 深圳宝安哪里祛痘好【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国�� �方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩� ��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹” 健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专�� �治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的� ��痘。 original issue reported on code google com by szft com on may at
1
33,759
7,237,732,059
IssuesEvent
2018-02-13 12:09:09
oozcitak/exiflibrary
https://api.github.com/repos/oozcitak/exiflibrary
closed
Question on recompression
Priority-Medium Type-Defect auto-migrated
``` When i use the library to modify EXIF data ("comment" for example) and save, does it recompress the picture? Thanks Filip ``` Original issue reported on code.google.com by `Filip.Je...@gmail.com` on 27 Aug 2013 at 4:06
1.0
Question on recompression - ``` When i use the library to modify EXIF data ("comment" for example) and save, does it recompress the picture? Thanks Filip ``` Original issue reported on code.google.com by `Filip.Je...@gmail.com` on 27 Aug 2013 at 4:06
defect
question on recompression when i use the library to modify exif data comment for example and save does it recompress the picture thanks filip original issue reported on code google com by filip je gmail com on aug at
1
40,778
10,154,006,446
IssuesEvent
2019-08-06 06:50:43
line/centraldogma
https://api.github.com/repos/line/centraldogma
opened
Some server log messages are not logged to Logback
defect
I found this from the stderr output: ``` Aug 06, 2019 3:18:22 PM com.github.benmanes.caffeine.cache.LocalAsyncCache lambda$handleCompletion$3 WARNING: Exception thrown during asynchronous load java.util.concurrent.CompletionException: com.linecorp.centraldogma.server.internal.storage.RequestAlreadyTimedOutException: Request already timed out. at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319) at java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1702) at io.micrometer.core.instrument.AbstractTimer.record(AbstractTimer.java:160) at io.micrometer.core.instrument.Timer.lambda$wrap$0(Timer.java:144) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:834) Caused by: com.linecorp.centraldogma.server.internal.storage.RequestAlreadyTimedOutException: Request already timed out. ```
1.0
Some server log messages are not logged to Logback - I found this from the stderr output: ``` Aug 06, 2019 3:18:22 PM com.github.benmanes.caffeine.cache.LocalAsyncCache lambda$handleCompletion$3 WARNING: Exception thrown during asynchronous load java.util.concurrent.CompletionException: com.linecorp.centraldogma.server.internal.storage.RequestAlreadyTimedOutException: Request already timed out. at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319) at java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1702) at io.micrometer.core.instrument.AbstractTimer.record(AbstractTimer.java:160) at io.micrometer.core.instrument.Timer.lambda$wrap$0(Timer.java:144) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:834) Caused by: com.linecorp.centraldogma.server.internal.storage.RequestAlreadyTimedOutException: Request already timed out. ```
defect
some server log messages are not logged to logback i found this from the stderr output aug pm com github benmanes caffeine cache localasynccache lambda handlecompletion warning exception thrown during asynchronous load java util concurrent completionexception com linecorp centraldogma server internal storage requestalreadytimedoutexception request already timed out at java base java util concurrent completablefuture encodethrowable completablefuture java at java base java util concurrent completablefuture completethrowable completablefuture java at java base java util concurrent completablefuture asyncsupply run completablefuture java at io micrometer core instrument abstracttimer record abstracttimer java at io micrometer core instrument timer lambda wrap timer java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java caused by com linecorp centraldogma server internal storage requestalreadytimedoutexception request already timed out
1
265,087
8,337,106,047
IssuesEvent
2018-09-28 09:57:44
opencaching/opencaching-pl
https://api.github.com/repos/opencaching/opencaching-pl
closed
vars in translations
Component_i18n Priority_Low Type_Enhancement
There is need to handle vars in translation texts. "Old style" template system provides possibility of including vars as: `'activate_mail_subject' => 'New user registration - {site_name}',` Ideas: "new" way can handle it in "printf" style so the translation can look like: `'activate_mail_subject' => 'New user registration - %s',` and the code `tr('activate_mail_subject', $view->siteName)`
1.0
vars in translations - There is need to handle vars in translation texts. "Old style" template system provides possibility of including vars as: `'activate_mail_subject' => 'New user registration - {site_name}',` Ideas: "new" way can handle it in "printf" style so the translation can look like: `'activate_mail_subject' => 'New user registration - %s',` and the code `tr('activate_mail_subject', $view->siteName)`
non_defect
vars in translations there is need to handle vars in translation texts old style template system provides possibility of including vars as activate mail subject new user registration site name ideas new way can handle it in printf style so the translation can look like activate mail subject new user registration s and the code tr activate mail subject view sitename
0
231,981
7,647,517,067
IssuesEvent
2018-05-09 04:32:12
AdChain/AdChainRegistryDapp
https://api.github.com/repos/AdChain/AdChainRegistryDapp
closed
Committing a Vote- Step 2 in Walkthrough
Priority: Medium Status: Review Needed
"The filtered domains are all in the Voting Commit stage." <img width="1181" alt="screen shot 2018-04-03 at 10 46 45 am" src="https://user-images.githubusercontent.com/35276813/38266202-e60c0e20-372c-11e8-8f6f-c9de64eb4048.png"> However, the image shows domains in various other stages such as "In Registry", "Reveal-Pending", and "Application (Pending)".
1.0
Committing a Vote- Step 2 in Walkthrough - "The filtered domains are all in the Voting Commit stage." <img width="1181" alt="screen shot 2018-04-03 at 10 46 45 am" src="https://user-images.githubusercontent.com/35276813/38266202-e60c0e20-372c-11e8-8f6f-c9de64eb4048.png"> However, the image shows domains in various other stages such as "In Registry", "Reveal-Pending", and "Application (Pending)".
non_defect
committing a vote step in walkthrough the filtered domains are all in the voting commit stage img width alt screen shot at am src however the image shows domains in various other stages such as in registry reveal pending and application pending
0
307,095
9,414,186,019
IssuesEvent
2019-04-10 09:33:38
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Feature: button to open profile folder
Feature Medium Priority
Files from this folder regularry needs for support, playtime.eco need for first server configuration... And finding this folder is not easy. `%appdata%\..\LocalLow\Strange Loop Games\Eco`
1.0
Feature: button to open profile folder - Files from this folder regularry needs for support, playtime.eco need for first server configuration... And finding this folder is not easy. `%appdata%\..\LocalLow\Strange Loop Games\Eco`
non_defect
feature button to open profile folder files from this folder regularry needs for support playtime eco need for first server configuration and finding this folder is not easy appdata locallow strange loop games eco
0
71,138
9,480,507,364
IssuesEvent
2019-04-20 18:25:50
smithjd/sql-pet
https://api.github.com/repos/smithjd/sql-pet
closed
Investigate DiagrammeR integration with bookdown book
documentation
The tools of choice for creating diagrams in RStudio / R Markdown is the `DiagrammeR` R package (http://rich-iannone.github.io/DiagrammeR/index.html). 1. Diagrams are code - you can edit them and version-control them. 2. RStudio has provisions to view / render them as you build them. What I don't know (yet) is how one can integrate them into a bookdown book.
1.0
Investigate DiagrammeR integration with bookdown book - The tools of choice for creating diagrams in RStudio / R Markdown is the `DiagrammeR` R package (http://rich-iannone.github.io/DiagrammeR/index.html). 1. Diagrams are code - you can edit them and version-control them. 2. RStudio has provisions to view / render them as you build them. What I don't know (yet) is how one can integrate them into a bookdown book.
non_defect
investigate diagrammer integration with bookdown book the tools of choice for creating diagrams in rstudio r markdown is the diagrammer r package diagrams are code you can edit them and version control them rstudio has provisions to view render them as you build them what i don t know yet is how one can integrate them into a bookdown book
0
65,436
19,515,508,094
IssuesEvent
2021-12-29 09:32:13
ontop/ontop
https://api.github.com/repos/ontop/ontop
closed
bootstrap fails when there is a unique index that uses md5() of a column value
type: defect status: fixed w: db support
### Description bootstrap fails when there is a unique index that uses md5() of a column value with the following error: ``` 20:42:22.607 [main] DEBUG o.s.owlapi.utilities.Injector - Injecting values [[org.semanticweb.owlapi.rdf.rdfxml.renderer.RDFXMLStorerFactory@13e698c7, org.semanticweb.owlapi.functional.renderer.FunctionalSyntaxStorerFactory@267bbe1a, org.semanticweb.owlapi.manchestersyntax.renderer.ManchesterSyntaxStorerFactory@1f12e153, org.semanticweb.owlapi.rio.RioJsonStorerFactory@6eafb10e, org.semanticweb.owlapi.rio.RioNTriplesStorerFactory@2873d672, org.semanticweb.owlapi.dlsyntax.renderer.DLSyntaxHTMLStorerFactory@6b410923, org.semanticweb.owlapi.krss2.renderer.KRSS2OWLSyntaxStorerFactory@389562d6, org.semanticweb.owlapi.rio.RioNQuadsStorerFactory@464a4442, org.semanticweb.owlapi.rio.RioJsonLDStorerFactory@1d3e6d34, org.semanticweb.owlapi.rdf.turtle.renderer.TurtleStorerFactory@5a101b1c, org.semanticweb.owlapi.rio.RioN3StorerFactory@26a94fa5, org.semanticweb.owlapi.rio.RioTurtleStorerFactory@757529a4, org.semanticweb.owlapi.rio.RioTrixStorerFactory@4303b7f0, org.semanticweb.owlapi.rio.RioRDFXMLStorerFactory@3bc735b3, org.semanticweb.owlapi.latex.renderer.LatexStorerFactory@3a60c416, org.semanticweb.owlapi.rio.RioTrigStorerFactory@577f9109, org.semanticweb.owlapi.oboformat.OBOFormatStorerFactory@6d64b553, org.semanticweb.owlapi.rio.RioBinaryRdfStorerFactory@53667cbe, org.semanticweb.owlapi.owlxml.renderer.OWLXMLStorerFactory@aed0151, org.semanticweb.owlapi.dlsyntax.renderer.DLSyntaxStorerFactory@e8fadb0]] on method public void uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.setOntologyStorers(java.util.Set). Error occurred during bootstrapping: it.unibz.inf.ontop.exception.MetadataExtractionException: it.unibz.inf.ontop.dbschema.AttributeNotFoundException Debugging information for developers: it.unibz.inf.ontop.exception.MappingBootstrappingException: it.unibz.inf.ontop.exception.MetadataExtractionException: it.unibz.inf.ontop.dbschema.AttributeNotFoundException at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.bootstrapMappingAndOntology(DirectMappingEngine.java:148) at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.bootstrap(DirectMappingEngine.java:91) at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DefaultDirectMappingBootstrapper.bootstrap(DefaultDirectMappingBootstrapper.java:16) at it.unibz.inf.ontop.cli.OntopBootstrap.run(OntopBootstrap.java:56) at it.unibz.inf.ontop.cli.Ontop.main(Ontop.java:20) Caused by: it.unibz.inf.ontop.exception.MetadataExtractionException: it.unibz.inf.ontop.dbschema.AttributeNotFoundException at it.unibz.inf.ontop.dbschema.impl.AbstractDBMetadataProvider.insertUniqueAttributes(AbstractDBMetadataProvider.java:243) at it.unibz.inf.ontop.dbschema.impl.AbstractDBMetadataProvider.insertIntegrityConstraints(AbstractDBMetadataProvider.java:154) at it.unibz.inf.ontop.dbschema.impl.CachingMetadataLookup.extractImmutableMetadata(CachingMetadataLookup.java:44) at it.unibz.inf.ontop.dbschema.ImmutableMetadata.extractImmutableMetadata(ImmutableMetadata.java:17) at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.extractPPMapping(DirectMappingEngine.java:166) at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.bootstrapMappingAndOntology(DirectMappingEngine.java:123) ... 4 more Caused by: it.unibz.inf.ontop.dbschema.AttributeNotFoundException at it.unibz.inf.ontop.dbschema.impl.AbstractRelationDefinition.getAttribute(AbstractRelationDefinition.java:49) at it.unibz.inf.ontop.dbschema.impl.UniqueConstraintImpl$UniqueConstraintBuilder.addDeterminant(UniqueConstraintImpl.java:31) at it.unibz.inf.ontop.dbschema.impl.AbstractDBMetadataProvider.insertUniqueAttributes(AbstractDBMetadataProvider.java:234) ... 9 more ``` ### Steps to Reproduce load this into a postgres database and then run the ontop bootstrap command against it: ``` -- -- PostgreSQL database dump -- -- Dumped from database version 11.12 (Debian 11.12-1.pgdg90+1) -- Dumped by pg_dump version 11.12 (Debian 11.12-1.pgdg90+1) SET statement_timeout = 0; SET lock_timeout = 0; SET idle_in_transaction_session_timeout = 0; SET client_encoding = 'UTF8'; SET standard_conforming_strings = on; SELECT pg_catalog.set_config('search_path', '', false); SET check_function_bodies = false; SET xmloption = content; SET client_min_messages = warning; SET row_security = off; SET default_tablespace = ''; SET default_with_oids = false; -- -- Name: cat; Type: TABLE; Schema: public; Owner: postgres -- CREATE TABLE public.cat ( id integer NOT NULL, name character varying, age integer ); ALTER TABLE public.cat OWNER TO postgres; -- -- Data for Name: cat; Type: TABLE DATA; Schema: public; Owner: postgres -- COPY public.cat (id, name, age) FROM stdin; 5 freddo 12 \. -- -- Name: indycar; Type: INDEX; Schema: public; Owner: postgres -- CREATE UNIQUE INDEX indycar ON public.cat USING btree (age, md5((name)::text)); -- -- PostgreSQL database dump complete -- ``` **Expected behavior:** [What you expect to see] i expect to get a mapping file from the bootstrap process **Actual behavior:** [What actually happens] an error is emitted. but if you remove this ``` CREATE UNIQUE INDEX indycar ON public.cat USING btree (age, md5((name)::text)); ``` then the bootstrap is able to run successfully. **Reproduces how often:** [Does it happen every time?] always ### Versions the HEAD of the version4 branch. and i tried some older versions of ontop and had the same result.
1.0
bootstrap fails when there is a unique index that uses md5() of a column value - ### Description bootstrap fails when there is a unique index that uses md5() of a column value with the following error: ``` 20:42:22.607 [main] DEBUG o.s.owlapi.utilities.Injector - Injecting values [[org.semanticweb.owlapi.rdf.rdfxml.renderer.RDFXMLStorerFactory@13e698c7, org.semanticweb.owlapi.functional.renderer.FunctionalSyntaxStorerFactory@267bbe1a, org.semanticweb.owlapi.manchestersyntax.renderer.ManchesterSyntaxStorerFactory@1f12e153, org.semanticweb.owlapi.rio.RioJsonStorerFactory@6eafb10e, org.semanticweb.owlapi.rio.RioNTriplesStorerFactory@2873d672, org.semanticweb.owlapi.dlsyntax.renderer.DLSyntaxHTMLStorerFactory@6b410923, org.semanticweb.owlapi.krss2.renderer.KRSS2OWLSyntaxStorerFactory@389562d6, org.semanticweb.owlapi.rio.RioNQuadsStorerFactory@464a4442, org.semanticweb.owlapi.rio.RioJsonLDStorerFactory@1d3e6d34, org.semanticweb.owlapi.rdf.turtle.renderer.TurtleStorerFactory@5a101b1c, org.semanticweb.owlapi.rio.RioN3StorerFactory@26a94fa5, org.semanticweb.owlapi.rio.RioTurtleStorerFactory@757529a4, org.semanticweb.owlapi.rio.RioTrixStorerFactory@4303b7f0, org.semanticweb.owlapi.rio.RioRDFXMLStorerFactory@3bc735b3, org.semanticweb.owlapi.latex.renderer.LatexStorerFactory@3a60c416, org.semanticweb.owlapi.rio.RioTrigStorerFactory@577f9109, org.semanticweb.owlapi.oboformat.OBOFormatStorerFactory@6d64b553, org.semanticweb.owlapi.rio.RioBinaryRdfStorerFactory@53667cbe, org.semanticweb.owlapi.owlxml.renderer.OWLXMLStorerFactory@aed0151, org.semanticweb.owlapi.dlsyntax.renderer.DLSyntaxStorerFactory@e8fadb0]] on method public void uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.setOntologyStorers(java.util.Set). Error occurred during bootstrapping: it.unibz.inf.ontop.exception.MetadataExtractionException: it.unibz.inf.ontop.dbschema.AttributeNotFoundException Debugging information for developers: it.unibz.inf.ontop.exception.MappingBootstrappingException: it.unibz.inf.ontop.exception.MetadataExtractionException: it.unibz.inf.ontop.dbschema.AttributeNotFoundException at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.bootstrapMappingAndOntology(DirectMappingEngine.java:148) at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.bootstrap(DirectMappingEngine.java:91) at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DefaultDirectMappingBootstrapper.bootstrap(DefaultDirectMappingBootstrapper.java:16) at it.unibz.inf.ontop.cli.OntopBootstrap.run(OntopBootstrap.java:56) at it.unibz.inf.ontop.cli.Ontop.main(Ontop.java:20) Caused by: it.unibz.inf.ontop.exception.MetadataExtractionException: it.unibz.inf.ontop.dbschema.AttributeNotFoundException at it.unibz.inf.ontop.dbschema.impl.AbstractDBMetadataProvider.insertUniqueAttributes(AbstractDBMetadataProvider.java:243) at it.unibz.inf.ontop.dbschema.impl.AbstractDBMetadataProvider.insertIntegrityConstraints(AbstractDBMetadataProvider.java:154) at it.unibz.inf.ontop.dbschema.impl.CachingMetadataLookup.extractImmutableMetadata(CachingMetadataLookup.java:44) at it.unibz.inf.ontop.dbschema.ImmutableMetadata.extractImmutableMetadata(ImmutableMetadata.java:17) at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.extractPPMapping(DirectMappingEngine.java:166) at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.bootstrapMappingAndOntology(DirectMappingEngine.java:123) ... 4 more Caused by: it.unibz.inf.ontop.dbschema.AttributeNotFoundException at it.unibz.inf.ontop.dbschema.impl.AbstractRelationDefinition.getAttribute(AbstractRelationDefinition.java:49) at it.unibz.inf.ontop.dbschema.impl.UniqueConstraintImpl$UniqueConstraintBuilder.addDeterminant(UniqueConstraintImpl.java:31) at it.unibz.inf.ontop.dbschema.impl.AbstractDBMetadataProvider.insertUniqueAttributes(AbstractDBMetadataProvider.java:234) ... 9 more ``` ### Steps to Reproduce load this into a postgres database and then run the ontop bootstrap command against it: ``` -- -- PostgreSQL database dump -- -- Dumped from database version 11.12 (Debian 11.12-1.pgdg90+1) -- Dumped by pg_dump version 11.12 (Debian 11.12-1.pgdg90+1) SET statement_timeout = 0; SET lock_timeout = 0; SET idle_in_transaction_session_timeout = 0; SET client_encoding = 'UTF8'; SET standard_conforming_strings = on; SELECT pg_catalog.set_config('search_path', '', false); SET check_function_bodies = false; SET xmloption = content; SET client_min_messages = warning; SET row_security = off; SET default_tablespace = ''; SET default_with_oids = false; -- -- Name: cat; Type: TABLE; Schema: public; Owner: postgres -- CREATE TABLE public.cat ( id integer NOT NULL, name character varying, age integer ); ALTER TABLE public.cat OWNER TO postgres; -- -- Data for Name: cat; Type: TABLE DATA; Schema: public; Owner: postgres -- COPY public.cat (id, name, age) FROM stdin; 5 freddo 12 \. -- -- Name: indycar; Type: INDEX; Schema: public; Owner: postgres -- CREATE UNIQUE INDEX indycar ON public.cat USING btree (age, md5((name)::text)); -- -- PostgreSQL database dump complete -- ``` **Expected behavior:** [What you expect to see] i expect to get a mapping file from the bootstrap process **Actual behavior:** [What actually happens] an error is emitted. but if you remove this ``` CREATE UNIQUE INDEX indycar ON public.cat USING btree (age, md5((name)::text)); ``` then the bootstrap is able to run successfully. **Reproduces how often:** [Does it happen every time?] always ### Versions the HEAD of the version4 branch. and i tried some older versions of ontop and had the same result.
defect
bootstrap fails when there is a unique index that uses of a column value description bootstrap fails when there is a unique index that uses of a column value with the following error debug o s owlapi utilities injector injecting values on method public void uk ac manchester cs owl owlapi owlontologymanagerimpl setontologystorers java util set error occurred during bootstrapping it unibz inf ontop exception metadataextractionexception it unibz inf ontop dbschema attributenotfoundexception debugging information for developers it unibz inf ontop exception mappingbootstrappingexception it unibz inf ontop exception metadataextractionexception it unibz inf ontop dbschema attributenotfoundexception at it unibz inf ontop spec mapping bootstrap impl directmappingengine bootstrapmappingandontology directmappingengine java at it unibz inf ontop spec mapping bootstrap impl directmappingengine bootstrap directmappingengine java at it unibz inf ontop spec mapping bootstrap impl defaultdirectmappingbootstrapper bootstrap defaultdirectmappingbootstrapper java at it unibz inf ontop cli ontopbootstrap run ontopbootstrap java at it unibz inf ontop cli ontop main ontop java caused by it unibz inf ontop exception metadataextractionexception it unibz inf ontop dbschema attributenotfoundexception at it unibz inf ontop dbschema impl abstractdbmetadataprovider insertuniqueattributes abstractdbmetadataprovider java at it unibz inf ontop dbschema impl abstractdbmetadataprovider insertintegrityconstraints abstractdbmetadataprovider java at it unibz inf ontop dbschema impl cachingmetadatalookup extractimmutablemetadata cachingmetadatalookup java at it unibz inf ontop dbschema immutablemetadata extractimmutablemetadata immutablemetadata java at it unibz inf ontop spec mapping bootstrap impl directmappingengine extractppmapping directmappingengine java at it unibz inf ontop spec mapping bootstrap impl directmappingengine bootstrapmappingandontology directmappingengine java more caused by it unibz inf ontop dbschema attributenotfoundexception at it unibz inf ontop dbschema impl abstractrelationdefinition getattribute abstractrelationdefinition java at it unibz inf ontop dbschema impl uniqueconstraintimpl uniqueconstraintbuilder adddeterminant uniqueconstraintimpl java at it unibz inf ontop dbschema impl abstractdbmetadataprovider insertuniqueattributes abstractdbmetadataprovider java more steps to reproduce load this into a postgres database and then run the ontop bootstrap command against it postgresql database dump dumped from database version debian dumped by pg dump version debian set statement timeout set lock timeout set idle in transaction session timeout set client encoding set standard conforming strings on select pg catalog set config search path false set check function bodies false set xmloption content set client min messages warning set row security off set default tablespace set default with oids false name cat type table schema public owner postgres create table public cat id integer not null name character varying age integer alter table public cat owner to postgres data for name cat type table data schema public owner postgres copy public cat id name age from stdin freddo name indycar type index schema public owner postgres create unique index indycar on public cat using btree age name text postgresql database dump complete expected behavior i expect to get a mapping file from the bootstrap process actual behavior an error is emitted but if you remove this create unique index indycar on public cat using btree age name text then the bootstrap is able to run successfully reproduces how often always versions the head of the branch and i tried some older versions of ontop and had the same result
1
48,986
13,185,183,426
IssuesEvent
2020-08-12 20:53:20
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters (Trac #565)
Incomplete Migration Migrated from Trac defect tools/ports
<details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/565 , reported by kislat and owned by cgils</em></summary> <p> ```json { "status": "closed", "changetime": "2009-06-30T14:16:42", "description": "There is a bug in rootcint (ROOT 5.20.00 and later versions) that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable (after resolving symlinks) is longer than 128 characters.\n\nA patch for the file ROOT source file core/utils/src/rootcint.cxx for root_v5.20.00 is attached.", "reporter": "kislat", "cc": "", "resolution": "fixed", "_ts": "1246371402000000", "component": "tools/ports", "summary": "rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters", "priority": "normal", "keywords": "", "time": "2009-06-30T12:29:54", "milestone": "", "owner": "cgils", "type": "defect" } ``` </p> </details>
1.0
rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters (Trac #565) - <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/565 , reported by kislat and owned by cgils</em></summary> <p> ```json { "status": "closed", "changetime": "2009-06-30T14:16:42", "description": "There is a bug in rootcint (ROOT 5.20.00 and later versions) that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable (after resolving symlinks) is longer than 128 characters.\n\nA patch for the file ROOT source file core/utils/src/rootcint.cxx for root_v5.20.00 is attached.", "reporter": "kislat", "cc": "", "resolution": "fixed", "_ts": "1246371402000000", "component": "tools/ports", "summary": "rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters", "priority": "normal", "keywords": "", "time": "2009-06-30T12:29:54", "milestone": "", "owner": "cgils", "type": "defect" } ``` </p> </details>
defect
rootcint won t create dictionaries if rootsys is longer than characters trac migrated from reported by kislat and owned by cgils json status closed changetime description there is a bug in rootcint root and later versions that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable after resolving symlinks is longer than characters n na patch for the file root source file core utils src rootcint cxx for root is attached reporter kislat cc resolution fixed ts component tools ports summary rootcint won t create dictionaries if rootsys is longer than characters priority normal keywords time milestone owner cgils type defect
1
16,253
2,882,833,250
IssuesEvent
2015-06-11 08:25:32
ooskapenaar/awarenet
https://api.github.com/repos/ooskapenaar/awarenet
closed
chat doesn't work
auto-migrated Priority-Low Type-Defect
``` What steps will reproduce the problem? 1. visit "http://awarenet.eu/chat/" 2. click on "Global Chat" 3. What is the expected output? What do you see instead? I see this error message (see attached pic). It happend to me while using Firefox 31.0 (in VB), IE 11 and Opera 26. Its always the same error. Please use labels and text to provide additional information. ``` Original issue reported on code.google.com by `MrLordS...@gmail.com` on 16 Jan 2015 at 8:15 Attachments: * [error.png](https://storage.googleapis.com/google-code-attachments/awarenet/issue-395/comment-0/error.png)
1.0
chat doesn't work - ``` What steps will reproduce the problem? 1. visit "http://awarenet.eu/chat/" 2. click on "Global Chat" 3. What is the expected output? What do you see instead? I see this error message (see attached pic). It happend to me while using Firefox 31.0 (in VB), IE 11 and Opera 26. Its always the same error. Please use labels and text to provide additional information. ``` Original issue reported on code.google.com by `MrLordS...@gmail.com` on 16 Jan 2015 at 8:15 Attachments: * [error.png](https://storage.googleapis.com/google-code-attachments/awarenet/issue-395/comment-0/error.png)
defect
chat doesn t work what steps will reproduce the problem visit click on global chat what is the expected output what do you see instead i see this error message see attached pic it happend to me while using firefox in vb ie and opera its always the same error please use labels and text to provide additional information original issue reported on code google com by mrlords gmail com on jan at attachments
1
41,190
10,328,311,841
IssuesEvent
2019-09-02 09:14:11
vector-im/riot-web
https://api.github.com/repos/vector-im/riot-web
closed
Handle the case of no IS in features that require IS to lookup
bug defect phase:1 privacy privacy-sprint type:identity-server
Since it's now possible to be disconnected from an identity server entirely (your session has no active identity server), it's unclear how features that require an IS to function should behave. For the case of the Discovery section in Settings, it's easy enough: you have no IS, so we don't show any 3PIDs to control. However, there are other IS-using features in Riot, like: 1. Inviting by 3PID via the address picker <img width="792" alt="2019-08-23 at 16 02" src="https://user-images.githubusercontent.com/279572/63602530-8c9c7500-c5bf-11e9-8931-ec308e89de78.png"> 2. Inviting by 3PID via `/invite bob@bob.com` 3. ~~Checking an email invite in the invite preview bar to ensure the invited email is for the active account (in case you have multiple Matrix accounts in different tabs, etc.)~~ (Extracted feature 3 to https://github.com/vector-im/riot-web/issues/10669.) For features 1 and 2, I think we need some UX advice from @nadonomy on what they should do (communicate some kind of error to the user, etc.). For feature 3, I think we are at an earlier stage where @lampholder needs to work out how the product should function for this case.
1.0
Handle the case of no IS in features that require IS to lookup - Since it's now possible to be disconnected from an identity server entirely (your session has no active identity server), it's unclear how features that require an IS to function should behave. For the case of the Discovery section in Settings, it's easy enough: you have no IS, so we don't show any 3PIDs to control. However, there are other IS-using features in Riot, like: 1. Inviting by 3PID via the address picker <img width="792" alt="2019-08-23 at 16 02" src="https://user-images.githubusercontent.com/279572/63602530-8c9c7500-c5bf-11e9-8931-ec308e89de78.png"> 2. Inviting by 3PID via `/invite bob@bob.com` 3. ~~Checking an email invite in the invite preview bar to ensure the invited email is for the active account (in case you have multiple Matrix accounts in different tabs, etc.)~~ (Extracted feature 3 to https://github.com/vector-im/riot-web/issues/10669.) For features 1 and 2, I think we need some UX advice from @nadonomy on what they should do (communicate some kind of error to the user, etc.). For feature 3, I think we are at an earlier stage where @lampholder needs to work out how the product should function for this case.
defect
handle the case of no is in features that require is to lookup since it s now possible to be disconnected from an identity server entirely your session has no active identity server it s unclear how features that require an is to function should behave for the case of the discovery section in settings it s easy enough you have no is so we don t show any to control however there are other is using features in riot like inviting by via the address picker img width alt at src inviting by via invite bob bob com checking an email invite in the invite preview bar to ensure the invited email is for the active account in case you have multiple matrix accounts in different tabs etc extracted feature to for features and i think we need some ux advice from nadonomy on what they should do communicate some kind of error to the user etc for feature i think we are at an earlier stage where lampholder needs to work out how the product should function for this case
1
68,860
8,357,618,074
IssuesEvent
2018-10-02 22:15:00
nextcloud/server
https://api.github.com/repos/nextcloud/server
closed
Improve backup codes UI/UX
1. to develop Hacktoberfest design enhancement good first issue help wanted papercut
We recently added a new setting to the personal settings page – the ability to generate backup codes, which can be used in case users lose access to their second factor. At the moment, the user interface doesn't look that nice: ![bildschirmfoto von 2016-09-07 17-33-06](https://cloud.githubusercontent.com/assets/1374172/18318366/b9bbf462-7521-11e6-923e-8bad821984cd.png) ![bildschirmfoto von 2016-09-07 17-33-18](https://cloud.githubusercontent.com/assets/1374172/18318367/b9c0811c-7521-11e6-95a1-7b99406b71ff.png) ![bildschirmfoto von 2016-09-07 17-33-27](https://cloud.githubusercontent.com/assets/1374172/18318368/b9c2a7c6-7521-11e6-9fa5-3c8c6da61161.png) ![bildschirmfoto von 2016-09-07 17-33-29](https://cloud.githubusercontent.com/assets/1374172/18318369/b9c511dc-7521-11e6-9f4a-f35445eaa97f.png) @nextcloud/designers any idea how we could make a prettier user interface out of that? Would be great if someone could create a mockup :art: :rocket: ref https://github.com/nextcloud/server/pull/1171
1.0
Improve backup codes UI/UX - We recently added a new setting to the personal settings page – the ability to generate backup codes, which can be used in case users lose access to their second factor. At the moment, the user interface doesn't look that nice: ![bildschirmfoto von 2016-09-07 17-33-06](https://cloud.githubusercontent.com/assets/1374172/18318366/b9bbf462-7521-11e6-923e-8bad821984cd.png) ![bildschirmfoto von 2016-09-07 17-33-18](https://cloud.githubusercontent.com/assets/1374172/18318367/b9c0811c-7521-11e6-95a1-7b99406b71ff.png) ![bildschirmfoto von 2016-09-07 17-33-27](https://cloud.githubusercontent.com/assets/1374172/18318368/b9c2a7c6-7521-11e6-9fa5-3c8c6da61161.png) ![bildschirmfoto von 2016-09-07 17-33-29](https://cloud.githubusercontent.com/assets/1374172/18318369/b9c511dc-7521-11e6-9f4a-f35445eaa97f.png) @nextcloud/designers any idea how we could make a prettier user interface out of that? Would be great if someone could create a mockup :art: :rocket: ref https://github.com/nextcloud/server/pull/1171
non_defect
improve backup codes ui ux we recently added a new setting to the personal settings page – the ability to generate backup codes which can be used in case users lose access to their second factor at the moment the user interface doesn t look that nice nextcloud designers any idea how we could make a prettier user interface out of that would be great if someone could create a mockup art rocket ref
0
1,683
2,812,799,119
IssuesEvent
2015-05-18 11:03:28
facebook/fresco
https://api.github.com/repos/facebook/fresco
closed
Fresco + Realm.io library error
build duplicate needs-details
i use compile 'com.facebook.fresco:fresco:0.5.0+' java.lang.UnsatisfiedLinkError: dalvik.system.PathClassLoader[DexPathList[[zip file "/system/framework/multiwindow.jar", zip file "/system/framework/android.test.runner.jar", zip file "/data/app/com.couchgram.privacycall-1/base.apk"],nativeLibraryDirectories=[/data/app/com.couchgram.privacycall-1/lib/arm64, /vendor/lib64, /system/lib64]]] couldn't find "libmemchunk.so" at java.lang.Runtime.loadLibrary(Runtime.java:366) at java.lang.System.loadLibrary(System.java:989) at com.facebook.common.soloader.SoLoaderShim$DefaultHandler.loadLibrary(SoLoaderShim.java:34) at com.facebook.common.soloader.SoLoaderShim.loadLibrary(SoLoaderShim.java:56) at com.facebook.imagepipeline.memory.NativeMemoryChunk.<clinit>(NativeMemoryChunk.java:36) at com.facebook.imagepipeline.memory.NativeMemoryChunkPool.alloc(NativeMemoryChunkPool.java:60) at com.facebook.imagepipeline.memory.NativeMemoryChunkPool.alloc(NativeMemoryChunkPool.java:22) at com.facebook.imagepipeline.memory.BasePool.get(BasePool.java:260) at com.facebook.imagepipeline.memory.NativePooledByteBufferOutputStream.<init>(NativePooledByteBufferOutputStream.java:53) at com.facebook.imagepipeline.memory.NativePooledByteBufferOutputStream.<init>(NativePooledByteBufferOutputStream.java:36) at com.facebook.imagepipeline.memory.NativePooledByteBufferFactory.newByteBuffer(NativePooledByteBufferFactory.java:59) at com.facebook.imagepipeline.memory.NativePooledByteBufferFactory.newByteBuffer(NativePooledByteBufferFactory.java:26) at com.facebook.imagepipeline.producers.LocalFetchProducer$1.getResult(LocalFetchProducer.java:58) at com.facebook.imagepipeline.producers.LocalFetchProducer$1.getResult(LocalFetchProducer.java:49) at com.facebook.common.executors.StatefulRunnable.run(StatefulRunnable.java:45) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1112) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:587) at java.lang.Thread.run(Thread.java:818)
1.0
Fresco + Realm.io library error - i use compile 'com.facebook.fresco:fresco:0.5.0+' java.lang.UnsatisfiedLinkError: dalvik.system.PathClassLoader[DexPathList[[zip file "/system/framework/multiwindow.jar", zip file "/system/framework/android.test.runner.jar", zip file "/data/app/com.couchgram.privacycall-1/base.apk"],nativeLibraryDirectories=[/data/app/com.couchgram.privacycall-1/lib/arm64, /vendor/lib64, /system/lib64]]] couldn't find "libmemchunk.so" at java.lang.Runtime.loadLibrary(Runtime.java:366) at java.lang.System.loadLibrary(System.java:989) at com.facebook.common.soloader.SoLoaderShim$DefaultHandler.loadLibrary(SoLoaderShim.java:34) at com.facebook.common.soloader.SoLoaderShim.loadLibrary(SoLoaderShim.java:56) at com.facebook.imagepipeline.memory.NativeMemoryChunk.<clinit>(NativeMemoryChunk.java:36) at com.facebook.imagepipeline.memory.NativeMemoryChunkPool.alloc(NativeMemoryChunkPool.java:60) at com.facebook.imagepipeline.memory.NativeMemoryChunkPool.alloc(NativeMemoryChunkPool.java:22) at com.facebook.imagepipeline.memory.BasePool.get(BasePool.java:260) at com.facebook.imagepipeline.memory.NativePooledByteBufferOutputStream.<init>(NativePooledByteBufferOutputStream.java:53) at com.facebook.imagepipeline.memory.NativePooledByteBufferOutputStream.<init>(NativePooledByteBufferOutputStream.java:36) at com.facebook.imagepipeline.memory.NativePooledByteBufferFactory.newByteBuffer(NativePooledByteBufferFactory.java:59) at com.facebook.imagepipeline.memory.NativePooledByteBufferFactory.newByteBuffer(NativePooledByteBufferFactory.java:26) at com.facebook.imagepipeline.producers.LocalFetchProducer$1.getResult(LocalFetchProducer.java:58) at com.facebook.imagepipeline.producers.LocalFetchProducer$1.getResult(LocalFetchProducer.java:49) at com.facebook.common.executors.StatefulRunnable.run(StatefulRunnable.java:45) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1112) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:587) at java.lang.Thread.run(Thread.java:818)
non_defect
fresco realm io library error i use compile com facebook fresco fresco java lang unsatisfiedlinkerror dalvik system pathclassloader nativelibrarydirectories couldn t find libmemchunk so at java lang runtime loadlibrary runtime java at java lang system loadlibrary system java at com facebook common soloader soloadershim defaulthandler loadlibrary soloadershim java at com facebook common soloader soloadershim loadlibrary soloadershim java at com facebook imagepipeline memory nativememorychunk nativememorychunk java at com facebook imagepipeline memory nativememorychunkpool alloc nativememorychunkpool java at com facebook imagepipeline memory nativememorychunkpool alloc nativememorychunkpool java at com facebook imagepipeline memory basepool get basepool java at com facebook imagepipeline memory nativepooledbytebufferoutputstream nativepooledbytebufferoutputstream java at com facebook imagepipeline memory nativepooledbytebufferoutputstream nativepooledbytebufferoutputstream java at com facebook imagepipeline memory nativepooledbytebufferfactory newbytebuffer nativepooledbytebufferfactory java at com facebook imagepipeline memory nativepooledbytebufferfactory newbytebuffer nativepooledbytebufferfactory java at com facebook imagepipeline producers localfetchproducer getresult localfetchproducer java at com facebook imagepipeline producers localfetchproducer getresult localfetchproducer java at com facebook common executors statefulrunnable run statefulrunnable java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java
0
15,993
2,870,250,090
IssuesEvent
2015-06-07 00:33:52
pdelia/away3d
https://api.github.com/repos/pdelia/away3d
opened
Sprite3D Visual Glitches with Object3D when using Renderer.INTERSECTING_OBJECTS
auto-migrated Priority-Medium Type-Defect
#45 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:21Z ``` What steps will reproduce the problem? 1. view = new View3D({scene:scene, camera:camera, renderer:Renderer.INTERSECTING_OBJECTS}); 2. sprite = new Sprite2D(new Sprite_bmd(512, 512)); 3. cube = new Cube({width:2.5, height:2.5, depth:2.5, y:-1.25}); What is the expected output? What do you see instead? We should see a sprite cleanly cut by the cube at it's midway. Instead, the entire cube mesh begins to breakdown. This may be related to Issue35 http://code.google.com/p/away3d/issues/detail?id=35. Tests with deltaZ and ownSession will not solve the issue. The only available measure I've discovered to remedy the issue is to switch to Renderer.CORRECT_Z_ORDER, with some issues with Clipping. Perhaps Planes are the solution for the time being. What version of the product are you using? On what operating system? I'm publishing with FP9 and haven't tested the results in FP10. This is Mac OSX. Please provide any additional information below. ``` Original issue reported on code.google.com by `robinlhu...@gmail.com` on 1 Jun 2009 at 9:03
1.0
Sprite3D Visual Glitches with Object3D when using Renderer.INTERSECTING_OBJECTS - #45 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:21Z ``` What steps will reproduce the problem? 1. view = new View3D({scene:scene, camera:camera, renderer:Renderer.INTERSECTING_OBJECTS}); 2. sprite = new Sprite2D(new Sprite_bmd(512, 512)); 3. cube = new Cube({width:2.5, height:2.5, depth:2.5, y:-1.25}); What is the expected output? What do you see instead? We should see a sprite cleanly cut by the cube at it's midway. Instead, the entire cube mesh begins to breakdown. This may be related to Issue35 http://code.google.com/p/away3d/issues/detail?id=35. Tests with deltaZ and ownSession will not solve the issue. The only available measure I've discovered to remedy the issue is to switch to Renderer.CORRECT_Z_ORDER, with some issues with Clipping. Perhaps Planes are the solution for the time being. What version of the product are you using? On what operating system? I'm publishing with FP9 and haven't tested the results in FP10. This is Mac OSX. Please provide any additional information below. ``` Original issue reported on code.google.com by `robinlhu...@gmail.com` on 1 Jun 2009 at 9:03
defect
visual glitches with when using renderer intersecting objects issue by googlecodeexporter created on what steps will reproduce the problem view new scene scene camera camera renderer renderer intersecting objects sprite new new sprite bmd cube new cube width height depth y what is the expected output what do you see instead we should see a sprite cleanly cut by the cube at it s midway instead the entire cube mesh begins to breakdown this may be related to tests with deltaz and ownsession will not solve the issue the only available measure i ve discovered to remedy the issue is to switch to renderer correct z order with some issues with clipping perhaps planes are the solution for the time being what version of the product are you using on what operating system i m publishing with and haven t tested the results in this is mac osx please provide any additional information below original issue reported on code google com by robinlhu gmail com on jun at
1
74,549
25,170,310,908
IssuesEvent
2022-11-11 02:14:00
ascott18/TellMeWhen
https://api.github.com/repos/ascott18/TellMeWhen
closed
[Bug] Soul shard item count apparently doesn't work for Item Cooldown icons
T: defect V: classic
Reported by someone on Discord. Perhaps is due to different bag type for soul shard bags?
1.0
[Bug] Soul shard item count apparently doesn't work for Item Cooldown icons - Reported by someone on Discord. Perhaps is due to different bag type for soul shard bags?
defect
soul shard item count apparently doesn t work for item cooldown icons reported by someone on discord perhaps is due to different bag type for soul shard bags
1
17,887
3,013,563,736
IssuesEvent
2015-07-29 09:45:03
yawlfoundation/yawl
https://api.github.com/repos/yawlfoundation/yawl
closed
Can't save a specification with extended UTF-8 chars in data type definition
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Use for example the data type: <xs:simpleType name="country"> <xs:restriction base="xs:string"> <xs:enumeration value="ÅLAND ISLANDS" /> <xs:enumeration value="TEST" /> </xs:restriction> </xs:simpleType> in your specification. 2. Try to save the specification leads to "Could not export Specification due to missing or invalid user-definied datatype. Please see the problem list below for details." After rev 1755 there isn't an error in the data definition window anymore, but the handling of UTF-8 seems not to be correct. ``` Original issue reported on code.google.com by `felix.mannhardt` on 12 Aug 2011 at 3:21
1.0
Can't save a specification with extended UTF-8 chars in data type definition - ``` What steps will reproduce the problem? 1. Use for example the data type: <xs:simpleType name="country"> <xs:restriction base="xs:string"> <xs:enumeration value="ÅLAND ISLANDS" /> <xs:enumeration value="TEST" /> </xs:restriction> </xs:simpleType> in your specification. 2. Try to save the specification leads to "Could not export Specification due to missing or invalid user-definied datatype. Please see the problem list below for details." After rev 1755 there isn't an error in the data definition window anymore, but the handling of UTF-8 seems not to be correct. ``` Original issue reported on code.google.com by `felix.mannhardt` on 12 Aug 2011 at 3:21
defect
can t save a specification with extended utf chars in data type definition what steps will reproduce the problem use for example the data type in your specification try to save the specification leads to could not export specification due to missing or invalid user definied datatype please see the problem list below for details after rev there isn t an error in the data definition window anymore but the handling of utf seems not to be correct original issue reported on code google com by felix mannhardt on aug at
1
55,174
14,257,032,075
IssuesEvent
2020-11-20 02:32:13
naev/naev
https://api.github.com/repos/naev/naev
opened
make distcheck broken
Priority-High Type-Defect
``` $ make distcheck ... Making all in src Making all in tk Making all in widget CC button.o CC checkbox.o CC cust.o CC fader.o CC image.o CC imagearray.o CC input.o CC list.o CC rect.o CC tabwin.o CC text.o ar: `u' modifier ignored since `D' is the default (see `U') CC ai.o cc1: fatal error: ../src/attributes.h: No such file or directory compilation terminated. make[4]: *** [Makefile:925: ai.o] Error 1 make[3]: *** [Makefile:963: all-recursive] Error 1 make[2]: *** [Makefile:661: all] Error 2 make[1]: *** [Makefile:661: all-recursive] Error 1 make: *** [Makefile:879: distcheck] Error 1 ```
1.0
make distcheck broken - ``` $ make distcheck ... Making all in src Making all in tk Making all in widget CC button.o CC checkbox.o CC cust.o CC fader.o CC image.o CC imagearray.o CC input.o CC list.o CC rect.o CC tabwin.o CC text.o ar: `u' modifier ignored since `D' is the default (see `U') CC ai.o cc1: fatal error: ../src/attributes.h: No such file or directory compilation terminated. make[4]: *** [Makefile:925: ai.o] Error 1 make[3]: *** [Makefile:963: all-recursive] Error 1 make[2]: *** [Makefile:661: all] Error 2 make[1]: *** [Makefile:661: all-recursive] Error 1 make: *** [Makefile:879: distcheck] Error 1 ```
defect
make distcheck broken make distcheck making all in src making all in tk making all in widget cc button o cc checkbox o cc cust o cc fader o cc image o cc imagearray o cc input o cc list o cc rect o cc tabwin o cc text o ar u modifier ignored since d is the default see u cc ai o fatal error src attributes h no such file or directory compilation terminated make error make error make error make error make error
1
692,647
23,744,200,017
IssuesEvent
2022-08-31 14:42:51
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
ecerp01.ecil.gov.in - "Secure connection failed" when accessing the site
browser-firefox priority-normal severity-critical engine-gecko type-geolocation type-unsupported-tls
<!-- @browser: Firefox 78.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:78.0) Gecko/20100101 Firefox/78.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/53851 --> **URL**: https://ecerp01.ecil.gov.in/ecilerec **Browser / Version**: Firefox 78.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Edge **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: sap pop up code come and every thing come begining <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/6/07f6c5dc-741a-4cfc-b894-9007fa7929ba.jpeg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200604213430</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/6/9ec93488-11a7-407b-9786-45ce0b11ddd2) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
ecerp01.ecil.gov.in - "Secure connection failed" when accessing the site - <!-- @browser: Firefox 78.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:78.0) Gecko/20100101 Firefox/78.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/53851 --> **URL**: https://ecerp01.ecil.gov.in/ecilerec **Browser / Version**: Firefox 78.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Edge **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: sap pop up code come and every thing come begining <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/6/07f6c5dc-741a-4cfc-b894-9007fa7929ba.jpeg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200604213430</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/6/9ec93488-11a7-407b-9786-45ce0b11ddd2) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_defect
ecil gov in secure connection failed when accessing the site url browser version firefox operating system windows tested another browser yes edge problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce sap pop up code come and every thing come begining view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
22,783
3,698,519,087
IssuesEvent
2016-02-28 11:33:02
dnaumenko/java-diff-utils
https://api.github.com/repos/dnaumenko/java-diff-utils
closed
Please include a LICENSE file in the source tree
auto-migrated Priority-Medium Type-Defect
``` Including a LICENSE file makes it easier for downstream packagers and other users to ensure that they are using the software in a manner consistent with its license. Please include the Apache license in the source tree and in source distributions: http://www.apache.org/licenses/LICENSE-2.0 ``` Original issue reported on code.google.com by `willben...@gmail.com` on 21 Oct 2013 at 3:36
1.0
Please include a LICENSE file in the source tree - ``` Including a LICENSE file makes it easier for downstream packagers and other users to ensure that they are using the software in a manner consistent with its license. Please include the Apache license in the source tree and in source distributions: http://www.apache.org/licenses/LICENSE-2.0 ``` Original issue reported on code.google.com by `willben...@gmail.com` on 21 Oct 2013 at 3:36
defect
please include a license file in the source tree including a license file makes it easier for downstream packagers and other users to ensure that they are using the software in a manner consistent with its license please include the apache license in the source tree and in source distributions original issue reported on code google com by willben gmail com on oct at
1
71,659
23,748,071,691
IssuesEvent
2022-08-31 17:50:00
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Panel layout broke across restart
T-Defect
### Steps to reproduce 1. Minimise the spacepanel & leftpanel 2. Un-minimise them again 3. Restart the app (an upgrade, in this instance) 4. Observe that the app launches with minimised spacepanel & leftpanel ### Outcome #### What did you expect? Panel layout should persist across layouts #### What happened instead? Something failed to save when i expanded the panel layouts, it seems. ### Operating system macOS ### Application version nightly ### How did you install the app? nightly ### Homeserver matrix.org ### Will you send logs? No
1.0
Panel layout broke across restart - ### Steps to reproduce 1. Minimise the spacepanel & leftpanel 2. Un-minimise them again 3. Restart the app (an upgrade, in this instance) 4. Observe that the app launches with minimised spacepanel & leftpanel ### Outcome #### What did you expect? Panel layout should persist across layouts #### What happened instead? Something failed to save when i expanded the panel layouts, it seems. ### Operating system macOS ### Application version nightly ### How did you install the app? nightly ### Homeserver matrix.org ### Will you send logs? No
defect
panel layout broke across restart steps to reproduce minimise the spacepanel leftpanel un minimise them again restart the app an upgrade in this instance observe that the app launches with minimised spacepanel leftpanel outcome what did you expect panel layout should persist across layouts what happened instead something failed to save when i expanded the panel layouts it seems operating system macos application version nightly how did you install the app nightly homeserver matrix org will you send logs no
1
42,332
10,966,177,776
IssuesEvent
2019-11-28 06:08:22
vesoft-inc/nebula
https://api.github.com/repos/vesoft-inc/nebula
closed
Memory usage
defect-p3
![image](https://user-images.githubusercontent.com/13991775/69477534-5373db00-0e22-11ea-8e48-0a71907beada.png) nebula-storaged.conf: ![image](https://user-images.githubusercontent.com/13991775/69477680-fe38c900-0e23-11ea-85de-1d8ef3685663.png) partition: CREATE SPACE test(partition_num=1024, replica_factor=1); edge: 100 million **Your Environments(__must be provided__)** * OS: CentOS Linux release 7.4.1708
1.0
Memory usage - ![image](https://user-images.githubusercontent.com/13991775/69477534-5373db00-0e22-11ea-8e48-0a71907beada.png) nebula-storaged.conf: ![image](https://user-images.githubusercontent.com/13991775/69477680-fe38c900-0e23-11ea-85de-1d8ef3685663.png) partition: CREATE SPACE test(partition_num=1024, replica_factor=1); edge: 100 million **Your Environments(__must be provided__)** * OS: CentOS Linux release 7.4.1708
defect
memory usage nebula storaged conf: partition: create space test partition num replica factor edge: million your environments must be provided os centos linux release
1
4,737
2,610,153,974
IssuesEvent
2015-02-26 18:49:01
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
AT-TE
auto-migrated Priority-Medium Type-Defect
``` ATTE glides sometimes ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:08
1.0
AT-TE - ``` ATTE glides sometimes ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:08
defect
at te atte glides sometimes original issue reported on code google com by gmail com on jan at
1
448,756
31,811,398,823
IssuesEvent
2023-09-13 17:06:01
fmi-faim/napari-psf-analysis
https://api.github.com/repos/fmi-faim/napari-psf-analysis
opened
Improve UI design
documentation enhancement
* Add guides to improve plugin flow (e.g. divider) * Add plugin title, icon, link to documentation * Add hint if layers are missing
1.0
Improve UI design - * Add guides to improve plugin flow (e.g. divider) * Add plugin title, icon, link to documentation * Add hint if layers are missing
non_defect
improve ui design add guides to improve plugin flow e g divider add plugin title icon link to documentation add hint if layers are missing
0
255,823
8,126,543,129
IssuesEvent
2018-08-17 02:54:15
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
Add ability to specify up vector in x ray image query.
Expected Use: 3 - Occasional Feature Impact: 3 - Medium Priority: Normal
There is currently no way to set the up vector on the view normal. It sets the up vector somewhat arbitrarily. The user should really be able to set it. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1191 Status: Resolved Project: VisIt Tracker: Feature Priority: Normal Subject: Add ability to specify up vector in x ray image query. Assigned to: Kathleen Biagas Category: Target version: 2.6 Author: Eric Brugger Start: 09/05/2012 Due date: % Done: 0 Estimated time: 2.0 Created: 09/05/2012 11:40 am Updated: 10/17/2012 05:59 pm Likelihood: Severity: Found in version: Impact: 3 - Medium Expected Use: 3 - Occasional OS: All Support Group: Any Description: There is currently no way to set the up vector on the view normal. It sets the up vector somewhat arbitrarily. The user should really be able to set it. Comments: Added field to the XRay Image query widget. Added 'up_vector' parameter for setting via cli named arguments.M /src/avt/Filters/avtXRayFilter.hM /src/avt/Filters/avtXRayFilter.CM /src/avt/Queries/Queries/avtXRayImageQuery.CM /src/avt/Queries/Queries/avtXRayImageQuery.hM /src/gui/QvisXRayImageQueryWidget.CM /src/gui/QvisXRayImageQueryWidget.hM /src/visitpy/common/visitmodule.CM /src/resources/help/en_US/relnotes2.6.0.htmlM /test/tests/queries/xrayimage.pySVN Revisions 19382-19384.
1.0
Add ability to specify up vector in x ray image query. - There is currently no way to set the up vector on the view normal. It sets the up vector somewhat arbitrarily. The user should really be able to set it. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1191 Status: Resolved Project: VisIt Tracker: Feature Priority: Normal Subject: Add ability to specify up vector in x ray image query. Assigned to: Kathleen Biagas Category: Target version: 2.6 Author: Eric Brugger Start: 09/05/2012 Due date: % Done: 0 Estimated time: 2.0 Created: 09/05/2012 11:40 am Updated: 10/17/2012 05:59 pm Likelihood: Severity: Found in version: Impact: 3 - Medium Expected Use: 3 - Occasional OS: All Support Group: Any Description: There is currently no way to set the up vector on the view normal. It sets the up vector somewhat arbitrarily. The user should really be able to set it. Comments: Added field to the XRay Image query widget. Added 'up_vector' parameter for setting via cli named arguments.M /src/avt/Filters/avtXRayFilter.hM /src/avt/Filters/avtXRayFilter.CM /src/avt/Queries/Queries/avtXRayImageQuery.CM /src/avt/Queries/Queries/avtXRayImageQuery.hM /src/gui/QvisXRayImageQueryWidget.CM /src/gui/QvisXRayImageQueryWidget.hM /src/visitpy/common/visitmodule.CM /src/resources/help/en_US/relnotes2.6.0.htmlM /test/tests/queries/xrayimage.pySVN Revisions 19382-19384.
non_defect
add ability to specify up vector in x ray image query there is currently no way to set the up vector on the view normal it sets the up vector somewhat arbitrarily the user should really be able to set it redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker feature priority normal subject add ability to specify up vector in x ray image query assigned to kathleen biagas category target version author eric brugger start due date done estimated time created am updated pm likelihood severity found in version impact medium expected use occasional os all support group any description there is currently no way to set the up vector on the view normal it sets the up vector somewhat arbitrarily the user should really be able to set it comments added field to the xray image query widget added up vector parameter for setting via cli named arguments m src avt filters avtxrayfilter hm src avt filters avtxrayfilter cm src avt queries queries avtxrayimagequery cm src avt queries queries avtxrayimagequery hm src gui qvisxrayimagequerywidget cm src gui qvisxrayimagequerywidget hm src visitpy common visitmodule cm src resources help en us htmlm test tests queries xrayimage pysvn revisions
0
229,867
7,600,472,883
IssuesEvent
2018-04-28 02:33:49
adventurerscodex/uat
https://api.github.com/repos/adventurerscodex/uat
closed
test_preview_armor fails
browser-firefox severity-high-priority type-bug
def test_preview_armor(player_wizard, browser): # noqa """As a player, I can select a row in the armor table and view the item in the preview tab.""" print(('As a player, I can select a row in the armor table and view ' 'the item in the preview tab')) armor_add = armor.ArmorAddModal(browser) armor_table = armor.ArmorTable(browser) armor_preview = armor.ArmorPreviewModal(browser) tabs = Tabs(browser) tabs.equipment.click() WebDriverWait(browser, 10).until( EC.element_to_be_clickable( (By.ID, armor_table.add_id) ) ) armor_table.add.click() ut.select_from_autocomplete( armor_add, 'name', browser, has_search_term=False ) armor_add.add.click() WebDriverWait(browser, 10).until( modal_finished_closing(armor_add.modal_div_id) ) row = ut.get_table_row(armor_table, 'table', values=False) row[0].click() WebDriverWait(browser, 10).until( EC.text_to_be_present_in_element( > (By.ID, armor_preview.name_id), 'Breastplate' ) ) TIMEOUT Exception
1.0
test_preview_armor fails - def test_preview_armor(player_wizard, browser): # noqa """As a player, I can select a row in the armor table and view the item in the preview tab.""" print(('As a player, I can select a row in the armor table and view ' 'the item in the preview tab')) armor_add = armor.ArmorAddModal(browser) armor_table = armor.ArmorTable(browser) armor_preview = armor.ArmorPreviewModal(browser) tabs = Tabs(browser) tabs.equipment.click() WebDriverWait(browser, 10).until( EC.element_to_be_clickable( (By.ID, armor_table.add_id) ) ) armor_table.add.click() ut.select_from_autocomplete( armor_add, 'name', browser, has_search_term=False ) armor_add.add.click() WebDriverWait(browser, 10).until( modal_finished_closing(armor_add.modal_div_id) ) row = ut.get_table_row(armor_table, 'table', values=False) row[0].click() WebDriverWait(browser, 10).until( EC.text_to_be_present_in_element( > (By.ID, armor_preview.name_id), 'Breastplate' ) ) TIMEOUT Exception
non_defect
test preview armor fails def test preview armor player wizard browser noqa as a player i can select a row in the armor table and view the item in the preview tab print as a player i can select a row in the armor table and view the item in the preview tab armor add armor armoraddmodal browser armor table armor armortable browser armor preview armor armorpreviewmodal browser tabs tabs browser tabs equipment click webdriverwait browser until ec element to be clickable by id armor table add id armor table add click ut select from autocomplete armor add name browser has search term false armor add add click webdriverwait browser until modal finished closing armor add modal div id row ut get table row armor table table values false row click webdriverwait browser until ec text to be present in element by id armor preview name id breastplate timeout exception
0
35,371
7,721,055,227
IssuesEvent
2018-05-24 02:52:50
AtlasOfLivingAustralia/layers-service
https://api.github.com/repos/AtlasOfLivingAustralia/layers-service
closed
No download and no thumbnails on http://spatial.ala.org.au/layers#
priority-critical type-defect
Looks like neither of the downloads is working on http://spatial.ala.org.au/layers# and the list doesn't include the thumbnail.
1.0
No download and no thumbnails on http://spatial.ala.org.au/layers# - Looks like neither of the downloads is working on http://spatial.ala.org.au/layers# and the list doesn't include the thumbnail.
defect
no download and no thumbnails on looks like neither of the downloads is working on and the list doesn t include the thumbnail
1
30,106
6,025,660,231
IssuesEvent
2017-06-08 09:15:43
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
SelectOneButton: wrong unseletable behavior
defect
Currently if you click on a selected item, it will be unselected - also if unselectable is set to true.
1.0
SelectOneButton: wrong unseletable behavior - Currently if you click on a selected item, it will be unselected - also if unselectable is set to true.
defect
selectonebutton wrong unseletable behavior currently if you click on a selected item it will be unselected also if unselectable is set to true
1
70,124
22,956,984,080
IssuesEvent
2022-07-19 12:28:46
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
FileDownload/CommandButton inside facet of a datatable, not working at all
:lady_beetle: defect :bangbang: needs-triage
### Describe the bug In a dataTable I have facets for input and output. In a Blob field, for documents, I render a column with fileupload for input, all working there, and for output I render a commandButton with a fileDownload, no way to make it work. If I go for the actionListener route to prepare the Stream prior to be called by the filedownload component, it doesnt fire at all the actionlistener. Using process="@this" makes the actionlistenerr work, but crashes without throwing any error. If I try to process the blob and create de stream direct in the method called by the filedownload, nothing happens. This is the relevant sourc excerpt: ``` <p:column headerText="Documento"> <p:cellEditor id="conteudo"> <f:facet name="output"> <p:commandButton value="Baixar o documento" icon="pi pi-arrow-down" styleClass="ui-button-outlined" ajax="false"> <p:fileDownload value="#{cadastroEventoController.arquivo}"/> </p:commandButton> </f:facet> <f:facet name="input"> <p:fileUpload listener="#{cadastroEventoController.trataUploadDeDocumento}" mode="advanced" dragDropSupport="false" update="dtDocumentos msgs" sizeLimit="100000000" value="#{cadastroEventoController.arquivo}" fileLimit="1" > </p:fileUpload> </f:facet> </p:cellEditor> </p:column> ``` Using ActionListener ``` <p:commandButton value="Baixar o documento" icon="pi pi-arrow-down" styleClass="ui-button-outlined" process="@this" actionListener="#{cadastroEventoController.preparaArquivo()}" ajax="false"> <p:fileDownload value="#{cadastroEventoController.arquivo}"/> </p:commandButton> ``` Relevant methods: ``` // In Case direct use by filedownload public StreamedContent getArquivo() { FacesContext context = FacesContext.getCurrentInstance(); Integer linha = context.getApplication().evaluateExpressionGet(context, "#{linha}",Integer.class); DocumentoEntity doc = eventoDTO.getObjeto().getDocumentos().get(linha); ByteArrayInputStream stream = new ByteArrayInputStream(doc.getConteudo()); return new DefaultStreamedContent(stream, doc.getMime(), doc.getLegenda() + "." + doc.getExtensao()); } // In Case ActionListener works public void preparaArquivo() { FacesContext context = FacesContext.getCurrentInstance(); Integer linha = context.getApplication().evaluateExpressionGet(context, "#{linha}",Integer.class); DocumentoEntity doc = eventoDTO.getObjeto().getDocumentos().get(linha); ByteArrayInputStream stream = new ByteArrayInputStream(doc.getConteudo()); setArquivo(new DefaultStreamedContent(stream, doc.getMime(), doc.getLegenda() + "." + doc.getExtensao())); } ``` Extracting the CommandButton from the dataTable, but still inside the form, passing the index, makes any approach works to download the correct blob. ### Reproducer _No response_ ### Expected behavior I would expect the FileDownload, inside a CommandButton, inside a dataTable, to try to get de BLOB contents through the provided attribute and initiate the download. Behavior that works perfectly outside the dataTable. ### PrimeFaces edition Elite ### PrimeFaces version 8.0.16 ### Theme _No response_ ### JSF implementation _No response_ ### JSF version 2.3 ### Browser(s) _No response_
1.0
FileDownload/CommandButton inside facet of a datatable, not working at all - ### Describe the bug In a dataTable I have facets for input and output. In a Blob field, for documents, I render a column with fileupload for input, all working there, and for output I render a commandButton with a fileDownload, no way to make it work. If I go for the actionListener route to prepare the Stream prior to be called by the filedownload component, it doesnt fire at all the actionlistener. Using process="@this" makes the actionlistenerr work, but crashes without throwing any error. If I try to process the blob and create de stream direct in the method called by the filedownload, nothing happens. This is the relevant sourc excerpt: ``` <p:column headerText="Documento"> <p:cellEditor id="conteudo"> <f:facet name="output"> <p:commandButton value="Baixar o documento" icon="pi pi-arrow-down" styleClass="ui-button-outlined" ajax="false"> <p:fileDownload value="#{cadastroEventoController.arquivo}"/> </p:commandButton> </f:facet> <f:facet name="input"> <p:fileUpload listener="#{cadastroEventoController.trataUploadDeDocumento}" mode="advanced" dragDropSupport="false" update="dtDocumentos msgs" sizeLimit="100000000" value="#{cadastroEventoController.arquivo}" fileLimit="1" > </p:fileUpload> </f:facet> </p:cellEditor> </p:column> ``` Using ActionListener ``` <p:commandButton value="Baixar o documento" icon="pi pi-arrow-down" styleClass="ui-button-outlined" process="@this" actionListener="#{cadastroEventoController.preparaArquivo()}" ajax="false"> <p:fileDownload value="#{cadastroEventoController.arquivo}"/> </p:commandButton> ``` Relevant methods: ``` // In Case direct use by filedownload public StreamedContent getArquivo() { FacesContext context = FacesContext.getCurrentInstance(); Integer linha = context.getApplication().evaluateExpressionGet(context, "#{linha}",Integer.class); DocumentoEntity doc = eventoDTO.getObjeto().getDocumentos().get(linha); ByteArrayInputStream stream = new ByteArrayInputStream(doc.getConteudo()); return new DefaultStreamedContent(stream, doc.getMime(), doc.getLegenda() + "." + doc.getExtensao()); } // In Case ActionListener works public void preparaArquivo() { FacesContext context = FacesContext.getCurrentInstance(); Integer linha = context.getApplication().evaluateExpressionGet(context, "#{linha}",Integer.class); DocumentoEntity doc = eventoDTO.getObjeto().getDocumentos().get(linha); ByteArrayInputStream stream = new ByteArrayInputStream(doc.getConteudo()); setArquivo(new DefaultStreamedContent(stream, doc.getMime(), doc.getLegenda() + "." + doc.getExtensao())); } ``` Extracting the CommandButton from the dataTable, but still inside the form, passing the index, makes any approach works to download the correct blob. ### Reproducer _No response_ ### Expected behavior I would expect the FileDownload, inside a CommandButton, inside a dataTable, to try to get de BLOB contents through the provided attribute and initiate the download. Behavior that works perfectly outside the dataTable. ### PrimeFaces edition Elite ### PrimeFaces version 8.0.16 ### Theme _No response_ ### JSF implementation _No response_ ### JSF version 2.3 ### Browser(s) _No response_
defect
filedownload commandbutton inside facet of a datatable not working at all describe the bug in a datatable i have facets for input and output in a blob field for documents i render a column with fileupload for input all working there and for output i render a commandbutton with a filedownload no way to make it work if i go for the actionlistener route to prepare the stream prior to be called by the filedownload component it doesnt fire at all the actionlistener using process this makes the actionlistenerr work but crashes without throwing any error if i try to process the blob and create de stream direct in the method called by the filedownload nothing happens this is the relevant sourc excerpt p commandbutton value baixar o documento icon pi pi arrow down styleclass ui button outlined ajax false p fileupload listener cadastroeventocontroller tratauploaddedocumento mode advanced dragdropsupport false update dtdocumentos msgs sizelimit value cadastroeventocontroller arquivo filelimit using actionlistener p commandbutton value baixar o documento icon pi pi arrow down styleclass ui button outlined process this actionlistener cadastroeventocontroller preparaarquivo ajax false relevant methods in case direct use by filedownload public streamedcontent getarquivo facescontext context facescontext getcurrentinstance integer linha context getapplication evaluateexpressionget context linha integer class documentoentity doc eventodto getobjeto getdocumentos get linha bytearrayinputstream stream new bytearrayinputstream doc getconteudo return new defaultstreamedcontent stream doc getmime doc getlegenda doc getextensao in case actionlistener works public void preparaarquivo facescontext context facescontext getcurrentinstance integer linha context getapplication evaluateexpressionget context linha integer class documentoentity doc eventodto getobjeto getdocumentos get linha bytearrayinputstream stream new bytearrayinputstream doc getconteudo setarquivo new defaultstreamedcontent stream doc getmime doc getlegenda doc getextensao extracting the commandbutton from the datatable but still inside the form passing the index makes any approach works to download the correct blob reproducer no response expected behavior i would expect the filedownload inside a commandbutton inside a datatable to try to get de blob contents through the provided attribute and initiate the download behavior that works perfectly outside the datatable primefaces edition elite primefaces version theme no response jsf implementation no response jsf version browser s no response
1
276,960
30,581,322,092
IssuesEvent
2023-07-21 09:54:23
ministryofjustice/hmpps-probation-integration-services
https://api.github.com/repos/ministryofjustice/hmpps-probation-integration-services
closed
CVE-2023-34035 (pre-sentence-reports-to-delius)
dependencies security
Spring Security's authorization rules can be misconfigured when using multiple servlets * Project: pre-sentence-reports-to-delius * Package: `org.springframework.security:spring-security-config:6.1.1` * Location: `app/libs/spring-security-config-6.1.1.jar` >Spring Security versions 5.8 prior to 5.8.5, 6.0 prior to 6.0.5, and 6.1 prior to 6.1.2 could be susceptible to authorization rule misconfiguration if the application uses requestMatchers(String) and multiple servlets, one of them being Spring MVC’s DispatcherServlet. (DispatcherServlet is a Spring MVC component that maps HTTP endpoints to methods on @Controller-annotated classes.) Specifically, an application is vulnerable when all of the following are true: * Spring MVC is on the classpath * Spring Security is securing more than one servlet in a single application (one of them being Spring MVC’s DispatcherServlet) * The application uses requestMatchers(String) to refer to endpoints that are not Spring MVC endpoints An application is not vulnerable if any of the following is true: * The application does not have Spring MVC on the classpath * The application secures no servlets other than Spring MVC’s DispatcherServlet * The application uses requestMatchers(String) only for Spring MVC endpoints https://avd.aquasec.com/nvd/cve-2023-34035 If the vulnerability does not impact the `pre-sentence-reports-to-delius` project, you can suppress this alert by adding a comment starting with `Suppress`. For example, "Suppressed because we do not process any untrusted XML content".
True
CVE-2023-34035 (pre-sentence-reports-to-delius) - Spring Security's authorization rules can be misconfigured when using multiple servlets * Project: pre-sentence-reports-to-delius * Package: `org.springframework.security:spring-security-config:6.1.1` * Location: `app/libs/spring-security-config-6.1.1.jar` >Spring Security versions 5.8 prior to 5.8.5, 6.0 prior to 6.0.5, and 6.1 prior to 6.1.2 could be susceptible to authorization rule misconfiguration if the application uses requestMatchers(String) and multiple servlets, one of them being Spring MVC’s DispatcherServlet. (DispatcherServlet is a Spring MVC component that maps HTTP endpoints to methods on @Controller-annotated classes.) Specifically, an application is vulnerable when all of the following are true: * Spring MVC is on the classpath * Spring Security is securing more than one servlet in a single application (one of them being Spring MVC’s DispatcherServlet) * The application uses requestMatchers(String) to refer to endpoints that are not Spring MVC endpoints An application is not vulnerable if any of the following is true: * The application does not have Spring MVC on the classpath * The application secures no servlets other than Spring MVC’s DispatcherServlet * The application uses requestMatchers(String) only for Spring MVC endpoints https://avd.aquasec.com/nvd/cve-2023-34035 If the vulnerability does not impact the `pre-sentence-reports-to-delius` project, you can suppress this alert by adding a comment starting with `Suppress`. For example, "Suppressed because we do not process any untrusted XML content".
non_defect
cve pre sentence reports to delius spring security s authorization rules can be misconfigured when using multiple servlets project pre sentence reports to delius package org springframework security spring security config location app libs spring security config jar spring security versions  prior to  prior to  and  prior to  could be susceptible to authorization rule misconfiguration if the application uses requestmatchers string  and multiple servlets one of them being spring mvc’s dispatcherservlet   dispatcherservlet is a spring mvc component that maps http endpoints to methods on controller annotated classes specifically an application is vulnerable when all of the following are true spring mvc is on the classpath spring security is securing more than one servlet in a single application one of them being spring mvc’s dispatcherservlet the application uses requestmatchers string  to refer to endpoints that are not spring mvc endpoints an application is not vulnerable if any of the following is true the application does not have spring mvc on the classpath the application secures no servlets other than spring mvc’s dispatcherservlet the application uses requestmatchers string  only for spring mvc endpoints if the vulnerability does not impact the pre sentence reports to delius project you can suppress this alert by adding a comment starting with suppress for example suppressed because we do not process any untrusted xml content
0
434,990
30,480,764,155
IssuesEvent
2023-07-17 20:09:25
openxla/iree
https://api.github.com/repos/openxla/iree
closed
[New documentation] Write docs for "TensorFlow Lite bindings"
documentation bindings/java bindings/tflite
* Live preview of this page: https://scotttodd.github.io/iree/bindings/tensorflow-lite/, which will eventually live on https://google.github.io/iree/ after we switch over the docs. * ~~Source currently lives on the `staging-docs-rework` branch here: https://github.com/google/iree/blob/staging-docs-rework/mkdocs/docs/bindings/tensorflow-lite.md . PRs or direct commits to the branch are okay. We'll want to be careful about force pushes / rebases though.~~ Source lives in https://github.com/google/iree/blob/main/docs/website/docs/bindings/tensorflow-lite.md * See the README here for instructions on previewing changes locally: ~~https://github.com/google/iree/tree/staging-docs-rework/mkdocs~~ https://github.com/google/iree/tree/main/docs/website This will connect in some places with the corresponding frontend documentation (tracking issue: https://github.com/google/iree/issues/5455) and could discuss both the C API [bindings/tflite](https://github.com/google/iree/tree/main/bindings/tflite) and the Java API (https://github.com/google/iree/pull/5361). We should specifically speak to the pros/cons of using these bindings over the alternatives. If this is still being built / is unstable, we could omit the pages from the documentation site at first.
1.0
[New documentation] Write docs for "TensorFlow Lite bindings" - * Live preview of this page: https://scotttodd.github.io/iree/bindings/tensorflow-lite/, which will eventually live on https://google.github.io/iree/ after we switch over the docs. * ~~Source currently lives on the `staging-docs-rework` branch here: https://github.com/google/iree/blob/staging-docs-rework/mkdocs/docs/bindings/tensorflow-lite.md . PRs or direct commits to the branch are okay. We'll want to be careful about force pushes / rebases though.~~ Source lives in https://github.com/google/iree/blob/main/docs/website/docs/bindings/tensorflow-lite.md * See the README here for instructions on previewing changes locally: ~~https://github.com/google/iree/tree/staging-docs-rework/mkdocs~~ https://github.com/google/iree/tree/main/docs/website This will connect in some places with the corresponding frontend documentation (tracking issue: https://github.com/google/iree/issues/5455) and could discuss both the C API [bindings/tflite](https://github.com/google/iree/tree/main/bindings/tflite) and the Java API (https://github.com/google/iree/pull/5361). We should specifically speak to the pros/cons of using these bindings over the alternatives. If this is still being built / is unstable, we could omit the pages from the documentation site at first.
non_defect
write docs for tensorflow lite bindings live preview of this page which will eventually live on after we switch over the docs source currently lives on the staging docs rework branch here prs or direct commits to the branch are okay we ll want to be careful about force pushes rebases though source lives in see the readme here for instructions on previewing changes locally this will connect in some places with the corresponding frontend documentation tracking issue and could discuss both the c api and the java api we should specifically speak to the pros cons of using these bindings over the alternatives if this is still being built is unstable we could omit the pages from the documentation site at first
0
192,944
14,633,704,230
IssuesEvent
2020-12-24 02:50:33
MarlinFirmware/Marlin
https://api.github.com/repos/MarlinFirmware/Marlin
closed
[BUG] TMC2209 UART In dual Z mode Z2 (E1) changes direction randomly
Needs: Testing
### Bug Description In dual Z mode for MKS Robin Nano V2 with TMC2209 controlled over UART second Z motor attached to E1 from time to time changes its direction when new move command send (no matter if over LCD or via PROTERFACE gcommand) ### Configuration Files [ConfigHeaders.zip](https://github.com/MarlinFirmware/Marlin/files/5573281/ConfigHeaders.zip) I have tested it in CoreXY mode and regular (separated motor for x and y) ### Steps to Reproduce 1. Start pronterface / LCD menu 2. Send Z move command (up or down) - multiple time **Expected behavior:** Both motor rotate the same direction or opposite (if inverted) but stable. **Actual behavior:** From time to time second Z motor changes it direction (randomly)
1.0
[BUG] TMC2209 UART In dual Z mode Z2 (E1) changes direction randomly - ### Bug Description In dual Z mode for MKS Robin Nano V2 with TMC2209 controlled over UART second Z motor attached to E1 from time to time changes its direction when new move command send (no matter if over LCD or via PROTERFACE gcommand) ### Configuration Files [ConfigHeaders.zip](https://github.com/MarlinFirmware/Marlin/files/5573281/ConfigHeaders.zip) I have tested it in CoreXY mode and regular (separated motor for x and y) ### Steps to Reproduce 1. Start pronterface / LCD menu 2. Send Z move command (up or down) - multiple time **Expected behavior:** Both motor rotate the same direction or opposite (if inverted) but stable. **Actual behavior:** From time to time second Z motor changes it direction (randomly)
non_defect
uart in dual z mode changes direction randomly bug description in dual z mode for mks robin nano with controlled over uart second z motor attached to from time to time changes its direction when new move command send no matter if over lcd or via proterface gcommand configuration files i have tested it in corexy mode and regular separated motor for x and y steps to reproduce start pronterface lcd menu send z move command up or down multiple time expected behavior both motor rotate the same direction or opposite if inverted but stable actual behavior from time to time second z motor changes it direction randomly
0
179,816
21,581,943,932
IssuesEvent
2022-05-02 19:44:15
temporalio/temporal-ecommerce
https://api.github.com/repos/temporalio/temporal-ecommerce
closed
github.com/temporalio/sdk-go-v1.9.0: 2 vulnerabilities (highest severity is: 7.5) - autoclosed
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/temporalio/sdk-go-v1.9.0</b></p></summary> <p></p> <p> <p>Found in HEAD commit: <a href="https://github.com/temporalio/temporal-ecommerce/commit/91f5d7a169fd359bad9e33b954d19e6f8f7adac5">91f5d7a169fd359bad9e33b954d19e6f8f7adac5</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2019-0205](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | github.com/uber-go/tally-v3.3.17 | Transitive | N/A | &#10060; | | [CVE-2019-0210](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0210) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | github.com/uber-go/tally-v3.3.17 | Transitive | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-0205</summary> ### Vulnerable Library - <b>github.com/uber-go/tally-v3.3.17</b></p> <p>A Go metrics interface with fast buffered metrics and third party reporters</p> <p> Dependency Hierarchy: - github.com/temporalio/sdk-go-v1.9.0 (Root Library) - :x: **github.com/uber-go/tally-v3.3.17** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/temporal-ecommerce/commit/91f5d7a169fd359bad9e33b954d19e6f8f7adac5">91f5d7a169fd359bad9e33b954d19e6f8f7adac5</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Apache Thrift all versions up to and including 0.12.0, a server or client may run into an endless loop when feed with specific input data. Because the issue had already been partially fixed in version 0.11.0, depending on the installed version it affects only certain language bindings. <p>Publish Date: 2019-10-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205>CVE-2019-0205</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205</a></p> <p>Release Date: 2019-10-29</p> <p>Fix Resolution: org.apache.thrift:libthrift:0.13.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-0210</summary> ### Vulnerable Library - <b>github.com/uber-go/tally-v3.3.17</b></p> <p>A Go metrics interface with fast buffered metrics and third party reporters</p> <p> Dependency Hierarchy: - github.com/temporalio/sdk-go-v1.9.0 (Root Library) - :x: **github.com/uber-go/tally-v3.3.17** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/temporal-ecommerce/commit/91f5d7a169fd359bad9e33b954d19e6f8f7adac5">91f5d7a169fd359bad9e33b954d19e6f8f7adac5</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Apache Thrift 0.9.3 to 0.12.0, a server implemented in Go using TJSONProtocol or TSimpleJSONProtocol may panic when feed with invalid input data. <p>Publish Date: 2019-10-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0210>CVE-2019-0210</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://mail-archives.apache.org/mod_mbox/thrift-dev/201910.mbox/%3C277A46CA87494176B1BBCF5D72624A2A%40HAGGIS%3E">http://mail-archives.apache.org/mod_mbox/thrift-dev/201910.mbox/%3C277A46CA87494176B1BBCF5D72624A2A%40HAGGIS%3E</a></p> <p>Release Date: 2019-10-29</p> <p>Fix Resolution: 0.13.0</p> </p> <p></p> </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/uber-go/tally","packageVersion":"v3.3.17","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"github.com/temporalio/sdk-go:v1.9.0;github.com/uber-go/tally:v3.3.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.thrift:libthrift:0.13.0","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-0205","vulnerabilityDetails":"In Apache Thrift all versions up to and including 0.12.0, a server or client may run into an endless loop when feed with specific input data. Because the issue had already been partially fixed in version 0.11.0, depending on the installed version it affects only certain language bindings.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/uber-go/tally","packageVersion":"v3.3.17","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"github.com/temporalio/sdk-go:v1.9.0;github.com/uber-go/tally:v3.3.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.13.0","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-0210","vulnerabilityDetails":"In Apache Thrift 0.9.3 to 0.12.0, a server implemented in Go using TJSONProtocol or TSimpleJSONProtocol may panic when feed with invalid input data.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0210","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> -->
True
github.com/temporalio/sdk-go-v1.9.0: 2 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/temporalio/sdk-go-v1.9.0</b></p></summary> <p></p> <p> <p>Found in HEAD commit: <a href="https://github.com/temporalio/temporal-ecommerce/commit/91f5d7a169fd359bad9e33b954d19e6f8f7adac5">91f5d7a169fd359bad9e33b954d19e6f8f7adac5</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2019-0205](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | github.com/uber-go/tally-v3.3.17 | Transitive | N/A | &#10060; | | [CVE-2019-0210](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0210) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | github.com/uber-go/tally-v3.3.17 | Transitive | N/A | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-0205</summary> ### Vulnerable Library - <b>github.com/uber-go/tally-v3.3.17</b></p> <p>A Go metrics interface with fast buffered metrics and third party reporters</p> <p> Dependency Hierarchy: - github.com/temporalio/sdk-go-v1.9.0 (Root Library) - :x: **github.com/uber-go/tally-v3.3.17** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/temporal-ecommerce/commit/91f5d7a169fd359bad9e33b954d19e6f8f7adac5">91f5d7a169fd359bad9e33b954d19e6f8f7adac5</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Apache Thrift all versions up to and including 0.12.0, a server or client may run into an endless loop when feed with specific input data. Because the issue had already been partially fixed in version 0.11.0, depending on the installed version it affects only certain language bindings. <p>Publish Date: 2019-10-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205>CVE-2019-0205</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205</a></p> <p>Release Date: 2019-10-29</p> <p>Fix Resolution: org.apache.thrift:libthrift:0.13.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-0210</summary> ### Vulnerable Library - <b>github.com/uber-go/tally-v3.3.17</b></p> <p>A Go metrics interface with fast buffered metrics and third party reporters</p> <p> Dependency Hierarchy: - github.com/temporalio/sdk-go-v1.9.0 (Root Library) - :x: **github.com/uber-go/tally-v3.3.17** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/temporal-ecommerce/commit/91f5d7a169fd359bad9e33b954d19e6f8f7adac5">91f5d7a169fd359bad9e33b954d19e6f8f7adac5</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Apache Thrift 0.9.3 to 0.12.0, a server implemented in Go using TJSONProtocol or TSimpleJSONProtocol may panic when feed with invalid input data. <p>Publish Date: 2019-10-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0210>CVE-2019-0210</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://mail-archives.apache.org/mod_mbox/thrift-dev/201910.mbox/%3C277A46CA87494176B1BBCF5D72624A2A%40HAGGIS%3E">http://mail-archives.apache.org/mod_mbox/thrift-dev/201910.mbox/%3C277A46CA87494176B1BBCF5D72624A2A%40HAGGIS%3E</a></p> <p>Release Date: 2019-10-29</p> <p>Fix Resolution: 0.13.0</p> </p> <p></p> </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/uber-go/tally","packageVersion":"v3.3.17","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"github.com/temporalio/sdk-go:v1.9.0;github.com/uber-go/tally:v3.3.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.thrift:libthrift:0.13.0","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-0205","vulnerabilityDetails":"In Apache Thrift all versions up to and including 0.12.0, a server or client may run into an endless loop when feed with specific input data. Because the issue had already been partially fixed in version 0.11.0, depending on the installed version it affects only certain language bindings.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/uber-go/tally","packageVersion":"v3.3.17","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"github.com/temporalio/sdk-go:v1.9.0;github.com/uber-go/tally:v3.3.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.13.0","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-0210","vulnerabilityDetails":"In Apache Thrift 0.9.3 to 0.12.0, a server implemented in Go using TJSONProtocol or TSimpleJSONProtocol may panic when feed with invalid input data.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0210","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> -->
non_defect
github com temporalio sdk go vulnerabilities highest severity is autoclosed vulnerable library github com temporalio sdk go found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high github com uber go tally transitive n a high github com uber go tally transitive n a details cve vulnerable library github com uber go tally a go metrics interface with fast buffered metrics and third party reporters dependency hierarchy github com temporalio sdk go root library x github com uber go tally vulnerable library found in head commit a href found in base branch main vulnerability details in apache thrift all versions up to and including a server or client may run into an endless loop when feed with specific input data because the issue had already been partially fixed in version depending on the installed version it affects only certain language bindings publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache thrift libthrift cve vulnerable library github com uber go tally a go metrics interface with fast buffered metrics and third party reporters dependency hierarchy github com temporalio sdk go root library x github com uber go tally vulnerable library found in head commit a href found in base branch main vulnerability details in apache thrift to a server implemented in go using tjsonprotocol or tsimplejsonprotocol may panic when feed with invalid input data publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution istransitivedependency true dependencytree github com temporalio sdk go github com uber go tally isminimumfixversionavailable true minimumfixversion org apache thrift libthrift isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails in apache thrift all versions up to and including a server or client may run into an endless loop when feed with specific input data because the issue had already been partially fixed in version depending on the installed version it affects only certain language bindings vulnerabilityurl istransitivedependency true dependencytree github com temporalio sdk go github com uber go tally isminimumfixversionavailable true minimumfixversion isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails in apache thrift to a server implemented in go using tjsonprotocol or tsimplejsonprotocol may panic when feed with invalid input data vulnerabilityurl
0
14,790
25,732,942,363
IssuesEvent
2022-12-07 21:50:40
pypa/packaging
https://api.github.com/repos/pypa/packaging
closed
InvalidRequirement: name@ git+...@branch; (python_version=="...") and extra == "..."
bug packaging.requirements
Hello. I found a requirement that packaging deems invalid but I am not sure why. This is happening on 20.4, 20.9 and main branch @ 73507464faca545d83b748929e60ec67c0cb50b1. I see what I believe is a proper requirement generated by https://github.com/sqlobject/sqlobject/blob/3.9.1/setup.py#L118-L121 as: ``` [oursql:python_version=="2.7"] oursql@ git+https://github.com/sqlobject/oursql.git@master [oursql:python_version>="3.4"] oursql3@ git+https://github.com/sqlobject/oursql.git@py3k ``` Parsed by `importlib.metadata.PathDistribution.requires` as: ``` oursql@ git+https://github.com/sqlobject/oursql.git@master; (python_version=="2.7") and extra == "oursql" oursql3@ git+https://github.com/sqlobject/oursql.git@py3k; (python_version>="3.4") and extra == "oursql" ``` Failing with: ``` >>> from packaging.requirements import Requirement >>> Requirement('oursql@ git+https://github.com/sqlobject/oursql.git@master; (python_version=="2.7") and extra == "oursql"') Traceback (most recent call last): File "/home/churchyard/Dokumenty/pypa/packaging/packaging/requirements.py", line 102, in __init__ req = REQUIREMENT.parseString(requirement_string) File "/usr/lib/python3.9/site-packages/pyparsing.py", line 1955, in parseString raise exc File "/usr/lib/python3.9/site-packages/pyparsing.py", line 3814, in parseImpl raise ParseException(instring, loc, self.errmsg, self) pyparsing.ParseException: Expected stringEnd, found '(' (at char 60), (line:1, col:61) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/churchyard/Dokumenty/pypa/packaging/packaging/requirements.py", line 104, in __init__ raise InvalidRequirement( packaging.requirements.InvalidRequirement: Parse error at "'(python_'": Expected stringEnd ``` What's wrong with this requirement?
1.0
InvalidRequirement: name@ git+...@branch; (python_version=="...") and extra == "..." - Hello. I found a requirement that packaging deems invalid but I am not sure why. This is happening on 20.4, 20.9 and main branch @ 73507464faca545d83b748929e60ec67c0cb50b1. I see what I believe is a proper requirement generated by https://github.com/sqlobject/sqlobject/blob/3.9.1/setup.py#L118-L121 as: ``` [oursql:python_version=="2.7"] oursql@ git+https://github.com/sqlobject/oursql.git@master [oursql:python_version>="3.4"] oursql3@ git+https://github.com/sqlobject/oursql.git@py3k ``` Parsed by `importlib.metadata.PathDistribution.requires` as: ``` oursql@ git+https://github.com/sqlobject/oursql.git@master; (python_version=="2.7") and extra == "oursql" oursql3@ git+https://github.com/sqlobject/oursql.git@py3k; (python_version>="3.4") and extra == "oursql" ``` Failing with: ``` >>> from packaging.requirements import Requirement >>> Requirement('oursql@ git+https://github.com/sqlobject/oursql.git@master; (python_version=="2.7") and extra == "oursql"') Traceback (most recent call last): File "/home/churchyard/Dokumenty/pypa/packaging/packaging/requirements.py", line 102, in __init__ req = REQUIREMENT.parseString(requirement_string) File "/usr/lib/python3.9/site-packages/pyparsing.py", line 1955, in parseString raise exc File "/usr/lib/python3.9/site-packages/pyparsing.py", line 3814, in parseImpl raise ParseException(instring, loc, self.errmsg, self) pyparsing.ParseException: Expected stringEnd, found '(' (at char 60), (line:1, col:61) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/churchyard/Dokumenty/pypa/packaging/packaging/requirements.py", line 104, in __init__ raise InvalidRequirement( packaging.requirements.InvalidRequirement: Parse error at "'(python_'": Expected stringEnd ``` What's wrong with this requirement?
non_defect
invalidrequirement name git branch python version and extra hello i found a requirement that packaging deems invalid but i am not sure why this is happening on and main branch i see what i believe is a proper requirement generated by as oursql git git parsed by importlib metadata pathdistribution requires as oursql git python version and extra oursql git python version and extra oursql failing with from packaging requirements import requirement requirement oursql git python version and extra oursql traceback most recent call last file home churchyard dokumenty pypa packaging packaging requirements py line in init req requirement parsestring requirement string file usr lib site packages pyparsing py line in parsestring raise exc file usr lib site packages pyparsing py line in parseimpl raise parseexception instring loc self errmsg self pyparsing parseexception expected stringend found at char line col during handling of the above exception another exception occurred traceback most recent call last file line in file home churchyard dokumenty pypa packaging packaging requirements py line in init raise invalidrequirement packaging requirements invalidrequirement parse error at python expected stringend what s wrong with this requirement
0
186,758
15,083,338,636
IssuesEvent
2021-02-05 15:42:37
CyVerse-learning-materials/foss
https://api.github.com/repos/CyVerse-learning-materials/foss
opened
Fix nav bar to match course schedule from top to bottom
documentation
Related to issue #29 - [ ] Resolve Issue #29 links to remove - [ ] Reorder pages to match syntax of agenda - [ ] Check github repo link at the bottom of the page.
1.0
Fix nav bar to match course schedule from top to bottom - Related to issue #29 - [ ] Resolve Issue #29 links to remove - [ ] Reorder pages to match syntax of agenda - [ ] Check github repo link at the bottom of the page.
non_defect
fix nav bar to match course schedule from top to bottom related to issue resolve issue links to remove reorder pages to match syntax of agenda check github repo link at the bottom of the page
0
33,664
7,195,147,657
IssuesEvent
2018-02-04 14:15:44
ganeti/ganeti
https://api.github.com/repos/ganeti/ganeti
closed
gnt-backup does not attach ext-based snapshot
AutoMigrated Component-storage Status:Fixed Type-Defect
Originally reported of Google Code with ID 1198. ``` What software version are you running? Please provide the output of "gnt- cluster --version", "gnt-cluster version", and "hspace --version". gnt-cluster (ganeti 2.15.2-99ds) 2.15.2 hspace (ganeti) version 2.15.2-99ds compiled with ghc 7.6 running on linux x86_64 <b>What distribution are you using?</b> Debian Jessie 8.6 <b>What steps will reproduce the problem?</b> 1. gnt-instance start INSTANCE 2. wait until instance is fully started. 3. gnt-backup -n NODE --noshutdown INSTANCE <b>What is the expected output? What do you see instead?</b> Exporting...finished. <b>Please provide any additional information below.</b> I use my own extstorage provider which I released last year to public here... https://code.google.com/p/ganeti/wiki/ExtStorageProviders (see bottom > cLVM and "Shared" LVM). Below is the error I get. Fri Dec 16 09:52:47 2016 Creating a snapshot of disk/0 on node xen01.rkt-zentral.local Fri Dec 16 09:52:48 2016 Exporting snapshot/0 from xen01.rkt-zentral.local to xen01.rkt-zentral.local Fri Dec 16 09:52:51 2016 snapshot/0 is now listening, starting export Fri Dec 16 09:52:52 2016 - WARNING: export on xen01.rkt-zentral.local failed: Failed to start export on xen01.rkt-zentral.local: Block device '<Disk(type=ext, logical_id=('lvm', '1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap'), children=None, visible as /dev/disk/0, size=10240m)>' is not set up Fri Dec 16 09:52:52 2016 snapshot/0 failed to send data: Failed to start export on xen01.rkt-zentral.local: Block device '<Disk(type=ext, logical_id=('lvm', '1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap'), children=None, visible as /dev/disk/0, size=10240m)>' is not set up (recent output: None) Fri Dec 16 09:52:52 2016 Removing snapshot of disk/0 on node xen01.rkt-zentral.local Fri Dec 16 09:52:53 2016 - WARNING: Aborting import 'import-disk0-2016-12-16_09_52_48-WYeFav' on b661ad19-f12d-4631-9d8e-b913a68494cf Fri Dec 16 09:52:53 2016 - WARNING: import 'import-disk0-2016-12-16_09_52_48-WYeFav' on xen01.rkt-zentral.local failed: Exited due to signal 15 Fri Dec 16 09:52:53 2016 snapshot/0 failed to receive data: Exited due to signal 15 (recent output: socat: W exiting on signal 15) Fri Dec 16 09:52:53 2016 - WARNING: Some disk exports have failed; there may be leftover data for instance debian.rkt-zentral.local on node xen01.rkt-zentral.local All executables in this provider write into a log file /usr/share/ganeti/extstorage/lvm/var.log. There you can see, that the snapshot is created successfully, but then the snapshot cannot be found because gnt-backup seems not to export the VolumegroupName to the attach executable. Here is a snippet... >>>> --- attach rktstorage8 1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0 --- snapshot rktstorage8 1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0 10240 1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap --- attach rktstorage8 1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0 --- attach 1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap --- attach 1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap <<<<<< Here you can see that when it comes to attaching the snapshot "1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap" the parameter "vgname" is empty. Using gnt-backup without "--noshutdown" works, because no snapshot has to be taken. Please help. Kind regards Darius ``` Originally added on 2016-12-16 10:14:20 +0000 UTC.
1.0
gnt-backup does not attach ext-based snapshot - Originally reported of Google Code with ID 1198. ``` What software version are you running? Please provide the output of "gnt- cluster --version", "gnt-cluster version", and "hspace --version". gnt-cluster (ganeti 2.15.2-99ds) 2.15.2 hspace (ganeti) version 2.15.2-99ds compiled with ghc 7.6 running on linux x86_64 <b>What distribution are you using?</b> Debian Jessie 8.6 <b>What steps will reproduce the problem?</b> 1. gnt-instance start INSTANCE 2. wait until instance is fully started. 3. gnt-backup -n NODE --noshutdown INSTANCE <b>What is the expected output? What do you see instead?</b> Exporting...finished. <b>Please provide any additional information below.</b> I use my own extstorage provider which I released last year to public here... https://code.google.com/p/ganeti/wiki/ExtStorageProviders (see bottom > cLVM and "Shared" LVM). Below is the error I get. Fri Dec 16 09:52:47 2016 Creating a snapshot of disk/0 on node xen01.rkt-zentral.local Fri Dec 16 09:52:48 2016 Exporting snapshot/0 from xen01.rkt-zentral.local to xen01.rkt-zentral.local Fri Dec 16 09:52:51 2016 snapshot/0 is now listening, starting export Fri Dec 16 09:52:52 2016 - WARNING: export on xen01.rkt-zentral.local failed: Failed to start export on xen01.rkt-zentral.local: Block device '<Disk(type=ext, logical_id=('lvm', '1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap'), children=None, visible as /dev/disk/0, size=10240m)>' is not set up Fri Dec 16 09:52:52 2016 snapshot/0 failed to send data: Failed to start export on xen01.rkt-zentral.local: Block device '<Disk(type=ext, logical_id=('lvm', '1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap'), children=None, visible as /dev/disk/0, size=10240m)>' is not set up (recent output: None) Fri Dec 16 09:52:52 2016 Removing snapshot of disk/0 on node xen01.rkt-zentral.local Fri Dec 16 09:52:53 2016 - WARNING: Aborting import 'import-disk0-2016-12-16_09_52_48-WYeFav' on b661ad19-f12d-4631-9d8e-b913a68494cf Fri Dec 16 09:52:53 2016 - WARNING: import 'import-disk0-2016-12-16_09_52_48-WYeFav' on xen01.rkt-zentral.local failed: Exited due to signal 15 Fri Dec 16 09:52:53 2016 snapshot/0 failed to receive data: Exited due to signal 15 (recent output: socat: W exiting on signal 15) Fri Dec 16 09:52:53 2016 - WARNING: Some disk exports have failed; there may be leftover data for instance debian.rkt-zentral.local on node xen01.rkt-zentral.local All executables in this provider write into a log file /usr/share/ganeti/extstorage/lvm/var.log. There you can see, that the snapshot is created successfully, but then the snapshot cannot be found because gnt-backup seems not to export the VolumegroupName to the attach executable. Here is a snippet... >>>> --- attach rktstorage8 1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0 --- snapshot rktstorage8 1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0 10240 1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap --- attach rktstorage8 1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0 --- attach 1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap --- attach 1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap <<<<<< Here you can see that when it comes to attaching the snapshot "1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap" the parameter "vgname" is empty. Using gnt-backup without "--noshutdown" works, because no snapshot has to be taken. Please help. Kind regards Darius ``` Originally added on 2016-12-16 10:14:20 +0000 UTC.
defect
gnt backup does not attach ext based snapshot originally reported of google code with id what software version are you running please provide the output of gnt cluster version gnt cluster version and hspace version gnt cluster ganeti hspace ganeti version compiled with ghc running on linux what distribution are you using debian jessie what steps will reproduce the problem gnt instance start instance wait until instance is fully started gnt backup n node noshutdown instance what is the expected output what do you see instead exporting finished please provide any additional information below i use my own extstorage provider which i released last year to public here see bottom clvm and shared lvm below is the error i get fri dec creating a snapshot of disk on node rkt zentral local fri dec exporting snapshot from rkt zentral local to rkt zentral local fri dec snapshot is now listening starting export fri dec warning export on rkt zentral local failed failed to start export on rkt zentral local block device is not set up fri dec snapshot failed to send data failed to start export on rkt zentral local block device is not set up recent output none fri dec removing snapshot of disk on node rkt zentral local fri dec warning aborting import import wyefav on fri dec warning import import wyefav on rkt zentral local failed exited due to signal fri dec snapshot failed to receive data exited due to signal recent output socat w exiting on signal fri dec warning some disk exports have failed there may be leftover data for instance debian rkt zentral local on node rkt zentral local all executables in this provider write into a log file usr share ganeti extstorage lvm var log there you can see that the snapshot is created successfully but then the snapshot cannot be found because gnt backup seems not to export the volumegroupname to the attach executable here is a snippet attach ext snapshot ext ext snap attach ext attach ext snap attach ext snap here you can see that when it comes to attaching the snapshot ext snap the parameter vgname is empty using gnt backup without noshutdown works because no snapshot has to be taken please help kind regards darius originally added on utc
1
49,643
13,187,244,688
IssuesEvent
2020-08-13 02:48:19
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
[icetray] uninitialized value (Trac #1799)
Incomplete Migration Migrated from Trac combo core defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1799">https://code.icecube.wisc.edu/ticket/1799</a>, reported by kjmeagher and owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:12:38", "description": "found by static analysis: http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-581265.html#EndPath", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "_ts": "1550067158057333", "component": "combo core", "summary": "[icetray] uninitialized value", "priority": "normal", "keywords": "", "time": "2016-07-27T08:01:06", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
[icetray] uninitialized value (Trac #1799) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1799">https://code.icecube.wisc.edu/ticket/1799</a>, reported by kjmeagher and owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:12:38", "description": "found by static analysis: http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-581265.html#EndPath", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "_ts": "1550067158057333", "component": "combo core", "summary": "[icetray] uninitialized value", "priority": "normal", "keywords": "", "time": "2016-07-27T08:01:06", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
uninitialized value trac migrated from json status closed changetime description found by static analysis reporter kjmeagher cc resolution fixed ts component combo core summary uninitialized value priority normal keywords time milestone owner olivas type defect
1
49,600
13,187,238,627
IssuesEvent
2020-08-13 02:47:11
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
[tableio] named argument in class declaration (Trac #1734)
Incomplete Migration Migrated from Trac cmake defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1734">https://code.icecube.wisc.edu/ticket/1734</a>, reported by kjmeagher and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2016-06-10T08:01:03", "description": "The sphinx build gives the following error:\n{{{\nTraceback (most recent call last):\n File \"/private/var/folders/rc/g_4_lyp9039cj1586zzg88f40000gn/T/pip-build-A327aa/sphinx/sphinx/ext/autodoc.py\", line 385, in import_object\n File \"/Users/kmeagher/icecube/combo/release/lib/icecube/tableio/enum3.py\", line 43\n class enum(baseEnum, metaclass=metaEnum):\n ^\nSyntaxError: invalid syntax\n}}}\n", "reporter": "kjmeagher", "cc": "", "resolution": "wontfix", "_ts": "1465545663829039", "component": "cmake", "summary": "[tableio] named argument in class declaration", "priority": "normal", "keywords": "", "time": "2016-06-10T07:25:29", "milestone": "Long-Term Future", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
1.0
[tableio] named argument in class declaration (Trac #1734) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1734">https://code.icecube.wisc.edu/ticket/1734</a>, reported by kjmeagher and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2016-06-10T08:01:03", "description": "The sphinx build gives the following error:\n{{{\nTraceback (most recent call last):\n File \"/private/var/folders/rc/g_4_lyp9039cj1586zzg88f40000gn/T/pip-build-A327aa/sphinx/sphinx/ext/autodoc.py\", line 385, in import_object\n File \"/Users/kmeagher/icecube/combo/release/lib/icecube/tableio/enum3.py\", line 43\n class enum(baseEnum, metaclass=metaEnum):\n ^\nSyntaxError: invalid syntax\n}}}\n", "reporter": "kjmeagher", "cc": "", "resolution": "wontfix", "_ts": "1465545663829039", "component": "cmake", "summary": "[tableio] named argument in class declaration", "priority": "normal", "keywords": "", "time": "2016-06-10T07:25:29", "milestone": "Long-Term Future", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
defect
named argument in class declaration trac migrated from json status closed changetime description the sphinx build gives the following error n ntraceback most recent call last n file private var folders rc g t pip build sphinx sphinx ext autodoc py line in import object n file users kmeagher icecube combo release lib icecube tableio py line n class enum baseenum metaclass metaenum n nsyntaxerror invalid syntax n n reporter kjmeagher cc resolution wontfix ts component cmake summary named argument in class declaration priority normal keywords time milestone long term future owner jvansanten type defect
1
12,381
2,694,263,997
IssuesEvent
2015-04-01 19:16:21
google/google-api-go-client
https://api.github.com/repos/google/google-api-go-client
opened
Support for Standard Query Parameters (fields, quotaUser, userIp, etc) in Google APIs
new priority-medium type-defect
**shazow** on 5 Sep 2013 at 4:40: ``` Most (if not all) Google API's support standard query parameters, such as ones documented here: https://developers.google.com/discovery/v1/parameters https://developers.google.com/drive/query-parameters https://developers.google.com/analytics/devguides/reporting/realtime/v3/paramete rs Etc. There doesn't seem to be a way to set these using the google-api-go-client library. Would be great if there was native support for these fields. Perhaps it would make sense letting the user set arbitrary fields to the opt_ field map, for future workarounds to such limitations? Related bugs: #5 https://code.google.com/p/google-api-go-client/issues/detail?id=5 #38 https://code.google.com/p/google-api-go-client/issues/detail?id=38 I would consider this bug to be a superset of the other two. Thanks! ```
1.0
Support for Standard Query Parameters (fields, quotaUser, userIp, etc) in Google APIs - **shazow** on 5 Sep 2013 at 4:40: ``` Most (if not all) Google API's support standard query parameters, such as ones documented here: https://developers.google.com/discovery/v1/parameters https://developers.google.com/drive/query-parameters https://developers.google.com/analytics/devguides/reporting/realtime/v3/paramete rs Etc. There doesn't seem to be a way to set these using the google-api-go-client library. Would be great if there was native support for these fields. Perhaps it would make sense letting the user set arbitrary fields to the opt_ field map, for future workarounds to such limitations? Related bugs: #5 https://code.google.com/p/google-api-go-client/issues/detail?id=5 #38 https://code.google.com/p/google-api-go-client/issues/detail?id=38 I would consider this bug to be a superset of the other two. Thanks! ```
defect
support for standard query parameters fields quotauser userip etc in google apis shazow on sep at most if not all google api s support standard query parameters such as ones documented here rs etc there doesn t seem to be a way to set these using the google api go client library would be great if there was native support for these fields perhaps it would make sense letting the user set arbitrary fields to the opt field map for future workarounds to such limitations related bugs i would consider this bug to be a superset of the other two thanks
1
81,941
31,825,594,527
IssuesEvent
2023-09-14 07:17:27
DependencyTrack/dependency-track
https://api.github.com/repos/DependencyTrack/dependency-track
opened
Unable to upload sbom file
defect in triage
### Current Behavior When I try to upload the SBOM file via UI (Upload BOM), i get a successfully updated message. However i don't see any component on the project. While analysing the API log i found following error message. `2023-09-14 07:07:52,138 ERROR [BomUploadProcessingTask] Error while processing bom org.cyclonedx.exception.ParseException: Unable to parse BOM from byte array at org.cyclonedx.parsers.JsonParser.parse(JsonParser.java:70) at org.dependencytrack.tasks.BomUploadProcessingTask.inform(BomUploadProcessingTask.java:105) at alpine.event.framework.BaseEventService.lambda$publish$0(BaseEventService.java:101) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) Caused by: com.fasterxml.jackson.databind.exc.MismatchedInputException: Cannot deserialize value of type `java.util.ArrayList<org.cyclonedx.model.Tool>` from Object value (token `JsonToken.START_OBJECT`) at [Source: (byte[])"{ "bomFormat": "CycloneDX", "specVersion": "1.5", "serialNumber": "urn:uuid:a4da8087-25bb-4b89-98b9-f2b6c368d6a1", "version": 1, "metadata": { "timestamp": "2023-09-14T05:24:09.218Z", "tools": { "components": [ { "group": "@cyclonedx", "name": "cdxgen", "version": "9.3.1", "purl": "pkg:npm/%40cyclonedx/cdxgen@9.3.1", "type": "application", "bom-ref": "pkg:npm/@cyclonedx/cdxgen@9.3.1" } ] "[truncated 46201 bytes]; line: 8, column: 14] (through reference chain: org.cyclonedx.model.Bom["metadata"]->org.cyclonedx.model.Metadata["tools"]) at com.fasterxml.jackson.databind.exc.MismatchedInputException.from(MismatchedInputException.java:59) at com.fasterxml.jackson.databind.DeserializationContext.reportInputMismatch(DeserializationContext.java:1746) at com.fasterxml.jackson.databind.DeserializationContext.handleUnexpectedToken(DeserializationContext.java:1520) at com.fasterxml.jackson.databind.DeserializationContext.handleUnexpectedToken(DeserializationContext.java:1467) at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.handleNonArray(CollectionDeserializer.java:396) at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.deserialize(CollectionDeserializer.java:252) at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.deserialize(CollectionDeserializer.java:28) at com.fasterxml.jackson.databind.deser.impl.MethodProperty.deserializeAndSet(MethodProperty.java:129) at com.fasterxml.jackson.databind.deser.BeanDeserializer.vanillaDeserialize(BeanDeserializer.java:314) at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:177) at com.fasterxml.jackson.databind.deser.impl.MethodProperty.deserializeAndSet(MethodProperty.java:129) at com.fasterxml.jackson.databind.deser.BeanDeserializer.vanillaDeserialize(BeanDeserializer.java:314) at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:177) at com.fasterxml.jackson.databind.deser.DefaultDeserializationContext.readRootValue(DefaultDeserializationContext.java:323) at com.fasterxml.jackson.databind.ObjectMapper._readMapAndClose(ObjectMapper.java:4730) at com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3738) at org.cyclonedx.parsers.JsonParser.parse(JsonParser.java:68) ... 5 common frames omitted 2023-09-14 07:07:52,175 DEBUG [NotificationService] Dispatching notification: class alpine.notification.Notification 2023-09-14 07:07:52,175 DEBUG [NotificationService] Alerting subscriber org.dependencytrack.notification.NotificationRouter 2023-09-14 07:07:52,175 DEBUG [EventService] Dispatching event: class org.dependencytrack.event.IndexEvent 2023-09-14 07:07:52,175 DEBUG [EventService] Alerting subscriber org.dependencytrack.tasks.IndexTask 2023-09-14 07:07:52,179 DEBUG [EventService] Dispatching event: class org.dependencytrack.event.IndexEvent 2023-09-14 07:07:52,179 DEBUG [EventService] Alerting subscriber org.dependencytrack.tasks.IndexTask 2023-09-14 07:07:52,179 DEBUG [EventService] Calling onSuccess 2023-09-14 07:07:52,179 DEBUG [EventService] Calling onSuccess 2023-09-14 07:07:52,185 DEBUG [EventService] Calling onSuccess` ### Steps to Reproduce 1. Login into DependencyTrack UI 2. Navigate to the Project > Components 3. Click on Upload BOM 4. Select the BOM file (file [attached)](url) [bom.zip](https://github.com/DependencyTrack/dependency-track/files/12605879/bom.zip) 5. Click Upload, a success message will appear on the UI but the upload will fail at the api level ### Expected Behavior BOM file need to uploaded and processed successfully ### Dependency-Track Version 4.8.2 ### Dependency-Track Distribution Container Image ### Database Server PostgreSQL ### Database Server Version _No response_ ### Browser Google Chrome ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
1.0
Unable to upload sbom file - ### Current Behavior When I try to upload the SBOM file via UI (Upload BOM), i get a successfully updated message. However i don't see any component on the project. While analysing the API log i found following error message. `2023-09-14 07:07:52,138 ERROR [BomUploadProcessingTask] Error while processing bom org.cyclonedx.exception.ParseException: Unable to parse BOM from byte array at org.cyclonedx.parsers.JsonParser.parse(JsonParser.java:70) at org.dependencytrack.tasks.BomUploadProcessingTask.inform(BomUploadProcessingTask.java:105) at alpine.event.framework.BaseEventService.lambda$publish$0(BaseEventService.java:101) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) Caused by: com.fasterxml.jackson.databind.exc.MismatchedInputException: Cannot deserialize value of type `java.util.ArrayList<org.cyclonedx.model.Tool>` from Object value (token `JsonToken.START_OBJECT`) at [Source: (byte[])"{ "bomFormat": "CycloneDX", "specVersion": "1.5", "serialNumber": "urn:uuid:a4da8087-25bb-4b89-98b9-f2b6c368d6a1", "version": 1, "metadata": { "timestamp": "2023-09-14T05:24:09.218Z", "tools": { "components": [ { "group": "@cyclonedx", "name": "cdxgen", "version": "9.3.1", "purl": "pkg:npm/%40cyclonedx/cdxgen@9.3.1", "type": "application", "bom-ref": "pkg:npm/@cyclonedx/cdxgen@9.3.1" } ] "[truncated 46201 bytes]; line: 8, column: 14] (through reference chain: org.cyclonedx.model.Bom["metadata"]->org.cyclonedx.model.Metadata["tools"]) at com.fasterxml.jackson.databind.exc.MismatchedInputException.from(MismatchedInputException.java:59) at com.fasterxml.jackson.databind.DeserializationContext.reportInputMismatch(DeserializationContext.java:1746) at com.fasterxml.jackson.databind.DeserializationContext.handleUnexpectedToken(DeserializationContext.java:1520) at com.fasterxml.jackson.databind.DeserializationContext.handleUnexpectedToken(DeserializationContext.java:1467) at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.handleNonArray(CollectionDeserializer.java:396) at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.deserialize(CollectionDeserializer.java:252) at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.deserialize(CollectionDeserializer.java:28) at com.fasterxml.jackson.databind.deser.impl.MethodProperty.deserializeAndSet(MethodProperty.java:129) at com.fasterxml.jackson.databind.deser.BeanDeserializer.vanillaDeserialize(BeanDeserializer.java:314) at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:177) at com.fasterxml.jackson.databind.deser.impl.MethodProperty.deserializeAndSet(MethodProperty.java:129) at com.fasterxml.jackson.databind.deser.BeanDeserializer.vanillaDeserialize(BeanDeserializer.java:314) at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:177) at com.fasterxml.jackson.databind.deser.DefaultDeserializationContext.readRootValue(DefaultDeserializationContext.java:323) at com.fasterxml.jackson.databind.ObjectMapper._readMapAndClose(ObjectMapper.java:4730) at com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3738) at org.cyclonedx.parsers.JsonParser.parse(JsonParser.java:68) ... 5 common frames omitted 2023-09-14 07:07:52,175 DEBUG [NotificationService] Dispatching notification: class alpine.notification.Notification 2023-09-14 07:07:52,175 DEBUG [NotificationService] Alerting subscriber org.dependencytrack.notification.NotificationRouter 2023-09-14 07:07:52,175 DEBUG [EventService] Dispatching event: class org.dependencytrack.event.IndexEvent 2023-09-14 07:07:52,175 DEBUG [EventService] Alerting subscriber org.dependencytrack.tasks.IndexTask 2023-09-14 07:07:52,179 DEBUG [EventService] Dispatching event: class org.dependencytrack.event.IndexEvent 2023-09-14 07:07:52,179 DEBUG [EventService] Alerting subscriber org.dependencytrack.tasks.IndexTask 2023-09-14 07:07:52,179 DEBUG [EventService] Calling onSuccess 2023-09-14 07:07:52,179 DEBUG [EventService] Calling onSuccess 2023-09-14 07:07:52,185 DEBUG [EventService] Calling onSuccess` ### Steps to Reproduce 1. Login into DependencyTrack UI 2. Navigate to the Project > Components 3. Click on Upload BOM 4. Select the BOM file (file [attached)](url) [bom.zip](https://github.com/DependencyTrack/dependency-track/files/12605879/bom.zip) 5. Click Upload, a success message will appear on the UI but the upload will fail at the api level ### Expected Behavior BOM file need to uploaded and processed successfully ### Dependency-Track Version 4.8.2 ### Dependency-Track Distribution Container Image ### Database Server PostgreSQL ### Database Server Version _No response_ ### Browser Google Chrome ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
defect
unable to upload sbom file current behavior when i try to upload the sbom file via ui upload bom i get a successfully updated message however i don t see any component on the project while analysing the api log i found following error message error error while processing bom org cyclonedx exception parseexception unable to parse bom from byte array at org cyclonedx parsers jsonparser parse jsonparser java at org dependencytrack tasks bomuploadprocessingtask inform bomuploadprocessingtask java at alpine event framework baseeventservice lambda publish baseeventservice java at java base java util concurrent threadpoolexecutor runworker unknown source at java base java util concurrent threadpoolexecutor worker run unknown source at java base java lang thread run unknown source caused by com fasterxml jackson databind exc mismatchedinputexception cannot deserialize value of type java util arraylist from object value token jsontoken start object at bomformat cyclonedx specversion serialnumber urn uuid version metadata timestamp tools components group cyclonedx name cdxgen version purl pkg npm cdxgen type application bom ref pkg npm cyclonedx cdxgen line column through reference chain org cyclonedx model bom org cyclonedx model metadata at com fasterxml jackson databind exc mismatchedinputexception from mismatchedinputexception java at com fasterxml jackson databind deserializationcontext reportinputmismatch deserializationcontext java at com fasterxml jackson databind deserializationcontext handleunexpectedtoken deserializationcontext java at com fasterxml jackson databind deserializationcontext handleunexpectedtoken deserializationcontext java at com fasterxml jackson databind deser std collectiondeserializer handlenonarray collectiondeserializer java at com fasterxml jackson databind deser std collectiondeserializer deserialize collectiondeserializer java at com fasterxml jackson databind deser std collectiondeserializer deserialize collectiondeserializer java at com fasterxml jackson databind deser impl methodproperty deserializeandset methodproperty java at com fasterxml jackson databind deser beandeserializer vanilladeserialize beandeserializer java at com fasterxml jackson databind deser beandeserializer deserialize beandeserializer java at com fasterxml jackson databind deser impl methodproperty deserializeandset methodproperty java at com fasterxml jackson databind deser beandeserializer vanilladeserialize beandeserializer java at com fasterxml jackson databind deser beandeserializer deserialize beandeserializer java at com fasterxml jackson databind deser defaultdeserializationcontext readrootvalue defaultdeserializationcontext java at com fasterxml jackson databind objectmapper readmapandclose objectmapper java at com fasterxml jackson databind objectmapper readvalue objectmapper java at org cyclonedx parsers jsonparser parse jsonparser java common frames omitted debug dispatching notification class alpine notification notification debug alerting subscriber org dependencytrack notification notificationrouter debug dispatching event class org dependencytrack event indexevent debug alerting subscriber org dependencytrack tasks indextask debug dispatching event class org dependencytrack event indexevent debug alerting subscriber org dependencytrack tasks indextask debug calling onsuccess debug calling onsuccess debug calling onsuccess steps to reproduce login into dependencytrack ui navigate to the project components click on upload bom select the bom file file url click upload a success message will appear on the ui but the upload will fail at the api level expected behavior bom file need to uploaded and processed successfully dependency track version dependency track distribution container image database server postgresql database server version no response browser google chrome checklist i have read and understand the i have checked the for whether this defect was already reported
1
66,143
20,016,796,631
IssuesEvent
2022-02-01 12:55:18
BOINC/boinc
https://api.github.com/repos/BOINC/boinc
closed
Problem with <max_concurrent>0</max_concurrent>
C: Client - Scheduler Policy P: Undetermined R: invalid T: Defect
**Describe the bug** I have set this app_config, but boinc still computes rosetta app in addition to rosetta_python_projects [app_config.txt](https://github.com/BOINC/boinc/files/7966612/app_config.txt) **Steps To Reproduce** 1. 2. **Expected behavior** <max_concurrent>0</max_concurrent> should disallow computing of given app **Screenshots** **System Information** - OS: Microsoft Windows [Version 10.0.22000.469] - BOINC Version: 7.16.20 **Additional context**
1.0
Problem with <max_concurrent>0</max_concurrent> - **Describe the bug** I have set this app_config, but boinc still computes rosetta app in addition to rosetta_python_projects [app_config.txt](https://github.com/BOINC/boinc/files/7966612/app_config.txt) **Steps To Reproduce** 1. 2. **Expected behavior** <max_concurrent>0</max_concurrent> should disallow computing of given app **Screenshots** **System Information** - OS: Microsoft Windows [Version 10.0.22000.469] - BOINC Version: 7.16.20 **Additional context**
defect
problem with describe the bug i have set this app config but boinc still computes rosetta app in addition to rosetta python projects steps to reproduce expected behavior should disallow computing of given app screenshots system information os microsoft windows boinc version additional context
1
308,109
26,577,984,845
IssuesEvent
2023-01-22 03:18:30
pola-rs/polars
https://api.github.com/repos/pola-rs/polars
closed
Use pytest markers to distinguish between test types
python test
### Problem description I think the [custom marker functionality](https://docs.pytest.org/en/7.1.x/example/markers.html#marking-test-functions-and-selecting-them-for-a-run) from pytest could be nice for us. We could mark parametric/slow tests explicitly, rather than relying on folder structure. You can apply a mark to [all tests](https://docs.pytest.org/en/7.1.x/example/markers.html#marking-whole-classes-or-modules) in a module easily. Might be a worthwhile improvement.
1.0
Use pytest markers to distinguish between test types - ### Problem description I think the [custom marker functionality](https://docs.pytest.org/en/7.1.x/example/markers.html#marking-test-functions-and-selecting-them-for-a-run) from pytest could be nice for us. We could mark parametric/slow tests explicitly, rather than relying on folder structure. You can apply a mark to [all tests](https://docs.pytest.org/en/7.1.x/example/markers.html#marking-whole-classes-or-modules) in a module easily. Might be a worthwhile improvement.
non_defect
use pytest markers to distinguish between test types problem description i think the from pytest could be nice for us we could mark parametric slow tests explicitly rather than relying on folder structure you can apply a mark to in a module easily might be a worthwhile improvement
0
17,020
2,966,744,197
IssuesEvent
2015-07-12 06:45:59
ElvisChiang/welly
https://api.github.com/repos/ElvisChiang/welly
closed
中文介面下書籤功能出問題
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1.將作業系統介面改成繁體中文 What is the expected output? What do you see instead? 如圖所示 英文介面:https://www.dropbox.com/s/cz0ac9oj2io30jg/Screenshot%202014-07-28 %2009.47.45.png 中文介面:https://www.dropbox.com/s/vhcjws0hq2tovia/%E8%9E%A2%E5%B9%95%E6%8 8%AA%E5%9C%96%202014-07-28%2009.49.41.png What version of the product are you using? On what operating system? Welly版本從2.6一直用到最新版2.7.fix都有同樣問題 作業系統也是從10.6.8到目前最新的10.9.4 Please provide any additional information below. ``` Original issue reported on code.google.com by `swn...@gmail.com` on 28 Jul 2014 at 7:55
1.0
中文介面下書籤功能出問題 - ``` What steps will reproduce the problem? 1.將作業系統介面改成繁體中文 What is the expected output? What do you see instead? 如圖所示 英文介面:https://www.dropbox.com/s/cz0ac9oj2io30jg/Screenshot%202014-07-28 %2009.47.45.png 中文介面:https://www.dropbox.com/s/vhcjws0hq2tovia/%E8%9E%A2%E5%B9%95%E6%8 8%AA%E5%9C%96%202014-07-28%2009.49.41.png What version of the product are you using? On what operating system? Welly版本從2.6一直用到最新版2.7.fix都有同樣問題 作業系統也是從10.6.8到目前最新的10.9.4 Please provide any additional information below. ``` Original issue reported on code.google.com by `swn...@gmail.com` on 28 Jul 2014 at 7:55
defect
中文介面下書籤功能出問題 what steps will reproduce the problem 將作業系統介面改成繁體中文 what is the expected output what do you see instead 如圖所示 英文介面: png 中文介面: aa png what version of the product are you using on what operating system fix都有同樣問題 please provide any additional information below original issue reported on code google com by swn gmail com on jul at
1
18,048
3,022,104,717
IssuesEvent
2015-07-31 18:22:32
rolly32/support-tools
https://api.github.com/repos/rolly32/support-tools
reopened
Convert googlecode-issues-exporter.py to use python3
auto-migrated Priority-Low Type-Defect
``` python3 googlecode-issues-exporter/github_issue_converter.py <a lot of options> File "googlecode-issues-exporter/github_issue_converter.py", line 423 print "\nDone!\n" ^ SyntaxError: Missing parentheses in call to 'print' --- Reported on google-code-shutdown@ ``` Original issue reported on code.google.com by `chrsm...@google.com` on 13 Mar 2015 at 10:07
1.0
Convert googlecode-issues-exporter.py to use python3 - ``` python3 googlecode-issues-exporter/github_issue_converter.py <a lot of options> File "googlecode-issues-exporter/github_issue_converter.py", line 423 print "\nDone!\n" ^ SyntaxError: Missing parentheses in call to 'print' --- Reported on google-code-shutdown@ ``` Original issue reported on code.google.com by `chrsm...@google.com` on 13 Mar 2015 at 10:07
defect
convert googlecode issues exporter py to use googlecode issues exporter github issue converter py file googlecode issues exporter github issue converter py line print ndone n syntaxerror missing parentheses in call to print reported on google code shutdown original issue reported on code google com by chrsm google com on mar at
1
48,675
20,204,819,718
IssuesEvent
2022-02-11 19:02:05
codeoverflow-org/nodecg-io
https://api.github.com/repos/codeoverflow-org/nodecg-io
closed
Bundles using Reddit service fail to compile
bug minor service
### Description Newly generated bundles that use the reddit service currently fail to compile. The odd thing is that it does not affect the sample bundle which still builds fine. ### How to reproduce 1. Create a fresh nodecg install and install nodecg-io (current `main` or `0.2`) 2. Generate a new bundle that uses the reddit service using `nodecg-io generate` 3. Bundle fails to compile with the error below ### Error Log Compile failure: <!-- If the console contains some information that could help identifying the problem paste them in the next empty line --> ``` > tsc -b node_modules/reddit-ts/src/RedditAPI.ts:185:25 - error TS2345: Argument of type 'Child | undefined' is not assignable to parameter of type 'Child'. Type 'undefined' is not assignable to type 'Child'. 185 return map_search(data.children[0]) ~~~~~~~~~~~~~~~~ Found 1 error. ```
1.0
Bundles using Reddit service fail to compile - ### Description Newly generated bundles that use the reddit service currently fail to compile. The odd thing is that it does not affect the sample bundle which still builds fine. ### How to reproduce 1. Create a fresh nodecg install and install nodecg-io (current `main` or `0.2`) 2. Generate a new bundle that uses the reddit service using `nodecg-io generate` 3. Bundle fails to compile with the error below ### Error Log Compile failure: <!-- If the console contains some information that could help identifying the problem paste them in the next empty line --> ``` > tsc -b node_modules/reddit-ts/src/RedditAPI.ts:185:25 - error TS2345: Argument of type 'Child | undefined' is not assignable to parameter of type 'Child'. Type 'undefined' is not assignable to type 'Child'. 185 return map_search(data.children[0]) ~~~~~~~~~~~~~~~~ Found 1 error. ```
non_defect
bundles using reddit service fail to compile description newly generated bundles that use the reddit service currently fail to compile the odd thing is that it does not affect the sample bundle which still builds fine how to reproduce create a fresh nodecg install and install nodecg io current main or generate a new bundle that uses the reddit service using nodecg io generate bundle fails to compile with the error below error log compile failure if the console contains some information that could help identifying the problem paste them in the next empty line tsc b node modules reddit ts src redditapi ts error argument of type child undefined is not assignable to parameter of type child type undefined is not assignable to type child return map search data children found error
0