Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
181,389 | 6,659,273,214 | IssuesEvent | 2017-10-01 09:03:53 | OperationCode/operationcode_frontend | https://api.github.com/repos/OperationCode/operationcode_frontend | opened | Broken link in CONTRIBUTING.md | beginner friendly hacktoberfest Priority: Medium Status: Available | # Bug Report
## What is the current behavior?
`CONTRIBUTING.md` contains:
`- [Running the development Server](#running-the-development-server)`
on line 29.
## What is the expected behavior?
It should contain:
- [Running the Development Environment](#running-the-development-environment)
| 1.0 | Broken link in CONTRIBUTING.md - # Bug Report
## What is the current behavior?
`CONTRIBUTING.md` contains:
`- [Running the development Server](#running-the-development-server)`
on line 29.
## What is the expected behavior?
It should contain:
- [Running the Development Environment](#running-the-development-environment)
| non_defect | broken link in contributing md bug report what is the current behavior contributing md contains running the development server on line what is the expected behavior it should contain running the development environment | 0 |
167 | 2,640,878,704 | IssuesEvent | 2015-03-11 14:56:06 | ploneintranet/ploneintranet.theme | https://api.github.com/repos/ploneintranet/ploneintranet.theme | closed | Scroll-bars on texareas new-post/comment/repost areas in IE | browser compatibility in progress | New post area:

Comment areas in the stream:

Repost text area

| True | Scroll-bars on texareas new-post/comment/repost areas in IE - New post area:

Comment areas in the stream:

Repost text area

| non_defect | scroll bars on texareas new post comment repost areas in ie new post area comment areas in the stream repost text area | 0 |
115,286 | 4,662,351,494 | IssuesEvent | 2016-10-05 03:14:00 | communityshare/communityshare | https://api.github.com/repos/communityshare/communityshare | closed | Alert text shown briefly on page load | [Effort] Good first issues [Priority] High [Skills] Angular [Type] Bug [Type] UX | When loading the app home page, there is an alert that briefly shows before the rest of the HTML is rendered. It's empty and looks like a bug when viewing the page. This alert should not appear.
Likely resolved inside of `static/index.html`

This may have been introduced in in [this commit](https://github.com/communityshare/communityshare/commit/d085fb091cdfeaeffd682c73906f9bc9f8e4a68e).
| 1.0 | Alert text shown briefly on page load - When loading the app home page, there is an alert that briefly shows before the rest of the HTML is rendered. It's empty and looks like a bug when viewing the page. This alert should not appear.
Likely resolved inside of `static/index.html`

This may have been introduced in in [this commit](https://github.com/communityshare/communityshare/commit/d085fb091cdfeaeffd682c73906f9bc9f8e4a68e).
| non_defect | alert text shown briefly on page load when loading the app home page there is an alert that briefly shows before the rest of the html is rendered it s empty and looks like a bug when viewing the page this alert should not appear likely resolved inside of static index html this may have been introduced in in | 0 |
171,937 | 13,254,437,731 | IssuesEvent | 2020-08-20 09:15:41 | google/go-safeweb | https://api.github.com/repos/google/go-safeweb | opened | The test harness for testing net/http parsing is very slow, again | invalid testing | As of Go 1.15 #59 has been become a problem again. The solution we found in #63 doesn't work. But at least we know what is causing the slowdown.`net/http.Server.Shutdown`s polling interval for checking if all connections are finished is 0.5 s and we don't have time to shut everything down before the first sleep starts. Therefore we get punished by 0.5 s every time although closing our connections takes way less than 0.5 s. | 1.0 | The test harness for testing net/http parsing is very slow, again - As of Go 1.15 #59 has been become a problem again. The solution we found in #63 doesn't work. But at least we know what is causing the slowdown.`net/http.Server.Shutdown`s polling interval for checking if all connections are finished is 0.5 s and we don't have time to shut everything down before the first sleep starts. Therefore we get punished by 0.5 s every time although closing our connections takes way less than 0.5 s. | non_defect | the test harness for testing net http parsing is very slow again as of go has been become a problem again the solution we found in doesn t work but at least we know what is causing the slowdown net http server shutdown s polling interval for checking if all connections are finished is s and we don t have time to shut everything down before the first sleep starts therefore we get punished by s every time although closing our connections takes way less than s | 0 |
304,430 | 23,065,639,033 | IssuesEvent | 2022-07-25 13:43:06 | zeroback-song/zero-service | https://api.github.com/repos/zeroback-song/zero-service | closed | 제로백 rest api 정의 | documentation | 제로백 rest api 정의
* [x] 힙합 rest api 정의
* [x] 발라드 rest api 정의
* [x] 팝송 rest api 정의
* [x] 재즈 rest api 정의 | 1.0 | 제로백 rest api 정의 - 제로백 rest api 정의
* [x] 힙합 rest api 정의
* [x] 발라드 rest api 정의
* [x] 팝송 rest api 정의
* [x] 재즈 rest api 정의 | non_defect | 제로백 rest api 정의 제로백 rest api 정의 힙합 rest api 정의 발라드 rest api 정의 팝송 rest api 정의 재즈 rest api 정의 | 0 |
2,985 | 3,275,549,269 | IssuesEvent | 2015-10-26 15:58:01 | CellProfiler/CellProfiler | https://api.github.com/repos/CellProfiler/CellProfiler | opened | Saving Preferences should warn user to restart Cellprofiler | Feature request Layout and usability | Python/CellProfiler menu -> Preferences don't get applied until the next restart, no? Or maybe for some settings, like 'Max memory for Java'? The user should be warned about this (e.g. a modal dialog after Save that says "Please restart CellProfiler for these settings to be applied") | True | Saving Preferences should warn user to restart Cellprofiler - Python/CellProfiler menu -> Preferences don't get applied until the next restart, no? Or maybe for some settings, like 'Max memory for Java'? The user should be warned about this (e.g. a modal dialog after Save that says "Please restart CellProfiler for these settings to be applied") | non_defect | saving preferences should warn user to restart cellprofiler python cellprofiler menu preferences don t get applied until the next restart no or maybe for some settings like max memory for java the user should be warned about this e g a modal dialog after save that says please restart cellprofiler for these settings to be applied | 0 |
19,810 | 10,532,671,798 | IssuesEvent | 2019-10-01 11:19:28 | woocommerce/woocommerce-gutenberg-products-block | https://api.github.com/repos/woocommerce/woocommerce-gutenberg-products-block | opened | Performance of Block Settings on every page load without caching | type: performance | This was reported in https://github.com/woocommerce/woocommerce/issues/24590
Blocks requires certain settings/data to function, and this needs to be present anywhere a block may be used (so admin and frontend, regardless of user type). This is output to JSON inline on every page load.
This can be problematic if doing large queries such as getting terms, and counting post types.
Currently, the JSON is generated in Assets.php here:
https://github.com/woocommerce/woocommerce-gutenberg-products-block/blob/master/src/Assets.php#L86-L123
Most of this is harmless, with the exception of:
1. get_terms - this is querying all categories in the store and generating links for each.
2. wp_count_posts - less intensive but could still be cached.
We should look at either caching this data, using APIs instead as needed instead of grabbing data that may not be used, or making some of this conditonal. | True | Performance of Block Settings on every page load without caching - This was reported in https://github.com/woocommerce/woocommerce/issues/24590
Blocks requires certain settings/data to function, and this needs to be present anywhere a block may be used (so admin and frontend, regardless of user type). This is output to JSON inline on every page load.
This can be problematic if doing large queries such as getting terms, and counting post types.
Currently, the JSON is generated in Assets.php here:
https://github.com/woocommerce/woocommerce-gutenberg-products-block/blob/master/src/Assets.php#L86-L123
Most of this is harmless, with the exception of:
1. get_terms - this is querying all categories in the store and generating links for each.
2. wp_count_posts - less intensive but could still be cached.
We should look at either caching this data, using APIs instead as needed instead of grabbing data that may not be used, or making some of this conditonal. | non_defect | performance of block settings on every page load without caching this was reported in blocks requires certain settings data to function and this needs to be present anywhere a block may be used so admin and frontend regardless of user type this is output to json inline on every page load this can be problematic if doing large queries such as getting terms and counting post types currently the json is generated in assets php here most of this is harmless with the exception of get terms this is querying all categories in the store and generating links for each wp count posts less intensive but could still be cached we should look at either caching this data using apis instead as needed instead of grabbing data that may not be used or making some of this conditonal | 0 |
278,588 | 24,162,102,620 | IssuesEvent | 2022-09-22 12:30:15 | Tencent/bk-job | https://api.github.com/repos/Tencent/bk-job | closed | bugfix: 已上线的脚本,为啥还提示未保存的内容 | kind/bug stage/test accepted | **Version / Branch / tag**
≥3.5.1
**出了什么问题?(What Happened?)**

已上线的脚本,为啥还提示未保存的内容,操作方式是 新建版本,选了一个版本,保存,上线,再点右上角的叉时,提示的这个
**如何复现?(How to reproduce?)**
新建版本,选了一个版本,保存,上线,再点右上角的叉时,提示的这个
**预期结果(What you expect?)**
上线后,因为已经是只读状态而非编辑状态了,这时离开不应该还要提示未保存信息的弹窗 | 1.0 | bugfix: 已上线的脚本,为啥还提示未保存的内容 - **Version / Branch / tag**
≥3.5.1
**出了什么问题?(What Happened?)**

已上线的脚本,为啥还提示未保存的内容,操作方式是 新建版本,选了一个版本,保存,上线,再点右上角的叉时,提示的这个
**如何复现?(How to reproduce?)**
新建版本,选了一个版本,保存,上线,再点右上角的叉时,提示的这个
**预期结果(What you expect?)**
上线后,因为已经是只读状态而非编辑状态了,这时离开不应该还要提示未保存信息的弹窗 | non_defect | bugfix 已上线的脚本,为啥还提示未保存的内容 version branch tag ≥ 出了什么问题? what happened 已上线的脚本,为啥还提示未保存的内容,操作方式是 新建版本,选了一个版本,保存,上线,再点右上角的叉时,提示的这个 如何复现? how to reproduce 新建版本,选了一个版本,保存,上线,再点右上角的叉时,提示的这个 预期结果 what you expect 上线后,因为已经是只读状态而非编辑状态了,这时离开不应该还要提示未保存信息的弹窗 | 0 |
655 | 2,823,166,139 | IssuesEvent | 2015-05-21 06:53:04 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Consider open sourcing ApiCompat. | enhancement Infrastructure | Internally we have some tooling to ensure that implementation assemblies are compatible with contracts. It would be a good idea to open source that and use it as part of the build or CI system or something. I'm not sure if this will happen naturally when we actually start generating contracts in the open or if we'll need to take extra steps. | 1.0 | Consider open sourcing ApiCompat. - Internally we have some tooling to ensure that implementation assemblies are compatible with contracts. It would be a good idea to open source that and use it as part of the build or CI system or something. I'm not sure if this will happen naturally when we actually start generating contracts in the open or if we'll need to take extra steps. | non_defect | consider open sourcing apicompat internally we have some tooling to ensure that implementation assemblies are compatible with contracts it would be a good idea to open source that and use it as part of the build or ci system or something i m not sure if this will happen naturally when we actually start generating contracts in the open or if we ll need to take extra steps | 0 |
93,501 | 3,901,182,294 | IssuesEvent | 2016-04-18 09:46:27 | natsys/tempesta | https://api.github.com/repos/natsys/tempesta | opened | Kernel warning when calling __alloc_skb from the soft-IRQ | enhancement low priority | Kernel warning was caught by me on the laptop:
~~~
[ 3.129833] ------------[ cut here ]------------
[ 3.129839] WARNING: CPU: 1 PID: 0 at kernel/softirq.c:150 __local_bh_enable_ip+0x72/0xa0()
[ 3.129841] Modules linked in: media btusb btbcm btintel bluetooth cdc_mbim intel_rapl iosf_mbi x86_pkg_temp_thermal intel_powerclamp coretemp cdc_ncm usbnet kvm_intel mii cdc_wdm cdc_acm arc4 kvm iwldvm iTCO_wdt iTCO_vendor_support crct10dif_pclmul crc32_pclmul crc32c_intel mac80211 aesni_intel aes_x86_64 iwlwifi lrw gf128mul glue_helper snd_hda_codec_hdmi ablk_helper cryptd psmouse serio_raw pcspkr cfg80211 snd_hda_codec_realtek snd_hda_codec_generic snd_hda_intel thermal snd_hda_controller snd_hda_codec wmi snd_hda_core snd_hwdep thinkpad_acpi snd_pcm snd_timer nvram ac rfkill tpm_tis mei_me tpm battery snd mei evdev i2c_i801 lpc_ich soundcore shpchp mac_hid processor sch_fq_codel vboxnetadp(O) vboxdrv(O) ip_tables x_tables ext4 crc16 mbcache jbd2 sd_mod atkbd libps2 sdhci_pci ahci libahci libata
[ 3.129883] scsi_mod xhci_pci ehci_pci sdhci ehci_hcd xhci_hcd led_class mmc_core usbcore usb_common i8042 serio i915 button intel_gtt i2c_algo_bit video drm_kms_helper drm
[ 3.129894] CPU: 1 PID: 0 Comm: swapper/1 Tainted: G O 4.1.21-1-tfw #1
[ 3.129897] Hardware name: LENOVO 34442KU/34442KU, BIOS G6ETB1WW (2.71 ) 04/23/2015
[ 3.129899] 0000000000000086 3f00b632d218473c ffff88011e243c98 ffffffff81588da2
[ 3.129902] 0000000000000000 ffffffff817129f0 ffff88011e243cd8 ffffffff8107be4a
[ 3.129904] ffff88011e2107e0 0000000000000200 0000000000000020 000077ff80000000
[ 3.129907] Call Trace:
[ 3.129909] <IRQ> [<ffffffff81588da2>] dump_stack+0x63/0x81
[ 3.129915] [<ffffffff8107be4a>] warn_slowpath_common+0x8a/0xc0
[ 3.129918] [<ffffffff8107bf7a>] warn_slowpath_null+0x1a/0x20
[ 3.129921] [<ffffffff8107fd12>] __local_bh_enable_ip+0x72/0xa0
[ 3.129925] [<ffffffff8147456a>] __pg_skb_alloc+0x2fa/0x3e0
[ 3.129928] [<ffffffff8147469e>] __alloc_skb+0x4e/0x250
[ 3.129932] [<ffffffffa07a9506>] btusb_recv_intr+0x126/0x170 [btusb]
[ 3.129935] [<ffffffffa07a9600>] btusb_intr_complete+0xb0/0x140 [btusb]
[ 3.129943] [<ffffffffa01fd2b5>] __usb_hcd_giveback_urb+0x85/0x140 [usbcore]
[ 3.129950] [<ffffffffa01fd416>] usb_giveback_urb_bh+0xa6/0x100 [usbcore]
[ 3.129953] [<ffffffff81080677>] tasklet_hi_action+0xb7/0xd0
[ 3.129955] [<ffffffff8107fe51>] __do_softirq+0x111/0x2d0
[ 3.129958] [<ffffffff8108018d>] irq_exit+0xad/0xb0
[ 3.129961] [<ffffffff81590d48>] do_IRQ+0x58/0xf0
[ 3.129965] [<ffffffff8158ed2e>] common_interrupt+0x6e/0x6e
[ 3.129966] <EOI> [<ffffffff8101eb0d>] ? native_sched_clock+0x2d/0x90
[ 3.129972] [<ffffffff81442b63>] ? cpuidle_enter_state+0x93/0x240
[ 3.129975] [<ffffffff81442b36>] ? cpuidle_enter_state+0x66/0x240
[ 3.129977] [<ffffffff81442d47>] cpuidle_enter+0x17/0x20
[ 3.129981] [<ffffffff810bdf7e>] cpu_startup_entry+0x31e/0x430
[ 3.129984] [<ffffffff8104ea6e>] start_secondary+0x18e/0x1d0
[ 3.129986] ---[ end trace df240af6b2a08d04 ]---
~~~
[__local_bh_enable_ip](http://lxr.free-electrons.com/source/kernel/softirq.c?v=4.1#L148)
~~~
148 void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
149 {
150 WARN_ON_ONCE(in_irq() || irqs_disabled());
~~~
The reason is [btusb_recv_intr](http://lxr.free-electrons.com/source/drivers/bluetooth/btusb.c?v=4.1#L396) which runs under tasklet and calls `alloc_skb` which calls `__pg_skb_alloc`.
| 1.0 | Kernel warning when calling __alloc_skb from the soft-IRQ - Kernel warning was caught by me on the laptop:
~~~
[ 3.129833] ------------[ cut here ]------------
[ 3.129839] WARNING: CPU: 1 PID: 0 at kernel/softirq.c:150 __local_bh_enable_ip+0x72/0xa0()
[ 3.129841] Modules linked in: media btusb btbcm btintel bluetooth cdc_mbim intel_rapl iosf_mbi x86_pkg_temp_thermal intel_powerclamp coretemp cdc_ncm usbnet kvm_intel mii cdc_wdm cdc_acm arc4 kvm iwldvm iTCO_wdt iTCO_vendor_support crct10dif_pclmul crc32_pclmul crc32c_intel mac80211 aesni_intel aes_x86_64 iwlwifi lrw gf128mul glue_helper snd_hda_codec_hdmi ablk_helper cryptd psmouse serio_raw pcspkr cfg80211 snd_hda_codec_realtek snd_hda_codec_generic snd_hda_intel thermal snd_hda_controller snd_hda_codec wmi snd_hda_core snd_hwdep thinkpad_acpi snd_pcm snd_timer nvram ac rfkill tpm_tis mei_me tpm battery snd mei evdev i2c_i801 lpc_ich soundcore shpchp mac_hid processor sch_fq_codel vboxnetadp(O) vboxdrv(O) ip_tables x_tables ext4 crc16 mbcache jbd2 sd_mod atkbd libps2 sdhci_pci ahci libahci libata
[ 3.129883] scsi_mod xhci_pci ehci_pci sdhci ehci_hcd xhci_hcd led_class mmc_core usbcore usb_common i8042 serio i915 button intel_gtt i2c_algo_bit video drm_kms_helper drm
[ 3.129894] CPU: 1 PID: 0 Comm: swapper/1 Tainted: G O 4.1.21-1-tfw #1
[ 3.129897] Hardware name: LENOVO 34442KU/34442KU, BIOS G6ETB1WW (2.71 ) 04/23/2015
[ 3.129899] 0000000000000086 3f00b632d218473c ffff88011e243c98 ffffffff81588da2
[ 3.129902] 0000000000000000 ffffffff817129f0 ffff88011e243cd8 ffffffff8107be4a
[ 3.129904] ffff88011e2107e0 0000000000000200 0000000000000020 000077ff80000000
[ 3.129907] Call Trace:
[ 3.129909] <IRQ> [<ffffffff81588da2>] dump_stack+0x63/0x81
[ 3.129915] [<ffffffff8107be4a>] warn_slowpath_common+0x8a/0xc0
[ 3.129918] [<ffffffff8107bf7a>] warn_slowpath_null+0x1a/0x20
[ 3.129921] [<ffffffff8107fd12>] __local_bh_enable_ip+0x72/0xa0
[ 3.129925] [<ffffffff8147456a>] __pg_skb_alloc+0x2fa/0x3e0
[ 3.129928] [<ffffffff8147469e>] __alloc_skb+0x4e/0x250
[ 3.129932] [<ffffffffa07a9506>] btusb_recv_intr+0x126/0x170 [btusb]
[ 3.129935] [<ffffffffa07a9600>] btusb_intr_complete+0xb0/0x140 [btusb]
[ 3.129943] [<ffffffffa01fd2b5>] __usb_hcd_giveback_urb+0x85/0x140 [usbcore]
[ 3.129950] [<ffffffffa01fd416>] usb_giveback_urb_bh+0xa6/0x100 [usbcore]
[ 3.129953] [<ffffffff81080677>] tasklet_hi_action+0xb7/0xd0
[ 3.129955] [<ffffffff8107fe51>] __do_softirq+0x111/0x2d0
[ 3.129958] [<ffffffff8108018d>] irq_exit+0xad/0xb0
[ 3.129961] [<ffffffff81590d48>] do_IRQ+0x58/0xf0
[ 3.129965] [<ffffffff8158ed2e>] common_interrupt+0x6e/0x6e
[ 3.129966] <EOI> [<ffffffff8101eb0d>] ? native_sched_clock+0x2d/0x90
[ 3.129972] [<ffffffff81442b63>] ? cpuidle_enter_state+0x93/0x240
[ 3.129975] [<ffffffff81442b36>] ? cpuidle_enter_state+0x66/0x240
[ 3.129977] [<ffffffff81442d47>] cpuidle_enter+0x17/0x20
[ 3.129981] [<ffffffff810bdf7e>] cpu_startup_entry+0x31e/0x430
[ 3.129984] [<ffffffff8104ea6e>] start_secondary+0x18e/0x1d0
[ 3.129986] ---[ end trace df240af6b2a08d04 ]---
~~~
[__local_bh_enable_ip](http://lxr.free-electrons.com/source/kernel/softirq.c?v=4.1#L148)
~~~
148 void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
149 {
150 WARN_ON_ONCE(in_irq() || irqs_disabled());
~~~
The reason is [btusb_recv_intr](http://lxr.free-electrons.com/source/drivers/bluetooth/btusb.c?v=4.1#L396) which runs under tasklet and calls `alloc_skb` which calls `__pg_skb_alloc`.
| non_defect | kernel warning when calling alloc skb from the soft irq kernel warning was caught by me on the laptop warning cpu pid at kernel softirq c local bh enable ip modules linked in media btusb btbcm btintel bluetooth cdc mbim intel rapl iosf mbi pkg temp thermal intel powerclamp coretemp cdc ncm usbnet kvm intel mii cdc wdm cdc acm kvm iwldvm itco wdt itco vendor support pclmul pclmul intel aesni intel aes iwlwifi lrw glue helper snd hda codec hdmi ablk helper cryptd psmouse serio raw pcspkr snd hda codec realtek snd hda codec generic snd hda intel thermal snd hda controller snd hda codec wmi snd hda core snd hwdep thinkpad acpi snd pcm snd timer nvram ac rfkill tpm tis mei me tpm battery snd mei evdev lpc ich soundcore shpchp mac hid processor sch fq codel vboxnetadp o vboxdrv o ip tables x tables mbcache sd mod atkbd sdhci pci ahci libahci libata scsi mod xhci pci ehci pci sdhci ehci hcd xhci hcd led class mmc core usbcore usb common serio button intel gtt algo bit video drm kms helper drm cpu pid comm swapper tainted g o tfw hardware name lenovo bios call trace dump stack warn slowpath common warn slowpath null local bh enable ip pg skb alloc alloc skb btusb recv intr btusb intr complete usb hcd giveback urb usb giveback urb bh tasklet hi action do softirq irq exit do irq common interrupt native sched clock cpuidle enter state cpuidle enter state cpuidle enter cpu startup entry start secondary void local bh enable ip unsigned long ip unsigned int cnt warn on once in irq irqs disabled the reason is which runs under tasklet and calls alloc skb which calls pg skb alloc | 0 |
13,964 | 2,789,803,377 | IssuesEvent | 2015-05-08 21:35:35 | google/google-visualization-api-issues | https://api.github.com/repos/google/google-visualization-api-issues | opened | LineChart select event to open URL does not work on Internet Explorer | Priority-Medium Type-Defect | Original [issue 157](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=157) created by orwant on 2009-12-29T22:33:16.000Z:
<b>What steps will reproduce the problem? Please provide a link to a</b>
<b>demonstration page if at all possible, or attach code.</b>
1. Code is attached. Hover over data points to see the pop up tooltip, and
when clicked, it will go to a URL (in this case, www.google.com).
It seems to work fine in Firefox, Safari, and Chrome but in Internet
Explorer, the clicking of the point to go to the URL doesn't work.
You can see the highlighting of the point in the other browsers, but
not in IE. This same URL functionality works fine in the PieChart component
on Internet Explorer.
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
LineChart
<b>Are you using the test environment (version 1.1)?</b>
No.
<b>What operating system and browser are you using?</b>
Windows XP & Internet Explorer 7 & 8.
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
| 1.0 | LineChart select event to open URL does not work on Internet Explorer - Original [issue 157](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=157) created by orwant on 2009-12-29T22:33:16.000Z:
<b>What steps will reproduce the problem? Please provide a link to a</b>
<b>demonstration page if at all possible, or attach code.</b>
1. Code is attached. Hover over data points to see the pop up tooltip, and
when clicked, it will go to a URL (in this case, www.google.com).
It seems to work fine in Firefox, Safari, and Chrome but in Internet
Explorer, the clicking of the point to go to the URL doesn't work.
You can see the highlighting of the point in the other browsers, but
not in IE. This same URL functionality works fine in the PieChart component
on Internet Explorer.
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
LineChart
<b>Are you using the test environment (version 1.1)?</b>
No.
<b>What operating system and browser are you using?</b>
Windows XP & Internet Explorer 7 & 8.
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
| defect | linechart select event to open url does not work on internet explorer original created by orwant on what steps will reproduce the problem please provide a link to a demonstration page if at all possible or attach code code is attached hover over data points to see the pop up tooltip and when clicked it will go to a url in this case it seems to work fine in firefox safari and chrome but in internet explorer the clicking of the point to go to the url doesn t work you can see the highlighting of the point in the other browsers but not in ie this same url functionality works fine in the piechart component on internet explorer what component is this issue related to piechart linechart datatable query etc linechart are you using the test environment version no what operating system and browser are you using windows xp amp internet explorer amp for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved | 1 |
48,498 | 13,100,884,246 | IssuesEvent | 2020-08-04 01:59:51 | FoldingAtHome/fah-issues | https://api.github.com/repos/FoldingAtHome/fah-issues | closed | Incomplete fix for Core_Outdated with two slots | 1.Type - Defect 3.Component - FAHClient 4.OS - All | Core_outdated means FAHClient needs to download a new core. I have two slots using that core. One was running and I unpaused the other. Naturally, the file was is use so the update couldn't proceed.
Pausing the active slot gives me two slots with status Update_Core. That.s correct ... except that now the download needs to proceed but it doesn't
Windows 7.6.9
Core had been deleted looking for a fresh copy. After a long wait, I paused both slots and then unpaused them and something worked. I'm not sure if that unjammed it or if it just took ~10 minutes to decide to proceed.
| 1.0 | Incomplete fix for Core_Outdated with two slots - Core_outdated means FAHClient needs to download a new core. I have two slots using that core. One was running and I unpaused the other. Naturally, the file was is use so the update couldn't proceed.
Pausing the active slot gives me two slots with status Update_Core. That.s correct ... except that now the download needs to proceed but it doesn't
Windows 7.6.9
Core had been deleted looking for a fresh copy. After a long wait, I paused both slots and then unpaused them and something worked. I'm not sure if that unjammed it or if it just took ~10 minutes to decide to proceed.
| defect | incomplete fix for core outdated with two slots core outdated means fahclient needs to download a new core i have two slots using that core one was running and i unpaused the other naturally the file was is use so the update couldn t proceed pausing the active slot gives me two slots with status update core that s correct except that now the download needs to proceed but it doesn t windows core had been deleted looking for a fresh copy after a long wait i paused both slots and then unpaused them and something worked i m not sure if that unjammed it or if it just took minutes to decide to proceed | 1 |
10,636 | 2,622,178,205 | IssuesEvent | 2015-03-04 00:17:37 | byzhang/leveldb | https://api.github.com/repos/byzhang/leveldb | closed | Add version info in include/c.h | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. There isn't any version info in b binding file include/c.h
What is the expected output? What do you see instead?
I can get version info in c interface
What version of the product are you using? On what operating system?
ANY
I'm working on a php binding using c binding, but I can't get version info,
If it was supplied as a macro will be much appreciated.
```
Original issue reported on code.google.com by `reeze....@gmail.com` on 13 Jun 2012 at 3:42 | 1.0 | Add version info in include/c.h - ```
What steps will reproduce the problem?
1. There isn't any version info in b binding file include/c.h
What is the expected output? What do you see instead?
I can get version info in c interface
What version of the product are you using? On what operating system?
ANY
I'm working on a php binding using c binding, but I can't get version info,
If it was supplied as a macro will be much appreciated.
```
Original issue reported on code.google.com by `reeze....@gmail.com` on 13 Jun 2012 at 3:42 | defect | add version info in include c h what steps will reproduce the problem there isn t any version info in b binding file include c h what is the expected output what do you see instead i can get version info in c interface what version of the product are you using on what operating system any i m working on a php binding using c binding but i can t get version info if it was supplied as a macro will be much appreciated original issue reported on code google com by reeze gmail com on jun at | 1 |
58,466 | 16,546,038,260 | IssuesEvent | 2021-05-28 00:13:12 | idaholab/moose | https://api.github.com/repos/idaholab/moose | closed | Fixed PETSc alt update script | P: normal T: defect | ## Bug Description
<!--A clear and concise description of the problem (Note: A missing feature is not a bug).-->
```
❯ cd ~/projects/moose/
❯ rm -rf petsc
❯ cd scripts
❯ ./update_and_rebuild_petsc_alt.sh
*** WARNING ***
scripts/update_and_rebuild_petsc_alt.sh is intended for internal
use only. Please use scripts/update_and_rebuild_petsc.sh instead.
/Users/milljm/projects/moose/scripts
./update_and_rebuild_petsc_alt.sh: line 30: cd: /Users/milljm/projects/moose/scripts/../petsc: No such file or directory
error: pathspec 'v3.11.4' did not match any file(s) known to git
git checkout command failed, are your proxy settings correct?
```
## Fix
<!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)-->
We need to add "cd $SCRIPT_DIR/.."
So that we are always in the right directory
## Impact
<!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
| 1.0 | Fixed PETSc alt update script - ## Bug Description
<!--A clear and concise description of the problem (Note: A missing feature is not a bug).-->
```
❯ cd ~/projects/moose/
❯ rm -rf petsc
❯ cd scripts
❯ ./update_and_rebuild_petsc_alt.sh
*** WARNING ***
scripts/update_and_rebuild_petsc_alt.sh is intended for internal
use only. Please use scripts/update_and_rebuild_petsc.sh instead.
/Users/milljm/projects/moose/scripts
./update_and_rebuild_petsc_alt.sh: line 30: cd: /Users/milljm/projects/moose/scripts/../petsc: No such file or directory
error: pathspec 'v3.11.4' did not match any file(s) known to git
git checkout command failed, are your proxy settings correct?
```
## Fix
<!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)-->
We need to add "cd $SCRIPT_DIR/.."
So that we are always in the right directory
## Impact
<!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
| defect | fixed petsc alt update script bug description ❯ cd projects moose ❯ rm rf petsc ❯ cd scripts ❯ update and rebuild petsc alt sh warning scripts update and rebuild petsc alt sh is intended for internal use only please use scripts update and rebuild petsc sh instead users milljm projects moose scripts update and rebuild petsc alt sh line cd users milljm projects moose scripts petsc no such file or directory error pathspec did not match any file s known to git git checkout command failed are your proxy settings correct fix we need to add cd script dir so that we are always in the right directory impact | 1 |
451,118 | 32,008,908,622 | IssuesEvent | 2023-09-21 16:34:08 | NOAA-EMC/NCEPLIBS-ip | https://api.github.com/repos/NOAA-EMC/NCEPLIBS-ip | closed | Missing documentation for iplib and ip2lib header files. | documentation | In ip2lib_4.h we have:
```
/** @file
* @brief C interface to gdswzd() and gdswzd_grib1() functions for '4'
* library build.
* @author NOAA Programmer
*/
#ifndef IPLIB
#define IPLIB
/**
GDSWZD in C.
@param igdtnum
@param igdtmpl
@param igdtlen
@param iopt
@param npts
@param fill
@param *xpts
@param *ypts
@param *rlon
@param *rlat
@param *nret
@param *crot
@param *srot
@param *xlon
@param *xlat
@param *ylon
@param *ylat
@param *area
*/
void gdswzd(int igdtnum, int *igdtmpl, int igdtlen, int iopt,
int npts, float fill, float *xpts, float *ypts,
float *rlon, float *rlat, int *nret,
float *crot, float *srot, float *xlon, float *xlat,
float *ylon, float *ylat, float *area);
/**
GDSWZD_grib1 in C.
@param kgds
@param iopt
@param npts
@param *fill
@param *xpts
@param *ypts
@param *rlon
@param *rlat
@param nret
@param *crot
@param *srot
@param *xlon
@param *xlat
@param *ylon
@param *ylat
@param *area
*/
void gdswzd_grib1(int kgds, int iopt, int npts, float *fill, float *xpts,
float *ypts, float *rlon, float *rlat, int nret, float *crot,
float *srot, float *xlon, float *xlat, float *ylon,
float *ylat, float *area);
#endif
```
We need much better documentation than this.
Let's start by figuring out who the author is.
@GeorgeGayno-NOAA are you the author of this code? | 1.0 | Missing documentation for iplib and ip2lib header files. - In ip2lib_4.h we have:
```
/** @file
* @brief C interface to gdswzd() and gdswzd_grib1() functions for '4'
* library build.
* @author NOAA Programmer
*/
#ifndef IPLIB
#define IPLIB
/**
GDSWZD in C.
@param igdtnum
@param igdtmpl
@param igdtlen
@param iopt
@param npts
@param fill
@param *xpts
@param *ypts
@param *rlon
@param *rlat
@param *nret
@param *crot
@param *srot
@param *xlon
@param *xlat
@param *ylon
@param *ylat
@param *area
*/
void gdswzd(int igdtnum, int *igdtmpl, int igdtlen, int iopt,
int npts, float fill, float *xpts, float *ypts,
float *rlon, float *rlat, int *nret,
float *crot, float *srot, float *xlon, float *xlat,
float *ylon, float *ylat, float *area);
/**
GDSWZD_grib1 in C.
@param kgds
@param iopt
@param npts
@param *fill
@param *xpts
@param *ypts
@param *rlon
@param *rlat
@param nret
@param *crot
@param *srot
@param *xlon
@param *xlat
@param *ylon
@param *ylat
@param *area
*/
void gdswzd_grib1(int kgds, int iopt, int npts, float *fill, float *xpts,
float *ypts, float *rlon, float *rlat, int nret, float *crot,
float *srot, float *xlon, float *xlat, float *ylon,
float *ylat, float *area);
#endif
```
We need much better documentation than this.
Let's start by figuring out who the author is.
@GeorgeGayno-NOAA are you the author of this code? | non_defect | missing documentation for iplib and header files in h we have file brief c interface to gdswzd and gdswzd functions for library build author noaa programmer ifndef iplib define iplib gdswzd in c param igdtnum param igdtmpl param igdtlen param iopt param npts param fill param xpts param ypts param rlon param rlat param nret param crot param srot param xlon param xlat param ylon param ylat param area void gdswzd int igdtnum int igdtmpl int igdtlen int iopt int npts float fill float xpts float ypts float rlon float rlat int nret float crot float srot float xlon float xlat float ylon float ylat float area gdswzd in c param kgds param iopt param npts param fill param xpts param ypts param rlon param rlat param nret param crot param srot param xlon param xlat param ylon param ylat param area void gdswzd int kgds int iopt int npts float fill float xpts float ypts float rlon float rlat int nret float crot float srot float xlon float xlat float ylon float ylat float area endif we need much better documentation than this let s start by figuring out who the author is georgegayno noaa are you the author of this code | 0 |
534,879 | 15,651,035,094 | IssuesEvent | 2021-03-23 09:43:22 | leihs/leihs | https://api.github.com/repos/leihs/leihs | reopened | show a message when the user can login but has no access right | low priority | - [ ] add page in `my`
- [ ] redirect from legacy | 1.0 | show a message when the user can login but has no access right - - [ ] add page in `my`
- [ ] redirect from legacy | non_defect | show a message when the user can login but has no access right add page in my redirect from legacy | 0 |
354,749 | 25,174,839,999 | IssuesEvent | 2022-11-11 08:17:48 | TZL0/pe | https://api.github.com/repos/TZL0/pe | opened | Incorrect destination of PDF link | severity.Low type.DocumentationBug | Pressing the 'Command Line Interface' takes the user to Graphical User Interface


<!--session: 1668154547514-b0cf5391-b882-4c7a-8aeb-5d77963c2458-->
<!--Version: Web v3.4.4--> | 1.0 | Incorrect destination of PDF link - Pressing the 'Command Line Interface' takes the user to Graphical User Interface


<!--session: 1668154547514-b0cf5391-b882-4c7a-8aeb-5d77963c2458-->
<!--Version: Web v3.4.4--> | non_defect | incorrect destination of pdf link pressing the command line interface takes the user to graphical user interface | 0 |
68,305 | 21,628,211,685 | IssuesEvent | 2022-05-05 06:43:16 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | SQL Translation Error | T: Defect P: Medium E: All Editions C: Translator | ### SQL translator incorrectly translates this query:
```sql
CREATE PROCEDURE unluckyEmployees()
BEGIN
SET @rn =0 | 1.0 | SQL Translation Error - ### SQL translator incorrectly translates this query:
```sql
CREATE PROCEDURE unluckyEmployees()
BEGIN
SET @rn =0 | defect | sql translation error sql translator incorrectly translates this query sql create procedure unluckyemployees begin set rn | 1 |
48,347 | 13,068,455,172 | IssuesEvent | 2020-07-31 03:38:01 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | [clsim] GPU detection broken for RTX-series cards (Trac #2212) | Migrated from Trac combo simulation defect | Nvidia decided that ray-tracing was a thing, and renamed their GTX series "RTX", and e.g. the 2080 Ti advertises itself as "GeForce RTX 2080 Ti". clsim.traysegments.common.configureOpenCLDevices, however, special-cases cards named "GTX" to set the number of work items to something sane and enable native math. Is there any reason not to do away with the special cases, and simply scale the number of work items to be proportional to the global memory size, as well as enabling native math by default?
Migrated from https://code.icecube.wisc.edu/ticket/2212
```json
{
"status": "closed",
"changetime": "2019-03-06T20:58:01",
"description": "Nvidia decided that ray-tracing was a thing, and renamed their GTX series \"RTX\", and e.g. the 2080 Ti advertises itself as \"GeForce RTX 2080 Ti\". clsim.traysegments.common.configureOpenCLDevices, however, special-cases cards named \"GTX\" to set the number of work items to something sane and enable native math. Is there any reason not to do away with the special cases, and simply scale the number of work items to be proportional to the global memory size, as well as enabling native math by default?",
"reporter": "jvansanten",
"cc": "",
"resolution": "fixed",
"_ts": "1551905881354849",
"component": "combo simulation",
"summary": "[clsim] GPU detection broken for RTX-series cards",
"priority": "normal",
"keywords": "",
"time": "2018-11-29T14:14:01",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
| 1.0 | [clsim] GPU detection broken for RTX-series cards (Trac #2212) - Nvidia decided that ray-tracing was a thing, and renamed their GTX series "RTX", and e.g. the 2080 Ti advertises itself as "GeForce RTX 2080 Ti". clsim.traysegments.common.configureOpenCLDevices, however, special-cases cards named "GTX" to set the number of work items to something sane and enable native math. Is there any reason not to do away with the special cases, and simply scale the number of work items to be proportional to the global memory size, as well as enabling native math by default?
Migrated from https://code.icecube.wisc.edu/ticket/2212
```json
{
"status": "closed",
"changetime": "2019-03-06T20:58:01",
"description": "Nvidia decided that ray-tracing was a thing, and renamed their GTX series \"RTX\", and e.g. the 2080 Ti advertises itself as \"GeForce RTX 2080 Ti\". clsim.traysegments.common.configureOpenCLDevices, however, special-cases cards named \"GTX\" to set the number of work items to something sane and enable native math. Is there any reason not to do away with the special cases, and simply scale the number of work items to be proportional to the global memory size, as well as enabling native math by default?",
"reporter": "jvansanten",
"cc": "",
"resolution": "fixed",
"_ts": "1551905881354849",
"component": "combo simulation",
"summary": "[clsim] GPU detection broken for RTX-series cards",
"priority": "normal",
"keywords": "",
"time": "2018-11-29T14:14:01",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
| defect | gpu detection broken for rtx series cards trac nvidia decided that ray tracing was a thing and renamed their gtx series rtx and e g the ti advertises itself as geforce rtx ti clsim traysegments common configureopencldevices however special cases cards named gtx to set the number of work items to something sane and enable native math is there any reason not to do away with the special cases and simply scale the number of work items to be proportional to the global memory size as well as enabling native math by default migrated from json status closed changetime description nvidia decided that ray tracing was a thing and renamed their gtx series rtx and e g the ti advertises itself as geforce rtx ti clsim traysegments common configureopencldevices however special cases cards named gtx to set the number of work items to something sane and enable native math is there any reason not to do away with the special cases and simply scale the number of work items to be proportional to the global memory size as well as enabling native math by default reporter jvansanten cc resolution fixed ts component combo simulation summary gpu detection broken for rtx series cards priority normal keywords time milestone owner jvansanten type defect | 1 |
232,095 | 25,564,962,440 | IssuesEvent | 2022-11-30 13:40:30 | billmcchesney1/concord | https://api.github.com/repos/billmcchesney1/concord | opened | CVE-2022-38900 (High) detected in decode-uri-component-0.2.0.tgz | security vulnerability | ## CVE-2022-38900 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>decode-uri-component-0.2.0.tgz</b></p></summary>
<p>A better decodeURIComponent</p>
<p>Library home page: <a href="https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz">https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz</a></p>
<p>Path to dependency file: /console2/package.json</p>
<p>Path to vulnerable library: /console2/node_modules/decode-uri-component/package.json</p>
<p>
Dependency Hierarchy:
- query-string-6.13.5.tgz (Root Library)
- :x: **decode-uri-component-0.2.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
decode-uri-component 0.2.0 is vulnerable to Improper Input Validation resulting in DoS.
<p>Publish Date: 2022-11-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38900>CVE-2022-38900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
| True | CVE-2022-38900 (High) detected in decode-uri-component-0.2.0.tgz - ## CVE-2022-38900 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>decode-uri-component-0.2.0.tgz</b></p></summary>
<p>A better decodeURIComponent</p>
<p>Library home page: <a href="https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz">https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz</a></p>
<p>Path to dependency file: /console2/package.json</p>
<p>Path to vulnerable library: /console2/node_modules/decode-uri-component/package.json</p>
<p>
Dependency Hierarchy:
- query-string-6.13.5.tgz (Root Library)
- :x: **decode-uri-component-0.2.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
decode-uri-component 0.2.0 is vulnerable to Improper Input Validation resulting in DoS.
<p>Publish Date: 2022-11-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38900>CVE-2022-38900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
| non_defect | cve high detected in decode uri component tgz cve high severity vulnerability vulnerable library decode uri component tgz a better decodeuricomponent library home page a href path to dependency file package json path to vulnerable library node modules decode uri component package json dependency hierarchy query string tgz root library x decode uri component tgz vulnerable library found in base branch master vulnerability details decode uri component is vulnerable to improper input validation resulting in dos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href | 0 |
59,945 | 17,023,296,356 | IssuesEvent | 2021-07-03 01:17:22 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Handling of notConnectedSameTag don't work with juction with dirrerent ways | Component: osmarender Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 1.56pm, Wednesday, 17th September 2008]**
Take folowing test case: a rount about junction which is part of a primary road (so that way forming the round about itself is also tagged with same name a ways of primary road to an from the junction).
In this case the round about is not labelled as expected due to notConnectedSameTag on corresponding rule (see correct.svg).
But if I connect an other road with a different name to the round about, the labels reappears (see notcorrect.svg).
The attached path corrects this issue. | 1.0 | Handling of notConnectedSameTag don't work with juction with dirrerent ways - **[Submitted to the original trac issue database at 1.56pm, Wednesday, 17th September 2008]**
Take folowing test case: a rount about junction which is part of a primary road (so that way forming the round about itself is also tagged with same name a ways of primary road to an from the junction).
In this case the round about is not labelled as expected due to notConnectedSameTag on corresponding rule (see correct.svg).
But if I connect an other road with a different name to the round about, the labels reappears (see notcorrect.svg).
The attached path corrects this issue. | defect | handling of notconnectedsametag don t work with juction with dirrerent ways take folowing test case a rount about junction which is part of a primary road so that way forming the round about itself is also tagged with same name a ways of primary road to an from the junction in this case the round about is not labelled as expected due to notconnectedsametag on corresponding rule see correct svg but if i connect an other road with a different name to the round about the labels reappears see notcorrect svg the attached path corrects this issue | 1 |
592,952 | 17,934,668,281 | IssuesEvent | 2021-09-10 13:56:50 | mozilla/addons-server | https://api.github.com/repos/mozilla/addons-server | closed | Uploading an add-on with a GUID in non string format triggers a server error | component: devhub priority: p3 | ### Describe the problem and steps to reproduce it:
1. Submit an add-on with a GUID in non string format, for example:
```
"browser_specific_settings": {
"gecko": {
"id": 12345
```
2. Check the validation results
### What happened?
"There was a problem contacting the server" error message is received with a 500 response

### What did you expect to happen?
The error message should mention that the GUID needs to be in string format and no server error should be triggered
### Anything else we should know?
- reproduced on all AMO servers
- the linter already handles this exception and has an error message for it:
```
Code Message Description File Line Column
MANIFEST_FIELD_INVALID "/browser_specific_s… See https://mzl.la/1ZOhoEN (MDN Docs) for manifest.json
ettings/gecko/id" more information.
should be string
```
| 1.0 | Uploading an add-on with a GUID in non string format triggers a server error - ### Describe the problem and steps to reproduce it:
1. Submit an add-on with a GUID in non string format, for example:
```
"browser_specific_settings": {
"gecko": {
"id": 12345
```
2. Check the validation results
### What happened?
"There was a problem contacting the server" error message is received with a 500 response

### What did you expect to happen?
The error message should mention that the GUID needs to be in string format and no server error should be triggered
### Anything else we should know?
- reproduced on all AMO servers
- the linter already handles this exception and has an error message for it:
```
Code Message Description File Line Column
MANIFEST_FIELD_INVALID "/browser_specific_s… See https://mzl.la/1ZOhoEN (MDN Docs) for manifest.json
ettings/gecko/id" more information.
should be string
```
| non_defect | uploading an add on with a guid in non string format triggers a server error describe the problem and steps to reproduce it submit an add on with a guid in non string format for example browser specific settings gecko id check the validation results what happened there was a problem contacting the server error message is received with a response what did you expect to happen the error message should mention that the guid needs to be in string format and no server error should be triggered anything else we should know reproduced on all amo servers the linter already handles this exception and has an error message for it code message description file line column manifest field invalid browser specific s… see mdn docs for manifest json ettings gecko id more information should be string | 0 |
61,157 | 17,023,620,285 | IssuesEvent | 2021-07-03 02:57:49 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | oneway arrows not visible | Component: mapnik Priority: major Resolution: invalid Type: defect | **[Submitted to the original trac issue database at 10.07pm, Wednesday, 28th July 2010]**
It would seem when the road name just barely fits the length of the road it can completely hide the oneway arrows making navigation rather exciting. When you don't see the road names at least you know to zoom in to see it, but when you see oneway arrows all around it's not so obvious.
[http://www.openstreetmap.org/?lat=51.49191&lon=-0.19612&zoom=17&layers=M For example here] Nevern Square is oneway all the way; the north-south stretch is split to nortbound and southbound oneways, but it looks very much like two-way street. | 1.0 | oneway arrows not visible - **[Submitted to the original trac issue database at 10.07pm, Wednesday, 28th July 2010]**
It would seem when the road name just barely fits the length of the road it can completely hide the oneway arrows making navigation rather exciting. When you don't see the road names at least you know to zoom in to see it, but when you see oneway arrows all around it's not so obvious.
[http://www.openstreetmap.org/?lat=51.49191&lon=-0.19612&zoom=17&layers=M For example here] Nevern Square is oneway all the way; the north-south stretch is split to nortbound and southbound oneways, but it looks very much like two-way street. | defect | oneway arrows not visible it would seem when the road name just barely fits the length of the road it can completely hide the oneway arrows making navigation rather exciting when you don t see the road names at least you know to zoom in to see it but when you see oneway arrows all around it s not so obvious nevern square is oneway all the way the north south stretch is split to nortbound and southbound oneways but it looks very much like two way street | 1 |
26,474 | 4,726,256,857 | IssuesEvent | 2016-10-18 09:37:52 | PowerDNS/pdns | https://api.github.com/repos/PowerDNS/pdns | closed | API zone info returns an url without a starting / | auth defect web | When querying the API for zones with: `curl -H 'X-API-Key: secret' http://127.0.0.1:8081/api/v1/servers/localhost/zones` the returned json contains an url to the api zone info. This url does not start with a `/` whereas the docs specify that it should. (Other url fields from the api do start with a `/`).
The 3.4.x api does not seem to have this issue
Powerdns version 4.0.1
Distribution: Gentoo
Reproduce:
1. enable api
2. curl the api along the lines of `curl -H 'X-API-Key: secret' http://127.0.0.1:8081/api/v1/servers/localhost/zones`
expected behaviour:
URL field starts with a `/`, for example: `"url": "/api/v1/servers/localhost/zones/test.example.org."`
Observed behavior
URL does not start with a slash, for example: "url": "api/v1/servers/localhost/zones/test.example.org."
Proposed fix:
Add a `/` at the relevant place in getZoneInfo in ws-auth.cc | 1.0 | API zone info returns an url without a starting / - When querying the API for zones with: `curl -H 'X-API-Key: secret' http://127.0.0.1:8081/api/v1/servers/localhost/zones` the returned json contains an url to the api zone info. This url does not start with a `/` whereas the docs specify that it should. (Other url fields from the api do start with a `/`).
The 3.4.x api does not seem to have this issue
Powerdns version 4.0.1
Distribution: Gentoo
Reproduce:
1. enable api
2. curl the api along the lines of `curl -H 'X-API-Key: secret' http://127.0.0.1:8081/api/v1/servers/localhost/zones`
expected behaviour:
URL field starts with a `/`, for example: `"url": "/api/v1/servers/localhost/zones/test.example.org."`
Observed behavior
URL does not start with a slash, for example: "url": "api/v1/servers/localhost/zones/test.example.org."
Proposed fix:
Add a `/` at the relevant place in getZoneInfo in ws-auth.cc | defect | api zone info returns an url without a starting when querying the api for zones with curl h x api key secret the returned json contains an url to the api zone info this url does not start with a whereas the docs specify that it should other url fields from the api do start with a the x api does not seem to have this issue powerdns version distribution gentoo reproduce enable api curl the api along the lines of curl h x api key secret expected behaviour url field starts with a for example url api servers localhost zones test example org observed behavior url does not start with a slash for example url api servers localhost zones test example org proposed fix add a at the relevant place in getzoneinfo in ws auth cc | 1 |
16,228 | 2,878,875,574 | IssuesEvent | 2015-06-10 06:00:50 | msg4real/pygooglevoice | https://api.github.com/repos/msg4real/pygooglevoice | closed | The comment in settings.py about forwardingNumber is wrong | auto-migrated Priority-Medium Type-Defect | ```
# Number to place calls from (eg, your google voice number)
forwardingNumber=
/snip
The forwarding number is *NOT* your google voice number. It is the number you
want GV to connect the call to. This confuses people trying to setup gvoice to
make call by running 'gvoice call 8005551212'
```
Original issue reported on code.google.com by `prog...@gmail.com` on 22 Sep 2010 at 9:35 | 1.0 | The comment in settings.py about forwardingNumber is wrong - ```
# Number to place calls from (eg, your google voice number)
forwardingNumber=
/snip
The forwarding number is *NOT* your google voice number. It is the number you
want GV to connect the call to. This confuses people trying to setup gvoice to
make call by running 'gvoice call 8005551212'
```
Original issue reported on code.google.com by `prog...@gmail.com` on 22 Sep 2010 at 9:35 | defect | the comment in settings py about forwardingnumber is wrong number to place calls from eg your google voice number forwardingnumber snip the forwarding number is not your google voice number it is the number you want gv to connect the call to this confuses people trying to setup gvoice to make call by running gvoice call original issue reported on code google com by prog gmail com on sep at | 1 |
39,680 | 5,116,571,411 | IssuesEvent | 2017-01-07 05:18:02 | caseyg/knutepunkt2017 | https://api.github.com/repos/caseyg/knutepunkt2017 | opened | 8. Pettersson, Hamlets, Vampires, and the Italian Alps | design | Paper. Has images that we want to make bigger. | 1.0 | 8. Pettersson, Hamlets, Vampires, and the Italian Alps - Paper. Has images that we want to make bigger. | non_defect | pettersson hamlets vampires and the italian alps paper has images that we want to make bigger | 0 |
76,178 | 26,276,535,376 | IssuesEvent | 2023-01-06 22:48:58 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | closed | Linux Kernel 6.2 renamed get_acl to get_inode_acl | Type: Defect | ### System information
Type | Version/Name
--- | ---
Distribution Name | Gentoo
Distribution Version | -
Kernel Version | `next-20221220`
Architecture | LoongArch
OpenZFS Version | 2.1.99-1641_gc935fe2e9
### Describe the problem you're observing
```
checking whether iops->get_acl() exists... configure: error:
*** None of the expected "iops->get_acl()" interfaces were detected.
*** This may be because your kernel version is newer than what is
*** supported, or you are using a patched custom kernel with
*** incompatible modifications.
***
*** ZFS Version: zfs-2.1.99-1641_gc935fe2e9
*** Compatible Kernels: 3.10 - 6.0
```
⚠️ ⚠️ **Kernel renamed `get_acl` to `get_inode_acl`: https://github.com/torvalds/linux/commit/cac2f8b8d8b50ef32b3e34f6dcbbf08937e4f616** ⚠️ ⚠️
### Describe how to reproduce the problem
Build oepnzfs master with a post-6.1 kernel.
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
```
# make modules -C /usr/src/linux M=/var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu
make: Entering directory '/usr/src/linux-loongson'
CC [M] /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.o
/var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.c:53:36: error: initialization of ‘struct posix_acl * (*)(struct user_namespace *, struct dentry *, int)’ from incompatible pointer type ‘struct posix_acl * (*)(struct inode *, int, bool)’ {aka ‘struct posix_acl * (*)(struct inode *, int, _Bool)’} [-Werror=incompatible-pointer-types]
53 | .get_acl = get_acl_fn,
| ^~~~~~~~~~
/var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.c:53:36: note: (near initialization for ‘iops.get_acl’)
cc1: all warnings being treated as errors
make[1]: *** [scripts/Makefile.build:250: /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.o] Error 1
make: *** [Makefile:1994: /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu] Error 2
``` | 1.0 | Linux Kernel 6.2 renamed get_acl to get_inode_acl - ### System information
Type | Version/Name
--- | ---
Distribution Name | Gentoo
Distribution Version | -
Kernel Version | `next-20221220`
Architecture | LoongArch
OpenZFS Version | 2.1.99-1641_gc935fe2e9
### Describe the problem you're observing
```
checking whether iops->get_acl() exists... configure: error:
*** None of the expected "iops->get_acl()" interfaces were detected.
*** This may be because your kernel version is newer than what is
*** supported, or you are using a patched custom kernel with
*** incompatible modifications.
***
*** ZFS Version: zfs-2.1.99-1641_gc935fe2e9
*** Compatible Kernels: 3.10 - 6.0
```
⚠️ ⚠️ **Kernel renamed `get_acl` to `get_inode_acl`: https://github.com/torvalds/linux/commit/cac2f8b8d8b50ef32b3e34f6dcbbf08937e4f616** ⚠️ ⚠️
### Describe how to reproduce the problem
Build oepnzfs master with a post-6.1 kernel.
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
```
# make modules -C /usr/src/linux M=/var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu
make: Entering directory '/usr/src/linux-loongson'
CC [M] /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.o
/var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.c:53:36: error: initialization of ‘struct posix_acl * (*)(struct user_namespace *, struct dentry *, int)’ from incompatible pointer type ‘struct posix_acl * (*)(struct inode *, int, bool)’ {aka ‘struct posix_acl * (*)(struct inode *, int, _Bool)’} [-Werror=incompatible-pointer-types]
53 | .get_acl = get_acl_fn,
| ^~~~~~~~~~
/var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.c:53:36: note: (near initialization for ‘iops.get_acl’)
cc1: all warnings being treated as errors
make[1]: *** [scripts/Makefile.build:250: /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu/inode_operations_get_acl_rcu.o] Error 1
make: *** [Makefile:1994: /var/tmp/portage/sys-fs/zfs-loong-kmod-9999/work/zfs-loong-kmod-9999/build/inode_operations_get_acl_rcu] Error 2
``` | defect | linux kernel renamed get acl to get inode acl system information type version name distribution name gentoo distribution version kernel version next architecture loongarch openzfs version describe the problem you re observing checking whether iops get acl exists configure error none of the expected iops get acl interfaces were detected this may be because your kernel version is newer than what is supported or you are using a patched custom kernel with incompatible modifications zfs version zfs compatible kernels ⚠️ ⚠️ kernel renamed get acl to get inode acl ⚠️ ⚠️ describe how to reproduce the problem build oepnzfs master with a post kernel include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with make modules c usr src linux m var tmp portage sys fs zfs loong kmod work zfs loong kmod build inode operations get acl rcu make entering directory usr src linux loongson cc var tmp portage sys fs zfs loong kmod work zfs loong kmod build inode operations get acl rcu inode operations get acl rcu o var tmp portage sys fs zfs loong kmod work zfs loong kmod build inode operations get acl rcu inode operations get acl rcu c error initialization of ‘struct posix acl struct user namespace struct dentry int ’ from incompatible pointer type ‘struct posix acl struct inode int bool ’ aka ‘struct posix acl struct inode int bool ’ get acl get acl fn var tmp portage sys fs zfs loong kmod work zfs loong kmod build inode operations get acl rcu inode operations get acl rcu c note near initialization for ‘iops get acl’ all warnings being treated as errors make error make error | 1 |
19,006 | 10,312,367,584 | IssuesEvent | 2019-08-29 19:41:16 | jamijam/WebGoat-Legacy | https://api.github.com/repos/jamijam/WebGoat-Legacy | opened | CVE-2019-12384 (Medium) detected in jackson-databind-2.0.4.jar | security vulnerability | ## CVE-2019-12384 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.0.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: /WebGoat-Legacy/pom.xml</p>
<p>Path to vulnerable library: 2/repository/com/fasterxml/jackson/core/jackson-databind/2.0.4/jackson-databind-2.0.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.0.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jamijam/WebGoat-Legacy/commit/47e282c80c0e0057166440fc7140b77c4f06feb6">47e282c80c0e0057166440fc7140b77c4f06feb6</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.9.1 might allow attackers to have a variety of impacts by leveraging failure to block the logback-core class from polymorphic deserialization. Depending on the classpath content, remote code execution may be possible.
<p>Publish Date: 2019-06-24
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384>CVE-2019-12384</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384</a></p>
<p>Release Date: 2019-08-12</p>
<p>Fix Resolution: 2.9.9.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-12384 (Medium) detected in jackson-databind-2.0.4.jar - ## CVE-2019-12384 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.0.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: /WebGoat-Legacy/pom.xml</p>
<p>Path to vulnerable library: 2/repository/com/fasterxml/jackson/core/jackson-databind/2.0.4/jackson-databind-2.0.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.0.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jamijam/WebGoat-Legacy/commit/47e282c80c0e0057166440fc7140b77c4f06feb6">47e282c80c0e0057166440fc7140b77c4f06feb6</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.9.1 might allow attackers to have a variety of impacts by leveraging failure to block the logback-core class from polymorphic deserialization. Depending on the classpath content, remote code execution may be possible.
<p>Publish Date: 2019-06-24
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384>CVE-2019-12384</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12384</a></p>
<p>Release Date: 2019-08-12</p>
<p>Fix Resolution: 2.9.9.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to dependency file webgoat legacy pom xml path to vulnerable library repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before might allow attackers to have a variety of impacts by leveraging failure to block the logback core class from polymorphic deserialization depending on the classpath content remote code execution may be possible publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
680,817 | 23,286,793,111 | IssuesEvent | 2022-08-05 17:25:08 | mm-ninja-turtles/bulbasaur-express | https://api.github.com/repos/mm-ninja-turtles/bulbasaur-express | closed | [bug]: remove `response` object from resolver function args. | type: bug priority: high task: done | ## Problem
`resolver()` function currently provide express's `response` object and it should be removed since it will give the user the freedom to response as they wanted and it'll create the problem of response data is not matching with response schema.
## Solution
1. remove `response` object from args
1. #16 will solve the necessity of `response` object inside resolver function | 1.0 | [bug]: remove `response` object from resolver function args. - ## Problem
`resolver()` function currently provide express's `response` object and it should be removed since it will give the user the freedom to response as they wanted and it'll create the problem of response data is not matching with response schema.
## Solution
1. remove `response` object from args
1. #16 will solve the necessity of `response` object inside resolver function | non_defect | remove response object from resolver function args problem resolver function currently provide express s response object and it should be removed since it will give the user the freedom to response as they wanted and it ll create the problem of response data is not matching with response schema solution remove response object from args will solve the necessity of response object inside resolver function | 0 |
231,045 | 18,735,238,071 | IssuesEvent | 2021-11-04 06:15:17 | kartoza/django-bims | https://api.github.com/repos/kartoza/django-bims | closed | Remove date when uploading unpublished datasets | testing SIZE 2 Checked by FRC bug-report Source references Repeat offender - previous ticket | Please remove date when uploading unpublished data and database.
It is creating extra work for us having to go and combine unpublished datasets
I fixed all of these source reference a few weeks ago and now they are all wrong again :( | 1.0 | Remove date when uploading unpublished datasets - Please remove date when uploading unpublished data and database.
It is creating extra work for us having to go and combine unpublished datasets
I fixed all of these source reference a few weeks ago and now they are all wrong again :( | non_defect | remove date when uploading unpublished datasets please remove date when uploading unpublished data and database it is creating extra work for us having to go and combine unpublished datasets i fixed all of these source reference a few weeks ago and now they are all wrong again | 0 |
79,503 | 22,782,906,821 | IssuesEvent | 2022-07-08 22:33:22 | chaotic-aur/packages | https://api.github.com/repos/chaotic-aur/packages | closed | [Outdated] ffmpeg-full | request:new-pkg request:rebuild-pkg priority:medium bug:PKGBUILD | ### If available, link to the latest build
[ffmpeg-full.log](https://builds.garudalinux.org/repos/chaotic-aur/logs/ffmpeg-full.log)
### Package name
`ffmpeg-full`
### Latest build
`None`
### Latest version available
`5.0.1-2`
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### More information
Adding missing deps in https://github.com/chaotic-aur/packages/commit/e2072dd61dca0b0b8c9732e76803278bba06ee27 but `python-pocketsphinx-git` still won't build: [python-pocketsphinx-git.log](https://builds.garudalinux.org/repos/chaotic-aur/logs/python-pocketsphinx-git.log) | 2.0 | [Outdated] ffmpeg-full - ### If available, link to the latest build
[ffmpeg-full.log](https://builds.garudalinux.org/repos/chaotic-aur/logs/ffmpeg-full.log)
### Package name
`ffmpeg-full`
### Latest build
`None`
### Latest version available
`5.0.1-2`
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### More information
Adding missing deps in https://github.com/chaotic-aur/packages/commit/e2072dd61dca0b0b8c9732e76803278bba06ee27 but `python-pocketsphinx-git` still won't build: [python-pocketsphinx-git.log](https://builds.garudalinux.org/repos/chaotic-aur/logs/python-pocketsphinx-git.log) | non_defect | ffmpeg full if available link to the latest build package name ffmpeg full latest build none latest version available have you tested if the package builds in a clean chroot yes more information adding missing deps in but python pocketsphinx git still won t build | 0 |
9,137 | 2,615,133,741 | IssuesEvent | 2015-03-01 06:04:46 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | opened | With null accountName, get infinity loop | auto-migrated Priority-Medium Type-Defect | ```
Version of google-api-java-client (e.g. 1.15.0-rc)?
1.17.0-rc
Java environment (e.g. Java 6, Android 2.3, App Engine)?
Android 4.2.2
Describe the problem.
from source GoogleAccountCredential.java:
public String getToken() throws IOException, GoogleAuthException {
if (backOff != null) {
backOff.reset();
}
while (true) {
try {
return GoogleAuthUtil.getToken(context, accountName, scope);
} catch (IOException e) {
// network or server error, so retry using back-off policy
try {
if (backOff == null || !BackOffUtils.next(sleeper, backOff)) {
throw e;
}
} catch (InterruptedException e2) {
// ignore
}
}
}
If accountName = "" (something wrong, yes), GoogleAuthUtil.getToken runs
forever. No exception or something.
How would you expect it to be fixed?
Exception?
```
Original issue reported on code.google.com by `rusxa...@gmail.com` on 14 Oct 2013 at 1:01 | 1.0 | With null accountName, get infinity loop - ```
Version of google-api-java-client (e.g. 1.15.0-rc)?
1.17.0-rc
Java environment (e.g. Java 6, Android 2.3, App Engine)?
Android 4.2.2
Describe the problem.
from source GoogleAccountCredential.java:
public String getToken() throws IOException, GoogleAuthException {
if (backOff != null) {
backOff.reset();
}
while (true) {
try {
return GoogleAuthUtil.getToken(context, accountName, scope);
} catch (IOException e) {
// network or server error, so retry using back-off policy
try {
if (backOff == null || !BackOffUtils.next(sleeper, backOff)) {
throw e;
}
} catch (InterruptedException e2) {
// ignore
}
}
}
If accountName = "" (something wrong, yes), GoogleAuthUtil.getToken runs
forever. No exception or something.
How would you expect it to be fixed?
Exception?
```
Original issue reported on code.google.com by `rusxa...@gmail.com` on 14 Oct 2013 at 1:01 | defect | with null accountname get infinity loop version of google api java client e g rc rc java environment e g java android app engine android describe the problem from source googleaccountcredential java public string gettoken throws ioexception googleauthexception if backoff null backoff reset while true try return googleauthutil gettoken context accountname scope catch ioexception e network or server error so retry using back off policy try if backoff null backoffutils next sleeper backoff throw e catch interruptedexception ignore if accountname something wrong yes googleauthutil gettoken runs forever no exception or something how would you expect it to be fixed exception original issue reported on code google com by rusxa gmail com on oct at | 1 |
36,422 | 7,928,753,966 | IssuesEvent | 2018-07-06 12:53:41 | otros-systems/otroslogviewer | https://api.github.com/repos/otros-systems/otroslogviewer | closed | Listening to udp not possible | Priority-Medium Type-Defect | ```
When starting a socket listener, it always starts listening to tcp messages.
However, it would be great if also listening to udp messages was supported. For
example the log4judp logging backend of log4cplus could be used with the
otroslogviewer then.
```
Original issue reported on code.google.com by `jan.rue...@dacuda.com` on 25 Sep 2013 at 11:45
| 1.0 | Listening to udp not possible - ```
When starting a socket listener, it always starts listening to tcp messages.
However, it would be great if also listening to udp messages was supported. For
example the log4judp logging backend of log4cplus could be used with the
otroslogviewer then.
```
Original issue reported on code.google.com by `jan.rue...@dacuda.com` on 25 Sep 2013 at 11:45
| defect | listening to udp not possible when starting a socket listener it always starts listening to tcp messages however it would be great if also listening to udp messages was supported for example the logging backend of could be used with the otroslogviewer then original issue reported on code google com by jan rue dacuda com on sep at | 1 |
21,470 | 3,511,527,923 | IssuesEvent | 2016-01-10 10:29:53 | nielsAD/lape | https://api.github.com/repos/nielsAD/lape | closed | It's impossible with certain boolean comparisons. | auto-migrated Priority-High Type-Defect | ```
What will reproduce the problem?
if (SomeBool and True) then
What is the expected output? What do you see instead?
I get "It's impossible!" (Line 706 of lpvartypes_ord.pas)
Which version are you using?
c96f612e0066
Please provide any additional information below.
Seems to not be an issue if True is first.
```
Original issue reported on code.google.com by `Dgby...@gmail.com` on 7 Sep 2014 at 12:33 | 1.0 | It's impossible with certain boolean comparisons. - ```
What will reproduce the problem?
if (SomeBool and True) then
What is the expected output? What do you see instead?
I get "It's impossible!" (Line 706 of lpvartypes_ord.pas)
Which version are you using?
c96f612e0066
Please provide any additional information below.
Seems to not be an issue if True is first.
```
Original issue reported on code.google.com by `Dgby...@gmail.com` on 7 Sep 2014 at 12:33 | defect | it s impossible with certain boolean comparisons what will reproduce the problem if somebool and true then what is the expected output what do you see instead i get it s impossible line of lpvartypes ord pas which version are you using please provide any additional information below seems to not be an issue if true is first original issue reported on code google com by dgby gmail com on sep at | 1 |
28,079 | 5,184,504,936 | IssuesEvent | 2017-01-20 06:31:44 | GarageGames/Torque3D | https://api.github.com/repos/GarageGames/Torque3D | closed | Terrain materials are double sided | Defect | I noticed when I fly under the terrain, it is still visible, this means the terrain materials are all double sided and this cannot be changed, since terrain materials, do not have the value to change if they are double sided or not.
Since double sided materials cost more to render, the terrain should not be double sided.
I also checked one version before and the terrain materials were all single sided like it should be, so they got broken in 3.9 update.
| 1.0 | Terrain materials are double sided - I noticed when I fly under the terrain, it is still visible, this means the terrain materials are all double sided and this cannot be changed, since terrain materials, do not have the value to change if they are double sided or not.
Since double sided materials cost more to render, the terrain should not be double sided.
I also checked one version before and the terrain materials were all single sided like it should be, so they got broken in 3.9 update.
| defect | terrain materials are double sided i noticed when i fly under the terrain it is still visible this means the terrain materials are all double sided and this cannot be changed since terrain materials do not have the value to change if they are double sided or not since double sided materials cost more to render the terrain should not be double sided i also checked one version before and the terrain materials were all single sided like it should be so they got broken in update | 1 |
34,979 | 7,497,166,764 | IssuesEvent | 2018-04-08 17:00:56 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | [Bug?] Example of Bland's Rule for optimize.linprog (simplex) cycling/preventing termination | defect scipy.optimize | Forcing Bland's Rule with the "bland":True options seems to prevent the code example below from terminating, while the purpose of Bland's Rule is to make sure it terminates. The problem is solved quickly when not forcing Bland's Rule, giving an optimal value of -6.044533469014448.
By "not terminating" i mean that it's still running after running for a couple of hours on a MacBook Pro, while my own implementation of Bland's Rule finds it in seconds.
### Reproducing code example:
```
import numpy as np
import scipy.optimize as opt
import sys
np.random.seed(4)
m = int(np.round(10 * np.exp(np.log(50) * np.random.rand())))
n = int(np.round(10 * np.exp(np.log(50) * np.random.rand())))
c, A, b = np.round(10*np.random.randn(n)),np.round(10*np.random.randn(m,n)),np.round(10*np.abs(np.random.randn(m)))
opt.linprog(-1*c, A, b, options={"maxiter": sys.maxsize, "bland": True})
```
### Scipy/Numpy/Python version information:
1.0.0 1.14.0 sys.version_info(major=3, minor=6, micro=4, releaselevel='final', serial=0)
| 1.0 | [Bug?] Example of Bland's Rule for optimize.linprog (simplex) cycling/preventing termination - Forcing Bland's Rule with the "bland":True options seems to prevent the code example below from terminating, while the purpose of Bland's Rule is to make sure it terminates. The problem is solved quickly when not forcing Bland's Rule, giving an optimal value of -6.044533469014448.
By "not terminating" i mean that it's still running after running for a couple of hours on a MacBook Pro, while my own implementation of Bland's Rule finds it in seconds.
### Reproducing code example:
```
import numpy as np
import scipy.optimize as opt
import sys
np.random.seed(4)
m = int(np.round(10 * np.exp(np.log(50) * np.random.rand())))
n = int(np.round(10 * np.exp(np.log(50) * np.random.rand())))
c, A, b = np.round(10*np.random.randn(n)),np.round(10*np.random.randn(m,n)),np.round(10*np.abs(np.random.randn(m)))
opt.linprog(-1*c, A, b, options={"maxiter": sys.maxsize, "bland": True})
```
### Scipy/Numpy/Python version information:
1.0.0 1.14.0 sys.version_info(major=3, minor=6, micro=4, releaselevel='final', serial=0)
| defect | example of bland s rule for optimize linprog simplex cycling preventing termination forcing bland s rule with the bland true options seems to prevent the code example below from terminating while the purpose of bland s rule is to make sure it terminates the problem is solved quickly when not forcing bland s rule giving an optimal value of by not terminating i mean that it s still running after running for a couple of hours on a macbook pro while my own implementation of bland s rule finds it in seconds reproducing code example import numpy as np import scipy optimize as opt import sys np random seed m int np round np exp np log np random rand n int np round np exp np log np random rand c a b np round np random randn n np round np random randn m n np round np abs np random randn m opt linprog c a b options maxiter sys maxsize bland true scipy numpy python version information sys version info major minor micro releaselevel final serial | 1 |
12,444 | 2,700,140,850 | IssuesEvent | 2015-04-03 22:54:05 | netty/netty | https://api.github.com/repos/netty/netty | closed | HTTP/2 RST_STREAM frame for an IDLE stream should result in connection error | defect | http://http2.github.io/http2-spec/index.html#RST_STREAM
> RST_STREAM frames MUST NOT be sent for a stream in the "idle" state. If a RST_STREAM frame identifying an idle stream is received, the recipient MUST treat this as a connection error (Section 5.4.1) of type PROTOCOL_ERROR. | 1.0 | HTTP/2 RST_STREAM frame for an IDLE stream should result in connection error - http://http2.github.io/http2-spec/index.html#RST_STREAM
> RST_STREAM frames MUST NOT be sent for a stream in the "idle" state. If a RST_STREAM frame identifying an idle stream is received, the recipient MUST treat this as a connection error (Section 5.4.1) of type PROTOCOL_ERROR. | defect | http rst stream frame for an idle stream should result in connection error rst stream frames must not be sent for a stream in the idle state if a rst stream frame identifying an idle stream is received the recipient must treat this as a connection error section of type protocol error | 1 |
39,390 | 6,741,336,636 | IssuesEvent | 2017-10-20 00:03:09 | scrabill/how-many-days-until-halloween | https://api.github.com/repos/scrabill/how-many-days-until-halloween | closed | Documentation: Update CONTRIBUTE.md | documentation Hacktoberfest help wanted | Include the following
- How to contribute
- What is currently there / how it works
- Links to resources
| 1.0 | Documentation: Update CONTRIBUTE.md - Include the following
- How to contribute
- What is currently there / how it works
- Links to resources
| non_defect | documentation update contribute md include the following how to contribute what is currently there how it works links to resources | 0 |
51,772 | 13,211,304,546 | IssuesEvent | 2020-08-15 22:10:39 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | Link to PROPOSAL project (Trac #1019) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1019">https://code.icecube.wisc.edu/projects/icecube/ticket/1019</a>, reported by icecube</summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-06-11T18:38:27",
"_ts": "1434047907705473",
"description": "Hi,\n\nCould you please provide a link to the paper describing how the PROPOSAL icesim meta-project work?\n\nThanks.",
"reporter": "icecube",
"cc": "jtatar@lbl.gov",
"resolution": "wontfix",
"time": "2015-06-11T17:37:23",
"component": "cmake",
"summary": "Link to PROPOSAL project",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Link to PROPOSAL project (Trac #1019) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1019">https://code.icecube.wisc.edu/projects/icecube/ticket/1019</a>, reported by icecube</summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-06-11T18:38:27",
"_ts": "1434047907705473",
"description": "Hi,\n\nCould you please provide a link to the paper describing how the PROPOSAL icesim meta-project work?\n\nThanks.",
"reporter": "icecube",
"cc": "jtatar@lbl.gov",
"resolution": "wontfix",
"time": "2015-06-11T17:37:23",
"component": "cmake",
"summary": "Link to PROPOSAL project",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| defect | link to proposal project trac migrated from json status closed changetime ts description hi n ncould you please provide a link to the paper describing how the proposal icesim meta project work n nthanks reporter icecube cc jtatar lbl gov resolution wontfix time component cmake summary link to proposal project priority normal keywords milestone owner type defect | 1 |
338,745 | 24,597,597,255 | IssuesEvent | 2022-10-14 09:37:56 | Thaza-Kun/sarjana | https://api.github.com/repos/Thaza-Kun/sarjana | opened | Check The Unit for $a$ and $K$ | documentation | The values of $a$ should be the inverse of $K$ yet in `thesis\literature.qmd`, the units are not entirely the same:
$a=4.148 806 4239(11) \text{GH}^2\text{pc}^{-1}\text{cm}^3\text{ms}$
$K=241.033 1786(66) \text{GH}^{-2}\text{pc}\text{cm}^{-3}\text{s}$
Note: In $a$, it is ...ms while in $K$, it is ...s
<!-- Edit the body of your new issue then click the ✓ "Create Issue" button in the top right of the editor. The first line will be the issue title. Assignees and Labels follow after a blank line. Leave an empty line before beginning the body of the issue. --> | 1.0 | Check The Unit for $a$ and $K$ - The values of $a$ should be the inverse of $K$ yet in `thesis\literature.qmd`, the units are not entirely the same:
$a=4.148 806 4239(11) \text{GH}^2\text{pc}^{-1}\text{cm}^3\text{ms}$
$K=241.033 1786(66) \text{GH}^{-2}\text{pc}\text{cm}^{-3}\text{s}$
Note: In $a$, it is ...ms while in $K$, it is ...s
<!-- Edit the body of your new issue then click the ✓ "Create Issue" button in the top right of the editor. The first line will be the issue title. Assignees and Labels follow after a blank line. Leave an empty line before beginning the body of the issue. --> | non_defect | check the unit for a and k the values of a should be the inverse of k yet in thesis literature qmd the units are not entirely the same a text gh text pc text cm text ms k text gh text pc text cm text s note in a it is ms while in k it is s | 0 |
21,992 | 3,587,833,741 | IssuesEvent | 2016-01-30 16:15:46 | dkfans/keeperfx | https://api.github.com/repos/dkfans/keeperfx | reopened | Level Kari-Mar automatic victory due to Computer Player suicide | Component-CompPlayer Component-Maps Priority-Low Status-Duplicate Type-Defect | Originally reported on Google Code with ID 213
```
Free Play level Kari-Mar on version 0.4.4. Early in the game you suddenly win, without
ever seeing an enemy. Most likely map invites Computer Player to dig to strong heroes.
Savegame attached.
```
Reported by `Loobinex` on 2014-01-15 15:40:39
<hr>
* *Attachment: [fx1g0004.sav](https://storage.googleapis.com/google-code-attachments/keeperfx/issue-213/comment-0/fx1g0004.sav)* | 1.0 | Level Kari-Mar automatic victory due to Computer Player suicide - Originally reported on Google Code with ID 213
```
Free Play level Kari-Mar on version 0.4.4. Early in the game you suddenly win, without
ever seeing an enemy. Most likely map invites Computer Player to dig to strong heroes.
Savegame attached.
```
Reported by `Loobinex` on 2014-01-15 15:40:39
<hr>
* *Attachment: [fx1g0004.sav](https://storage.googleapis.com/google-code-attachments/keeperfx/issue-213/comment-0/fx1g0004.sav)* | defect | level kari mar automatic victory due to computer player suicide originally reported on google code with id free play level kari mar on version early in the game you suddenly win without ever seeing an enemy most likely map invites computer player to dig to strong heroes savegame attached reported by loobinex on attachment | 1 |
10,473 | 2,622,165,432 | IssuesEvent | 2015-03-04 00:12:05 | byzhang/graphchi | https://api.github.com/repos/byzhang/graphchi | opened | Cannot make sharder_basic - fixed by adding an extra include | auto-migrated Priority-Medium Type-Defect | ```
I tried to build the sharder_basic program with the latest download and it
failed with this error:
g++ -g -O3 -I/usr/local/include/ -I./src/ -fopenmp -Wall -Wno-strict-aliasing
src/preprocessing/sharder_basic.cpp -o bin/sharder_basic
In file included from ./src/preprocessing/conversions.hpp:36,
from src/preprocessing/sharder_basic.cpp:35:
./src/preprocessing/sharder.hpp: In member function ‘virtual void
graphchi::sharder<EdgeDataType>::write_shards()’:
./src/preprocessing/sharder.hpp:465: error: ‘degree’ was not declared in
this scope
./src/preprocessing/sharder.hpp:467: error: ‘degrees’ was not declared in
this scope
./src/preprocessing/sharder.hpp:469: error: expected primary-expression before
‘)’ token
./src/preprocessing/sharder.hpp:469: error: expected `;' before ‘calloc’
make: *** [sharder_basic] Error 1
Including the degree_data header seemed to fix this and I was able to build and
run basic_sharder after this
#include "engine/auxdata/degree_data.hpp"
```
Original issue reported on code.google.com by `Jennifer...@gmail.com` on 13 Jan 2013 at 3:16 | 1.0 | Cannot make sharder_basic - fixed by adding an extra include - ```
I tried to build the sharder_basic program with the latest download and it
failed with this error:
g++ -g -O3 -I/usr/local/include/ -I./src/ -fopenmp -Wall -Wno-strict-aliasing
src/preprocessing/sharder_basic.cpp -o bin/sharder_basic
In file included from ./src/preprocessing/conversions.hpp:36,
from src/preprocessing/sharder_basic.cpp:35:
./src/preprocessing/sharder.hpp: In member function ‘virtual void
graphchi::sharder<EdgeDataType>::write_shards()’:
./src/preprocessing/sharder.hpp:465: error: ‘degree’ was not declared in
this scope
./src/preprocessing/sharder.hpp:467: error: ‘degrees’ was not declared in
this scope
./src/preprocessing/sharder.hpp:469: error: expected primary-expression before
‘)’ token
./src/preprocessing/sharder.hpp:469: error: expected `;' before ‘calloc’
make: *** [sharder_basic] Error 1
Including the degree_data header seemed to fix this and I was able to build and
run basic_sharder after this
#include "engine/auxdata/degree_data.hpp"
```
Original issue reported on code.google.com by `Jennifer...@gmail.com` on 13 Jan 2013 at 3:16 | defect | cannot make sharder basic fixed by adding an extra include i tried to build the sharder basic program with the latest download and it failed with this error g g i usr local include i src fopenmp wall wno strict aliasing src preprocessing sharder basic cpp o bin sharder basic in file included from src preprocessing conversions hpp from src preprocessing sharder basic cpp src preprocessing sharder hpp in member function ‘virtual void graphchi sharder write shards ’ src preprocessing sharder hpp error ‘degree’ was not declared in this scope src preprocessing sharder hpp error ‘degrees’ was not declared in this scope src preprocessing sharder hpp error expected primary expression before ‘ ’ token src preprocessing sharder hpp error expected before ‘calloc’ make error including the degree data header seemed to fix this and i was able to build and run basic sharder after this include engine auxdata degree data hpp original issue reported on code google com by jennifer gmail com on jan at | 1 |
20,097 | 3,295,315,224 | IssuesEvent | 2015-10-31 20:48:32 | chief-atx/bcmon | https://api.github.com/repos/chief-atx/bcmon | closed | Galaxy s 4 | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
Dont finding anything.
What version of the product are you using? On what operating system?
galaxy s 4 cm 12.1 5.1.1
Please provide any additional information below.
Maybe u can help me to connect tl-wn722 cos i need mon mode on galaxy s 4
```
Original issue reported on code.google.com by `nshasher...@gmail.com` on 23 Aug 2015 at 12:23 | 1.0 | Galaxy s 4 - ```
What steps will reproduce the problem?
Dont finding anything.
What version of the product are you using? On what operating system?
galaxy s 4 cm 12.1 5.1.1
Please provide any additional information below.
Maybe u can help me to connect tl-wn722 cos i need mon mode on galaxy s 4
```
Original issue reported on code.google.com by `nshasher...@gmail.com` on 23 Aug 2015 at 12:23 | defect | galaxy s what steps will reproduce the problem dont finding anything what version of the product are you using on what operating system galaxy s cm please provide any additional information below maybe u can help me to connect tl cos i need mon mode on galaxy s original issue reported on code google com by nshasher gmail com on aug at | 1 |
63,728 | 17,872,009,895 | IssuesEvent | 2021-09-06 17:09:04 | martinrotter/rssguard | https://api.github.com/repos/martinrotter/rssguard | closed | Linking fails when version 3.9.2 is installed | Type-Defect Status-Not-Enough-Data | When building version 4.0.1 while version 3.9.2 is installed, linking fails, being unable to resolve `parseCmdArgumentsFromOtherInstance(const QString& message)` symbol. Removing or renaming `/usr/local/lib/librssguard.so` helps, but the following small change to `src/rssguard/rssguard.pro` would make the build more robust:
```patch
-unix: LIBS += -L$$OUT_PWD/../librssguard/ -lrssguard
+unix: LIBS += $$OUT_PWD/../librssguard/librssguard.so
``` | 1.0 | Linking fails when version 3.9.2 is installed - When building version 4.0.1 while version 3.9.2 is installed, linking fails, being unable to resolve `parseCmdArgumentsFromOtherInstance(const QString& message)` symbol. Removing or renaming `/usr/local/lib/librssguard.so` helps, but the following small change to `src/rssguard/rssguard.pro` would make the build more robust:
```patch
-unix: LIBS += -L$$OUT_PWD/../librssguard/ -lrssguard
+unix: LIBS += $$OUT_PWD/../librssguard/librssguard.so
``` | defect | linking fails when version is installed when building version while version is installed linking fails being unable to resolve parsecmdargumentsfromotherinstance const qstring message symbol removing or renaming usr local lib librssguard so helps but the following small change to src rssguard rssguard pro would make the build more robust patch unix libs l out pwd librssguard lrssguard unix libs out pwd librssguard librssguard so | 1 |
17,624 | 3,012,774,652 | IssuesEvent | 2015-07-29 02:27:30 | yawlfoundation/yawl | https://api.github.com/repos/yawlfoundation/yawl | closed | Importing adds to specification in case specification is erroneous | auto-migrated Category-Component-Editor Priority-Medium Type-Defect | ```
When I *open* specification new60.ywl I get an error and the editor states
that the load file will be discarded. However, the specification is shown
and when I subsequently *import* another specification (e.g. new61.xml),
two nets are shown. This is not possible with specifications which load
normally as the import option is then disabled.
```
Original issue reported on code.google.com by `arthurte...@gmail.com` on 20 Aug 2008 at 7:49
Attachments:
* [net60.ywl](https://storage.googleapis.com/google-code-attachments/yawl/issue-108/comment-0/net60.ywl)
* [new61.xml](https://storage.googleapis.com/google-code-attachments/yawl/issue-108/comment-0/new61.xml)
| 1.0 | Importing adds to specification in case specification is erroneous - ```
When I *open* specification new60.ywl I get an error and the editor states
that the load file will be discarded. However, the specification is shown
and when I subsequently *import* another specification (e.g. new61.xml),
two nets are shown. This is not possible with specifications which load
normally as the import option is then disabled.
```
Original issue reported on code.google.com by `arthurte...@gmail.com` on 20 Aug 2008 at 7:49
Attachments:
* [net60.ywl](https://storage.googleapis.com/google-code-attachments/yawl/issue-108/comment-0/net60.ywl)
* [new61.xml](https://storage.googleapis.com/google-code-attachments/yawl/issue-108/comment-0/new61.xml)
| defect | importing adds to specification in case specification is erroneous when i open specification ywl i get an error and the editor states that the load file will be discarded however the specification is shown and when i subsequently import another specification e g xml two nets are shown this is not possible with specifications which load normally as the import option is then disabled original issue reported on code google com by arthurte gmail com on aug at attachments | 1 |
38,995 | 5,207,021,621 | IssuesEvent | 2017-01-24 22:14:56 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | We should support Load Balancer or remove from drop-down list | area/webhooks kind/bug status/resolved status/to-test | **Rancher Versions:** master 1/17
**Steps to Reproduce:**
1. Add a load balancer
2. Go to Add webhook and click on drop-down
**Results:** Load balancer is in list
**Expected:** We should support it or remove it from drop-down.
| 1.0 | We should support Load Balancer or remove from drop-down list - **Rancher Versions:** master 1/17
**Steps to Reproduce:**
1. Add a load balancer
2. Go to Add webhook and click on drop-down
**Results:** Load balancer is in list
**Expected:** We should support it or remove it from drop-down.
| non_defect | we should support load balancer or remove from drop down list rancher versions master steps to reproduce add a load balancer go to add webhook and click on drop down results load balancer is in list expected we should support it or remove it from drop down | 0 |
796,903 | 28,131,175,889 | IssuesEvent | 2023-03-31 23:24:59 | integrations/terraform-provider-github | https://api.github.com/repos/integrations/terraform-provider-github | closed | The repository_file data source throws an error when trying to access a non-existent file | Type: Bug Priority: Normal | ### Affected Resource(s)
Please list the resources as a list, for example:
- `data.github_repository_file`
### Expected Behavior
The `repository_file` data source allows checking if a specific file exists in a repository. The same way that `branch`data source, for example, does (see: https://github.com/integrations/terraform-provider-github/blob/main/github/data_source_github_branch.go#L52).
### Actual Behavior
The `repository_file` data source throws an error when trying to get a file which does not exist (see: https://github.com/integrations/terraform-provider-github/blob/main/github/data_source_github_repository_file.go#L82).
### References
- https://github.com/integrations/terraform-provider-github/pull/1129
| 1.0 | The repository_file data source throws an error when trying to access a non-existent file - ### Affected Resource(s)
Please list the resources as a list, for example:
- `data.github_repository_file`
### Expected Behavior
The `repository_file` data source allows checking if a specific file exists in a repository. The same way that `branch`data source, for example, does (see: https://github.com/integrations/terraform-provider-github/blob/main/github/data_source_github_branch.go#L52).
### Actual Behavior
The `repository_file` data source throws an error when trying to get a file which does not exist (see: https://github.com/integrations/terraform-provider-github/blob/main/github/data_source_github_repository_file.go#L82).
### References
- https://github.com/integrations/terraform-provider-github/pull/1129
| non_defect | the repository file data source throws an error when trying to access a non existent file affected resource s please list the resources as a list for example data github repository file expected behavior the repository file data source allows checking if a specific file exists in a repository the same way that branch data source for example does see actual behavior the repository file data source throws an error when trying to get a file which does not exist see references | 0 |
75,505 | 25,888,415,428 | IssuesEvent | 2022-12-14 16:06:14 | DependencyTrack/dependency-track | https://api.github.com/repos/DependencyTrack/dependency-track | opened | After Component update, only CVEs related to updated CPE should be available | defect in triage | ### Current Behavior
If we update the version and CPE of an existing component, all old CVEs not related to the new CPE and their audit histories are still available in Vulnerability Audit.
### Steps to Reproduce
1. Edit any exist component with outdated version and CPE >
Note available CVEs
2. Change version number and CPE to new one
3. After some time check for new CVEs >
Old and new CVEs are observed
### Expected Behavior
After updating the version and CPE of an existing component, only new CVEs and previously found CVEs related to the new CPE and their audit histories should be available in Vulnerability Audit.
### Dependency-Track Version
4.6.2
### Dependency-Track Distribution
Container Image
### Database Server
PostgreSQL
### Database Server Version
_No response_
### Browser
Mozilla Firefox
### Checklist
- [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues)
- [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported | 1.0 | After Component update, only CVEs related to updated CPE should be available - ### Current Behavior
If we update the version and CPE of an existing component, all old CVEs not related to the new CPE and their audit histories are still available in Vulnerability Audit.
### Steps to Reproduce
1. Edit any exist component with outdated version and CPE >
Note available CVEs
2. Change version number and CPE to new one
3. After some time check for new CVEs >
Old and new CVEs are observed
### Expected Behavior
After updating the version and CPE of an existing component, only new CVEs and previously found CVEs related to the new CPE and their audit histories should be available in Vulnerability Audit.
### Dependency-Track Version
4.6.2
### Dependency-Track Distribution
Container Image
### Database Server
PostgreSQL
### Database Server Version
_No response_
### Browser
Mozilla Firefox
### Checklist
- [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues)
- [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported | defect | after component update only cves related to updated cpe should be available current behavior if we update the version and cpe of an existing component all old cves not related to the new cpe and their audit histories are still available in vulnerability audit steps to reproduce edit any exist component with outdated version and cpe note available cves change version number and cpe to new one after some time check for new cves old and new cves are observed expected behavior after updating the version and cpe of an existing component only new cves and previously found cves related to the new cpe and their audit histories should be available in vulnerability audit dependency track version dependency track distribution container image database server postgresql database server version no response browser mozilla firefox checklist i have read and understand the i have checked the for whether this defect was already reported | 1 |
16,739 | 2,941,305,424 | IssuesEvent | 2015-07-02 06:49:26 | tnt944445/reaver-wps | https://api.github.com/repos/tnt944445/reaver-wps | closed | Try multiple BSSIDs, round robin | auto-migrated Priority-Triage Type-Defect | ```
I found myself sitting an entire evening switching between two routers I was
testing. The first (Thomson) locks for five minutes after five attempts. The
second (dlink) reboots after a few minutes of testing. Try one until it locks,
swith to the other until it reboots, switch back to the first, etc ad nauseam.
It would be great if reaver could take arguments like -b
thefirst,thesecond,thethird -c 1,6,11 and do this switching by itself. Switch
to the next if it detects a lock or progress stops.
```
Original issue reported on code.google.com by `stur...@gmail.com` on 27 Feb 2013 at 9:05 | 1.0 | Try multiple BSSIDs, round robin - ```
I found myself sitting an entire evening switching between two routers I was
testing. The first (Thomson) locks for five minutes after five attempts. The
second (dlink) reboots after a few minutes of testing. Try one until it locks,
swith to the other until it reboots, switch back to the first, etc ad nauseam.
It would be great if reaver could take arguments like -b
thefirst,thesecond,thethird -c 1,6,11 and do this switching by itself. Switch
to the next if it detects a lock or progress stops.
```
Original issue reported on code.google.com by `stur...@gmail.com` on 27 Feb 2013 at 9:05 | defect | try multiple bssids round robin i found myself sitting an entire evening switching between two routers i was testing the first thomson locks for five minutes after five attempts the second dlink reboots after a few minutes of testing try one until it locks swith to the other until it reboots switch back to the first etc ad nauseam it would be great if reaver could take arguments like b thefirst thesecond thethird c and do this switching by itself switch to the next if it detects a lock or progress stops original issue reported on code google com by stur gmail com on feb at | 1 |
310,441 | 26,716,616,535 | IssuesEvent | 2023-01-28 15:54:07 | rust-lang/rust | https://api.github.com/repos/rust-lang/rust | closed | Error when trying use trait alias for Fn() in impl | A-traits E-needs-test T-compiler C-bug F-trait_alias | This code produce compiling error:
```rust
#![feature(trait_alias)]
struct MyStruct {}
trait MyFn = Fn(&MyStruct);
fn foo (_: impl MyFn) {}
fn main () {
foo(|_| {});
}
```
([Playground](https://play.rust-lang.org/?version=nightly&mode=debug&edition=2018&gist=6f1bfe3f9fca8240a2111c94b12a963d))
Errors:
```
Compiling playground v0.0.1 (/playground)
error[E0631]: type mismatch in closure arguments
--> src/main.rs:9:5
|
9 | foo(|_| {});
| ^^^ --- found signature of `fn(_) -> _`
| |
| expected signature of `for<'r> fn(&'r MyStruct) -> _`
|
note: required by `foo`
--> src/main.rs:6:1
|
6 | fn foo (_: impl MyFn) {}
| ^^^^^^^^^^^^^^^^^^^^^
error[E0271]: type mismatch resolving `for<'r> <[closure@src/main.rs:9:9: 9:15] as std::ops::FnOnce<(&'r MyStruct,)>>::Output == ()`
--> src/main.rs:9:5
|
9 | foo(|_| {});
| ^^^ expected bound lifetime parameter, found concrete lifetime
|
note: required by `foo`
--> src/main.rs:6:1
|
6 | fn foo (_: impl MyFn) {}
| ^^^^^^^^^^^^^^^^^^^^^
error: aborting due to 2 previous errors
For more information about this error, try `rustc --explain E0271`.
error: Could not compile `playground`.
To learn more, run the command again with --verbose.
```
But when non aliased trait used, then no error occured. That is when we replace
`fn foo (_: impl MyFn) {}`
with
`fn foo (_: impl Fn(&MyStruct)) {}` | 1.0 | Error when trying use trait alias for Fn() in impl - This code produce compiling error:
```rust
#![feature(trait_alias)]
struct MyStruct {}
trait MyFn = Fn(&MyStruct);
fn foo (_: impl MyFn) {}
fn main () {
foo(|_| {});
}
```
([Playground](https://play.rust-lang.org/?version=nightly&mode=debug&edition=2018&gist=6f1bfe3f9fca8240a2111c94b12a963d))
Errors:
```
Compiling playground v0.0.1 (/playground)
error[E0631]: type mismatch in closure arguments
--> src/main.rs:9:5
|
9 | foo(|_| {});
| ^^^ --- found signature of `fn(_) -> _`
| |
| expected signature of `for<'r> fn(&'r MyStruct) -> _`
|
note: required by `foo`
--> src/main.rs:6:1
|
6 | fn foo (_: impl MyFn) {}
| ^^^^^^^^^^^^^^^^^^^^^
error[E0271]: type mismatch resolving `for<'r> <[closure@src/main.rs:9:9: 9:15] as std::ops::FnOnce<(&'r MyStruct,)>>::Output == ()`
--> src/main.rs:9:5
|
9 | foo(|_| {});
| ^^^ expected bound lifetime parameter, found concrete lifetime
|
note: required by `foo`
--> src/main.rs:6:1
|
6 | fn foo (_: impl MyFn) {}
| ^^^^^^^^^^^^^^^^^^^^^
error: aborting due to 2 previous errors
For more information about this error, try `rustc --explain E0271`.
error: Could not compile `playground`.
To learn more, run the command again with --verbose.
```
But when non aliased trait used, then no error occured. That is when we replace
`fn foo (_: impl MyFn) {}`
with
`fn foo (_: impl Fn(&MyStruct)) {}` | non_defect | error when trying use trait alias for fn in impl this code produce compiling error rust struct mystruct trait myfn fn mystruct fn foo impl myfn fn main foo errors compiling playground playground error type mismatch in closure arguments src main rs foo found signature of fn expected signature of for fn r mystruct note required by foo src main rs fn foo impl myfn error type mismatch resolving for output src main rs foo expected bound lifetime parameter found concrete lifetime note required by foo src main rs fn foo impl myfn error aborting due to previous errors for more information about this error try rustc explain error could not compile playground to learn more run the command again with verbose but when non aliased trait used then no error occured that is when we replace fn foo impl myfn with fn foo impl fn mystruct | 0 |
49,001 | 13,185,189,493 | IssuesEvent | 2020-08-12 20:54:06 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | Increase robustness of ROOTSYS detection foo (Trac #587) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/587
, reported by blaufuss and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2013-06-28T13:01:06",
"description": "still often get strangish errors from ROOT voodoo (rootcint sometimes)\n\nOften attached to having non-ports ROOTSYS in path or LD_LIB path or\nenv.\n\nSee, for example, email from Seon-hee:\n\nHi all again,\n\nWhen I build before, I was setting \nROOTSYS=/data1/i3_ports-v3/root-v5.20.00 .\nWhen I do not set this as Henrik suggested,\nthe compilation goes fine.\n\nThanks!\n(and sorry for the spam,)\nSeon-Hee\n\nSeon-Hee Seo wrote:\nHi all again,\n\nI just needed to include dependent projects, simclasses and trigger-sim.\nNow it works for cmake part.\n(Thanks Henrik and RobertF.)\n\nBut my compile fails with the following error.\n\n-------------------------------------\n[ 56%] Generating \n/home/seo/icerec/V3-1-0/build/simclasses/CMakeFiles/simclassesDict.cxx \nwith rootcint\n/tmp/gab3A1_cint.cxx:1:13: error: C: No such file or directory\nError: external preprocessing failed. /usr/include/c++/4.1.3/exception:42:\nError: Too many '}' /usr/include/c++/4.1.3/exception:103:\n/tmp/FsefFB_cint.cxx:1:13: error: C: No such file or directory\nError: external preprocessing failed. /usr/include/c++/4.1.3/typeinfo:42:\nError: Symbol __constchar is not defined in current scope \n/usr/include/gconv.h:118:\nError: Symbol __modname is not defined in current scope \n/usr/include/gconv.h:118:\nError: class,struct,union or type __builtin_va_list not defined \n/usr/lib/gcc/x86_64-linux-gnu/4.1.3/include/stdarg.h:43:\nError: Too many '}' /usr/include/stdio.h:841:\nError: Too many '}' /usr/include/nl_types.h:53:\nSyntax error /usr/include/c++/4.1.3/x86_64-linux-gnu/bits/c++locale.h:55:\nError: unrecognized language construct \n/usr/include/c++/4.1.3/x86_64-linux-gnu/bits/c++locale.h:55:\nError: Symbol __uselocale is not defined in current scope \n/usr/include/c++/4.1.3/x86_64-linux-gnu/bits/c++locale.h:55:\nError: Symbol __constchar is not defined in current scope \n/usr/include/time.h:148:\nError: Symbol tm_zone is not defined in current scope \n/usr/include/time.h:148:\nError: Too many '}' /usr/include/pthread.h:1104:\nError: Too many '}' /usr/include/unistd.h:1069:\nSyntax Error: void* \n/usr/include/c++/4.1.3/x86_64-linux-gnu/bits/gthr-default.h:113:\nError: Too many '}' /usr/include/ctype.h:218:\nError: Too many '}' /usr/include/c++/4.1.3/bits/functexcept.h:91:\nError: class,struct,union or type ... not defined \n/usr/include/c++/4.1.3/bits/cpp_type_traits.h:83:\nInternal error: global function template arg type \n/usr/include/c++/4.1.3/bits/cpp_type_traits.h:83:\nError: Missing closing brace for the block opened around line 44.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 30.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 28.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 217.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 31.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 30.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 42.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nWarning: Error occurred during reading source files\nWarning: Error occurred during dictionary source generation\n!!!Removing \n/home/seo/icerec/V3-1-0/build/simclasses/CMakeFiles/simclassesDict.cxx \n/home/seo/icerec/V3-1-0/build/simclasses/CMakeFiles/simclassesDict.h !!!\nError: rootcint: error loading headers...\nmake[2]: *** [simclasses/CMakeFiles/simclassesDict.cxx] Error 1\nmake[1]: *** [simclasses/CMakeFiles/simclasses.dir/all] Error 2\nmake: *** [all] Error 2\n---------------------------------\n\nCould this be ROOT version issue?\nHow to deal with this?\n\nCheers,\nSeon-Hee\n\nSeon-Hee Seo wrote:\n\nHi all,\n\nI was able to build V03-01-00 icerec.\nBut when I include sim-services (most recent release V00-07-01),\nI get cmake error:\n\n--------------------------------------\n\n....\n-- sim-services\nCMake Error at cmake/project.cmake:82 (message):\n Attempt to use nonexistent project 'simclasses'\nCall Stack (most recent call first):\n cmake/project.cmake:208 (use_projects)\n sim-services/CMakeLists.txt:6 (i3_add_library)\n\n\n-- Configuring incomplete, errors occurred!\nmake: *** [rebuild_cache] Error 1\n----------------------------------------\n\nIn V2 icerec, I did not have this problem.\n\nI need sim-services project to change RunID, EventID\nto be able to use data which contains same RunID, EventID pair.\n(So that my analysis-tree handles the data properly.)\n\nThis data set is corsika set 2267 being used for filter study.\n\nAny suggestion?\n\nThanks,\nSeon-Hee\n_______________________________________________",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1372424466000000",
"component": "cmake",
"summary": "Increase robustness of ROOTSYS detection foo",
"priority": "normal",
"keywords": "",
"time": "2010-01-19T20:53:52",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Increase robustness of ROOTSYS detection foo (Trac #587) - <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/587
, reported by blaufuss and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2013-06-28T13:01:06",
"description": "still often get strangish errors from ROOT voodoo (rootcint sometimes)\n\nOften attached to having non-ports ROOTSYS in path or LD_LIB path or\nenv.\n\nSee, for example, email from Seon-hee:\n\nHi all again,\n\nWhen I build before, I was setting \nROOTSYS=/data1/i3_ports-v3/root-v5.20.00 .\nWhen I do not set this as Henrik suggested,\nthe compilation goes fine.\n\nThanks!\n(and sorry for the spam,)\nSeon-Hee\n\nSeon-Hee Seo wrote:\nHi all again,\n\nI just needed to include dependent projects, simclasses and trigger-sim.\nNow it works for cmake part.\n(Thanks Henrik and RobertF.)\n\nBut my compile fails with the following error.\n\n-------------------------------------\n[ 56%] Generating \n/home/seo/icerec/V3-1-0/build/simclasses/CMakeFiles/simclassesDict.cxx \nwith rootcint\n/tmp/gab3A1_cint.cxx:1:13: error: C: No such file or directory\nError: external preprocessing failed. /usr/include/c++/4.1.3/exception:42:\nError: Too many '}' /usr/include/c++/4.1.3/exception:103:\n/tmp/FsefFB_cint.cxx:1:13: error: C: No such file or directory\nError: external preprocessing failed. /usr/include/c++/4.1.3/typeinfo:42:\nError: Symbol __constchar is not defined in current scope \n/usr/include/gconv.h:118:\nError: Symbol __modname is not defined in current scope \n/usr/include/gconv.h:118:\nError: class,struct,union or type __builtin_va_list not defined \n/usr/lib/gcc/x86_64-linux-gnu/4.1.3/include/stdarg.h:43:\nError: Too many '}' /usr/include/stdio.h:841:\nError: Too many '}' /usr/include/nl_types.h:53:\nSyntax error /usr/include/c++/4.1.3/x86_64-linux-gnu/bits/c++locale.h:55:\nError: unrecognized language construct \n/usr/include/c++/4.1.3/x86_64-linux-gnu/bits/c++locale.h:55:\nError: Symbol __uselocale is not defined in current scope \n/usr/include/c++/4.1.3/x86_64-linux-gnu/bits/c++locale.h:55:\nError: Symbol __constchar is not defined in current scope \n/usr/include/time.h:148:\nError: Symbol tm_zone is not defined in current scope \n/usr/include/time.h:148:\nError: Too many '}' /usr/include/pthread.h:1104:\nError: Too many '}' /usr/include/unistd.h:1069:\nSyntax Error: void* \n/usr/include/c++/4.1.3/x86_64-linux-gnu/bits/gthr-default.h:113:\nError: Too many '}' /usr/include/ctype.h:218:\nError: Too many '}' /usr/include/c++/4.1.3/bits/functexcept.h:91:\nError: class,struct,union or type ... not defined \n/usr/include/c++/4.1.3/bits/cpp_type_traits.h:83:\nInternal error: global function template arg type \n/usr/include/c++/4.1.3/bits/cpp_type_traits.h:83:\nError: Missing closing brace for the block opened around line 44.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 30.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 28.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 217.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 31.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 30.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nError: Missing closing brace for the block opened around line 42.\nError: Unexpected end of file (G__exec_statement()) /tmp/Bd3P2a_cint.cxx:2:\nWarning: Error occurred during reading source files\nWarning: Error occurred during dictionary source generation\n!!!Removing \n/home/seo/icerec/V3-1-0/build/simclasses/CMakeFiles/simclassesDict.cxx \n/home/seo/icerec/V3-1-0/build/simclasses/CMakeFiles/simclassesDict.h !!!\nError: rootcint: error loading headers...\nmake[2]: *** [simclasses/CMakeFiles/simclassesDict.cxx] Error 1\nmake[1]: *** [simclasses/CMakeFiles/simclasses.dir/all] Error 2\nmake: *** [all] Error 2\n---------------------------------\n\nCould this be ROOT version issue?\nHow to deal with this?\n\nCheers,\nSeon-Hee\n\nSeon-Hee Seo wrote:\n\nHi all,\n\nI was able to build V03-01-00 icerec.\nBut when I include sim-services (most recent release V00-07-01),\nI get cmake error:\n\n--------------------------------------\n\n....\n-- sim-services\nCMake Error at cmake/project.cmake:82 (message):\n Attempt to use nonexistent project 'simclasses'\nCall Stack (most recent call first):\n cmake/project.cmake:208 (use_projects)\n sim-services/CMakeLists.txt:6 (i3_add_library)\n\n\n-- Configuring incomplete, errors occurred!\nmake: *** [rebuild_cache] Error 1\n----------------------------------------\n\nIn V2 icerec, I did not have this problem.\n\nI need sim-services project to change RunID, EventID\nto be able to use data which contains same RunID, EventID pair.\n(So that my analysis-tree handles the data properly.)\n\nThis data set is corsika set 2267 being used for filter study.\n\nAny suggestion?\n\nThanks,\nSeon-Hee\n_______________________________________________",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1372424466000000",
"component": "cmake",
"summary": "Increase robustness of ROOTSYS detection foo",
"priority": "normal",
"keywords": "",
"time": "2010-01-19T20:53:52",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | increase robustness of rootsys detection foo trac migrated from reported by blaufuss and owned by nega json status closed changetime description still often get strangish errors from root voodoo rootcint sometimes n noften attached to having non ports rootsys in path or ld lib path or nenv n nsee for example email from seon hee n nhi all again n nwhen i build before i was setting nrootsys ports root nwhen i do not set this as henrik suggested nthe compilation goes fine n nthanks n and sorry for the spam nseon hee n nseon hee seo wrote nhi all again n ni just needed to include dependent projects simclasses and trigger sim nnow it works for cmake part n thanks henrik and robertf n nbut my compile fails with the following error n n n generating n home seo icerec build simclasses cmakefiles simclassesdict cxx nwith rootcint n tmp cint cxx error c no such file or directory nerror external preprocessing failed usr include c exception nerror too many usr include c exception n tmp fseffb cint cxx error c no such file or directory nerror external preprocessing failed usr include c typeinfo nerror symbol constchar is not defined in current scope n usr include gconv h nerror symbol modname is not defined in current scope n usr include gconv h nerror class struct union or type builtin va list not defined n usr lib gcc linux gnu include stdarg h nerror too many usr include stdio h nerror too many usr include nl types h nsyntax error usr include c linux gnu bits c locale h nerror unrecognized language construct n usr include c linux gnu bits c locale h nerror symbol uselocale is not defined in current scope n usr include c linux gnu bits c locale h nerror symbol constchar is not defined in current scope n usr include time h nerror symbol tm zone is not defined in current scope n usr include time h nerror too many usr include pthread h nerror too many usr include unistd h nsyntax error void n usr include c linux gnu bits gthr default h nerror too many usr include ctype h nerror too many usr include c bits functexcept h nerror class struct union or type not defined n usr include c bits cpp type traits h ninternal error global function template arg type n usr include c bits cpp type traits h nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nerror missing closing brace for the block opened around line nerror unexpected end of file g exec statement tmp cint cxx nwarning error occurred during reading source files nwarning error occurred during dictionary source generation n removing n home seo icerec build simclasses cmakefiles simclassesdict cxx n home seo icerec build simclasses cmakefiles simclassesdict h nerror rootcint error loading headers nmake error nmake error nmake error n n ncould this be root version issue nhow to deal with this n ncheers nseon hee n nseon hee seo wrote n nhi all n ni was able to build icerec nbut when i include sim services most recent release ni get cmake error n n n n n sim services ncmake error at cmake project cmake message n attempt to use nonexistent project simclasses ncall stack most recent call first n cmake project cmake use projects n sim services cmakelists txt add library n n n configuring incomplete errors occurred nmake error n n nin icerec i did not have this problem n ni need sim services project to change runid eventid nto be able to use data which contains same runid eventid pair n so that my analysis tree handles the data properly n nthis data set is corsika set being used for filter study n nany suggestion n nthanks nseon hee n reporter blaufuss cc resolution fixed ts component cmake summary increase robustness of rootsys detection foo priority normal keywords time milestone owner nega type defect | 1 |
15,477 | 2,856,459,610 | IssuesEvent | 2015-06-02 15:03:53 | svalinn/DAGMC | https://api.github.com/repos/svalinn/DAGMC | closed | Change amalgamated pyne library name | Type: Defect Type: Enhancement | The compiled shared object library name clashes with standard pyne library name, it should be changed to something more appropriate that doesn't clash like `pyne_builtin`, something definitely not called pyne. | 1.0 | Change amalgamated pyne library name - The compiled shared object library name clashes with standard pyne library name, it should be changed to something more appropriate that doesn't clash like `pyne_builtin`, something definitely not called pyne. | defect | change amalgamated pyne library name the compiled shared object library name clashes with standard pyne library name it should be changed to something more appropriate that doesn t clash like pyne builtin something definitely not called pyne | 1 |
166,390 | 6,303,924,156 | IssuesEvent | 2017-07-21 14:48:17 | chez-nestor/backoffice | https://api.github.com/repos/chez-nestor/backoffice | closed | Write unit tests for complex Liana getters | [priority] P2 [type] enhancement | - [ ] load Liana collections ourselves from src/models ourselves,
- [ ] find a way to build objects that will have those getters | 1.0 | Write unit tests for complex Liana getters - - [ ] load Liana collections ourselves from src/models ourselves,
- [ ] find a way to build objects that will have those getters | non_defect | write unit tests for complex liana getters load liana collections ourselves from src models ourselves find a way to build objects that will have those getters | 0 |
77,773 | 27,156,303,056 | IssuesEvent | 2023-02-17 08:09:17 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | opened | DataExporter: Attrribute 'visibleOnly' not documented | :lady_beetle: defect :bangbang: needs-triage | ### Describe the bug
The [online documentation](https://primefaces.github.io/primefaces/12_0_0/#/components/dataexporter) of component DataExporter has no description for attribute 'visibleOnly'.
### Reproducer
_No response_
### Expected behavior
_No response_
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.3
### Java version
11
### Browser(s)
_No response_ | 1.0 | DataExporter: Attrribute 'visibleOnly' not documented - ### Describe the bug
The [online documentation](https://primefaces.github.io/primefaces/12_0_0/#/components/dataexporter) of component DataExporter has no description for attribute 'visibleOnly'.
### Reproducer
_No response_
### Expected behavior
_No response_
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.3
### Java version
11
### Browser(s)
_No response_ | defect | dataexporter attrribute visibleonly not documented describe the bug the of component dataexporter has no description for attribute visibleonly reproducer no response expected behavior no response primefaces edition community primefaces version theme no response jsf implementation mojarra jsf version java version browser s no response | 1 |
9,578 | 2,615,162,942 | IssuesEvent | 2015-03-01 06:42:23 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | opened | bssid.wpc empty | auto-migrated Priority-Triage Type-Defect | ```
System crashed and i need to reboot. Now the bssid.wpc is empty! The file
exist, sure reaver ask to restore the session but starts from the beginning.
It was about 18% done and Pin 20XXXXXX, it takes me 1 week to get that far. Do
i really have to start from the beginning?
Or maybe i can write something in the bssid.wpc file to start from pin
20000000? What else do i have to write in this file so that the restore option
from reaver will work and doesn´t start from the very beginning 0,0%.
Sry my english isn´t the best. regards
```
Original issue reported on code.google.com by `rkutsch...@gmail.com` on 14 Jun 2012 at 11:11 | 1.0 | bssid.wpc empty - ```
System crashed and i need to reboot. Now the bssid.wpc is empty! The file
exist, sure reaver ask to restore the session but starts from the beginning.
It was about 18% done and Pin 20XXXXXX, it takes me 1 week to get that far. Do
i really have to start from the beginning?
Or maybe i can write something in the bssid.wpc file to start from pin
20000000? What else do i have to write in this file so that the restore option
from reaver will work and doesn´t start from the very beginning 0,0%.
Sry my english isn´t the best. regards
```
Original issue reported on code.google.com by `rkutsch...@gmail.com` on 14 Jun 2012 at 11:11 | defect | bssid wpc empty system crashed and i need to reboot now the bssid wpc is empty the file exist sure reaver ask to restore the session but starts from the beginning it was about done and pin it takes me week to get that far do i really have to start from the beginning or maybe i can write something in the bssid wpc file to start from pin what else do i have to write in this file so that the restore option from reaver will work and doesn´t start from the very beginning sry my english isn´t the best regards original issue reported on code google com by rkutsch gmail com on jun at | 1 |
36,805 | 8,139,791,383 | IssuesEvent | 2018-08-20 18:52:40 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | Null-propagation does not work on nullable value tuples. | defect in-progress premium | ### Steps To Reproduce
https://deck.net/6574bfcdd0addabfc66d1cc0c3c7498f
Related to https://github.com/bridgedotnet/Bridge/issues/3645
```csharp
public class App
{
public static void Main()
{
(string Prop1, string Prop2)? val = ("test1", "test2");
Console.WriteLine(val.Value.Prop1);
Console.WriteLine(val?.Prop1);
}
}
```
### Expected Result
```
> test1
> test1
```
### Actual Result
```
> test1
>
```
| 1.0 | Null-propagation does not work on nullable value tuples. - ### Steps To Reproduce
https://deck.net/6574bfcdd0addabfc66d1cc0c3c7498f
Related to https://github.com/bridgedotnet/Bridge/issues/3645
```csharp
public class App
{
public static void Main()
{
(string Prop1, string Prop2)? val = ("test1", "test2");
Console.WriteLine(val.Value.Prop1);
Console.WriteLine(val?.Prop1);
}
}
```
### Expected Result
```
> test1
> test1
```
### Actual Result
```
> test1
>
```
| defect | null propagation does not work on nullable value tuples steps to reproduce related to csharp public class app public static void main string string val console writeline val value console writeline val expected result actual result | 1 |
3,900 | 2,610,083,707 | IssuesEvent | 2015-02-26 18:25:33 | chrsmith/dsdsdaadf | https://api.github.com/repos/chrsmith/dsdsdaadf | opened | 深圳宝安哪里祛痘好 | auto-migrated Priority-Medium Type-Defect | ```
深圳宝安哪里祛痘好【深圳韩方科颜全国热线400-869-1818,24小
时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国��
�方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩�
��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”
健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专��
�治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的�
��痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:51 | 1.0 | 深圳宝安哪里祛痘好 - ```
深圳宝安哪里祛痘好【深圳韩方科颜全国热线400-869-1818,24小
时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国��
�方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩�
��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”
健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专��
�治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的�
��痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:51 | defect | 深圳宝安哪里祛痘好 深圳宝安哪里祛痘好【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国�� �方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩� ��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹” 健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专�� �治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的� ��痘。 original issue reported on code google com by szft com on may at | 1 |
33,759 | 7,237,732,059 | IssuesEvent | 2018-02-13 12:09:09 | oozcitak/exiflibrary | https://api.github.com/repos/oozcitak/exiflibrary | closed | Question on recompression | Priority-Medium Type-Defect auto-migrated | ```
When i use the library to modify EXIF data ("comment" for example) and save,
does it recompress the picture?
Thanks
Filip
```
Original issue reported on code.google.com by `Filip.Je...@gmail.com` on 27 Aug 2013 at 4:06
| 1.0 | Question on recompression - ```
When i use the library to modify EXIF data ("comment" for example) and save,
does it recompress the picture?
Thanks
Filip
```
Original issue reported on code.google.com by `Filip.Je...@gmail.com` on 27 Aug 2013 at 4:06
| defect | question on recompression when i use the library to modify exif data comment for example and save does it recompress the picture thanks filip original issue reported on code google com by filip je gmail com on aug at | 1 |
40,778 | 10,154,006,446 | IssuesEvent | 2019-08-06 06:50:43 | line/centraldogma | https://api.github.com/repos/line/centraldogma | opened | Some server log messages are not logged to Logback | defect | I found this from the stderr output:
```
Aug 06, 2019 3:18:22 PM com.github.benmanes.caffeine.cache.LocalAsyncCache lambda$handleCompletion$3
WARNING: Exception thrown during asynchronous load
java.util.concurrent.CompletionException: com.linecorp.centraldogma.server.internal.storage.RequestAlreadyTimedOutException: Request already timed out.
at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
at java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1702)
at io.micrometer.core.instrument.AbstractTimer.record(AbstractTimer.java:160)
at io.micrometer.core.instrument.Timer.lambda$wrap$0(Timer.java:144)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:834)
Caused by: com.linecorp.centraldogma.server.internal.storage.RequestAlreadyTimedOutException: Request already timed out.
``` | 1.0 | Some server log messages are not logged to Logback - I found this from the stderr output:
```
Aug 06, 2019 3:18:22 PM com.github.benmanes.caffeine.cache.LocalAsyncCache lambda$handleCompletion$3
WARNING: Exception thrown during asynchronous load
java.util.concurrent.CompletionException: com.linecorp.centraldogma.server.internal.storage.RequestAlreadyTimedOutException: Request already timed out.
at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
at java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1702)
at io.micrometer.core.instrument.AbstractTimer.record(AbstractTimer.java:160)
at io.micrometer.core.instrument.Timer.lambda$wrap$0(Timer.java:144)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:834)
Caused by: com.linecorp.centraldogma.server.internal.storage.RequestAlreadyTimedOutException: Request already timed out.
``` | defect | some server log messages are not logged to logback i found this from the stderr output aug pm com github benmanes caffeine cache localasynccache lambda handlecompletion warning exception thrown during asynchronous load java util concurrent completionexception com linecorp centraldogma server internal storage requestalreadytimedoutexception request already timed out at java base java util concurrent completablefuture encodethrowable completablefuture java at java base java util concurrent completablefuture completethrowable completablefuture java at java base java util concurrent completablefuture asyncsupply run completablefuture java at io micrometer core instrument abstracttimer record abstracttimer java at io micrometer core instrument timer lambda wrap timer java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java caused by com linecorp centraldogma server internal storage requestalreadytimedoutexception request already timed out | 1 |
265,087 | 8,337,106,047 | IssuesEvent | 2018-09-28 09:57:44 | opencaching/opencaching-pl | https://api.github.com/repos/opencaching/opencaching-pl | closed | vars in translations | Component_i18n Priority_Low Type_Enhancement | There is need to handle vars in translation texts.
"Old style" template system provides possibility of including vars as:
`'activate_mail_subject' => 'New user registration - {site_name}',`
Ideas:
"new" way can handle it in "printf" style so the translation can look like:
`'activate_mail_subject' => 'New user registration - %s',`
and the code
`tr('activate_mail_subject', $view->siteName)`
| 1.0 | vars in translations - There is need to handle vars in translation texts.
"Old style" template system provides possibility of including vars as:
`'activate_mail_subject' => 'New user registration - {site_name}',`
Ideas:
"new" way can handle it in "printf" style so the translation can look like:
`'activate_mail_subject' => 'New user registration - %s',`
and the code
`tr('activate_mail_subject', $view->siteName)`
| non_defect | vars in translations there is need to handle vars in translation texts old style template system provides possibility of including vars as activate mail subject new user registration site name ideas new way can handle it in printf style so the translation can look like activate mail subject new user registration s and the code tr activate mail subject view sitename | 0 |
231,981 | 7,647,517,067 | IssuesEvent | 2018-05-09 04:32:12 | AdChain/AdChainRegistryDapp | https://api.github.com/repos/AdChain/AdChainRegistryDapp | closed | Committing a Vote- Step 2 in Walkthrough | Priority: Medium Status: Review Needed | "The filtered domains are all in the Voting Commit stage."
<img width="1181" alt="screen shot 2018-04-03 at 10 46 45 am" src="https://user-images.githubusercontent.com/35276813/38266202-e60c0e20-372c-11e8-8f6f-c9de64eb4048.png">
However, the image shows domains in various other stages such as "In Registry", "Reveal-Pending",
and "Application (Pending)".
| 1.0 | Committing a Vote- Step 2 in Walkthrough - "The filtered domains are all in the Voting Commit stage."
<img width="1181" alt="screen shot 2018-04-03 at 10 46 45 am" src="https://user-images.githubusercontent.com/35276813/38266202-e60c0e20-372c-11e8-8f6f-c9de64eb4048.png">
However, the image shows domains in various other stages such as "In Registry", "Reveal-Pending",
and "Application (Pending)".
| non_defect | committing a vote step in walkthrough the filtered domains are all in the voting commit stage img width alt screen shot at am src however the image shows domains in various other stages such as in registry reveal pending and application pending | 0 |
307,095 | 9,414,186,019 | IssuesEvent | 2019-04-10 09:33:38 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Feature: button to open profile folder | Feature Medium Priority | Files from this folder regularry needs for support, playtime.eco need for first server configuration...
And finding this folder is not easy.
`%appdata%\..\LocalLow\Strange Loop Games\Eco` | 1.0 | Feature: button to open profile folder - Files from this folder regularry needs for support, playtime.eco need for first server configuration...
And finding this folder is not easy.
`%appdata%\..\LocalLow\Strange Loop Games\Eco` | non_defect | feature button to open profile folder files from this folder regularry needs for support playtime eco need for first server configuration and finding this folder is not easy appdata locallow strange loop games eco | 0 |
71,138 | 9,480,507,364 | IssuesEvent | 2019-04-20 18:25:50 | smithjd/sql-pet | https://api.github.com/repos/smithjd/sql-pet | closed | Investigate DiagrammeR integration with bookdown book | documentation | The tools of choice for creating diagrams in RStudio / R Markdown is the `DiagrammeR` R package (http://rich-iannone.github.io/DiagrammeR/index.html).
1. Diagrams are code - you can edit them and version-control them.
2. RStudio has provisions to view / render them as you build them.
What I don't know (yet) is how one can integrate them into a bookdown book. | 1.0 | Investigate DiagrammeR integration with bookdown book - The tools of choice for creating diagrams in RStudio / R Markdown is the `DiagrammeR` R package (http://rich-iannone.github.io/DiagrammeR/index.html).
1. Diagrams are code - you can edit them and version-control them.
2. RStudio has provisions to view / render them as you build them.
What I don't know (yet) is how one can integrate them into a bookdown book. | non_defect | investigate diagrammer integration with bookdown book the tools of choice for creating diagrams in rstudio r markdown is the diagrammer r package diagrams are code you can edit them and version control them rstudio has provisions to view render them as you build them what i don t know yet is how one can integrate them into a bookdown book | 0 |
65,436 | 19,515,508,094 | IssuesEvent | 2021-12-29 09:32:13 | ontop/ontop | https://api.github.com/repos/ontop/ontop | closed | bootstrap fails when there is a unique index that uses md5() of a column value | type: defect status: fixed w: db support | ### Description
bootstrap fails when there is a unique index that uses md5() of a column value with the following error:
```
20:42:22.607 [main] DEBUG o.s.owlapi.utilities.Injector - Injecting values [[org.semanticweb.owlapi.rdf.rdfxml.renderer.RDFXMLStorerFactory@13e698c7, org.semanticweb.owlapi.functional.renderer.FunctionalSyntaxStorerFactory@267bbe1a, org.semanticweb.owlapi.manchestersyntax.renderer.ManchesterSyntaxStorerFactory@1f12e153, org.semanticweb.owlapi.rio.RioJsonStorerFactory@6eafb10e, org.semanticweb.owlapi.rio.RioNTriplesStorerFactory@2873d672, org.semanticweb.owlapi.dlsyntax.renderer.DLSyntaxHTMLStorerFactory@6b410923, org.semanticweb.owlapi.krss2.renderer.KRSS2OWLSyntaxStorerFactory@389562d6, org.semanticweb.owlapi.rio.RioNQuadsStorerFactory@464a4442, org.semanticweb.owlapi.rio.RioJsonLDStorerFactory@1d3e6d34, org.semanticweb.owlapi.rdf.turtle.renderer.TurtleStorerFactory@5a101b1c, org.semanticweb.owlapi.rio.RioN3StorerFactory@26a94fa5, org.semanticweb.owlapi.rio.RioTurtleStorerFactory@757529a4, org.semanticweb.owlapi.rio.RioTrixStorerFactory@4303b7f0, org.semanticweb.owlapi.rio.RioRDFXMLStorerFactory@3bc735b3, org.semanticweb.owlapi.latex.renderer.LatexStorerFactory@3a60c416, org.semanticweb.owlapi.rio.RioTrigStorerFactory@577f9109, org.semanticweb.owlapi.oboformat.OBOFormatStorerFactory@6d64b553, org.semanticweb.owlapi.rio.RioBinaryRdfStorerFactory@53667cbe, org.semanticweb.owlapi.owlxml.renderer.OWLXMLStorerFactory@aed0151, org.semanticweb.owlapi.dlsyntax.renderer.DLSyntaxStorerFactory@e8fadb0]] on method public void uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.setOntologyStorers(java.util.Set).
Error occurred during bootstrapping: it.unibz.inf.ontop.exception.MetadataExtractionException: it.unibz.inf.ontop.dbschema.AttributeNotFoundException
Debugging information for developers:
it.unibz.inf.ontop.exception.MappingBootstrappingException: it.unibz.inf.ontop.exception.MetadataExtractionException: it.unibz.inf.ontop.dbschema.AttributeNotFoundException
at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.bootstrapMappingAndOntology(DirectMappingEngine.java:148)
at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.bootstrap(DirectMappingEngine.java:91)
at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DefaultDirectMappingBootstrapper.bootstrap(DefaultDirectMappingBootstrapper.java:16)
at it.unibz.inf.ontop.cli.OntopBootstrap.run(OntopBootstrap.java:56)
at it.unibz.inf.ontop.cli.Ontop.main(Ontop.java:20)
Caused by: it.unibz.inf.ontop.exception.MetadataExtractionException: it.unibz.inf.ontop.dbschema.AttributeNotFoundException
at it.unibz.inf.ontop.dbschema.impl.AbstractDBMetadataProvider.insertUniqueAttributes(AbstractDBMetadataProvider.java:243)
at it.unibz.inf.ontop.dbschema.impl.AbstractDBMetadataProvider.insertIntegrityConstraints(AbstractDBMetadataProvider.java:154)
at it.unibz.inf.ontop.dbschema.impl.CachingMetadataLookup.extractImmutableMetadata(CachingMetadataLookup.java:44)
at it.unibz.inf.ontop.dbschema.ImmutableMetadata.extractImmutableMetadata(ImmutableMetadata.java:17)
at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.extractPPMapping(DirectMappingEngine.java:166)
at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.bootstrapMappingAndOntology(DirectMappingEngine.java:123)
... 4 more
Caused by: it.unibz.inf.ontop.dbschema.AttributeNotFoundException
at it.unibz.inf.ontop.dbschema.impl.AbstractRelationDefinition.getAttribute(AbstractRelationDefinition.java:49)
at it.unibz.inf.ontop.dbschema.impl.UniqueConstraintImpl$UniqueConstraintBuilder.addDeterminant(UniqueConstraintImpl.java:31)
at it.unibz.inf.ontop.dbschema.impl.AbstractDBMetadataProvider.insertUniqueAttributes(AbstractDBMetadataProvider.java:234)
... 9 more
```
### Steps to Reproduce
load this into a postgres database and then run the ontop bootstrap command against it:
```
--
-- PostgreSQL database dump
--
-- Dumped from database version 11.12 (Debian 11.12-1.pgdg90+1)
-- Dumped by pg_dump version 11.12 (Debian 11.12-1.pgdg90+1)
SET statement_timeout = 0;
SET lock_timeout = 0;
SET idle_in_transaction_session_timeout = 0;
SET client_encoding = 'UTF8';
SET standard_conforming_strings = on;
SELECT pg_catalog.set_config('search_path', '', false);
SET check_function_bodies = false;
SET xmloption = content;
SET client_min_messages = warning;
SET row_security = off;
SET default_tablespace = '';
SET default_with_oids = false;
--
-- Name: cat; Type: TABLE; Schema: public; Owner: postgres
--
CREATE TABLE public.cat (
id integer NOT NULL,
name character varying,
age integer
);
ALTER TABLE public.cat OWNER TO postgres;
--
-- Data for Name: cat; Type: TABLE DATA; Schema: public; Owner: postgres
--
COPY public.cat (id, name, age) FROM stdin;
5 freddo 12
\.
--
-- Name: indycar; Type: INDEX; Schema: public; Owner: postgres
--
CREATE UNIQUE INDEX indycar ON public.cat USING btree (age, md5((name)::text));
--
-- PostgreSQL database dump complete
--
```
**Expected behavior:** [What you expect to see]
i expect to get a mapping file from the bootstrap process
**Actual behavior:** [What actually happens]
an error is emitted. but if you remove this
```
CREATE UNIQUE INDEX indycar ON public.cat USING btree (age, md5((name)::text));
```
then the bootstrap is able to run successfully.
**Reproduces how often:** [Does it happen every time?]
always
### Versions
the HEAD of the version4 branch. and i tried some older versions of ontop and had the same result.
| 1.0 | bootstrap fails when there is a unique index that uses md5() of a column value - ### Description
bootstrap fails when there is a unique index that uses md5() of a column value with the following error:
```
20:42:22.607 [main] DEBUG o.s.owlapi.utilities.Injector - Injecting values [[org.semanticweb.owlapi.rdf.rdfxml.renderer.RDFXMLStorerFactory@13e698c7, org.semanticweb.owlapi.functional.renderer.FunctionalSyntaxStorerFactory@267bbe1a, org.semanticweb.owlapi.manchestersyntax.renderer.ManchesterSyntaxStorerFactory@1f12e153, org.semanticweb.owlapi.rio.RioJsonStorerFactory@6eafb10e, org.semanticweb.owlapi.rio.RioNTriplesStorerFactory@2873d672, org.semanticweb.owlapi.dlsyntax.renderer.DLSyntaxHTMLStorerFactory@6b410923, org.semanticweb.owlapi.krss2.renderer.KRSS2OWLSyntaxStorerFactory@389562d6, org.semanticweb.owlapi.rio.RioNQuadsStorerFactory@464a4442, org.semanticweb.owlapi.rio.RioJsonLDStorerFactory@1d3e6d34, org.semanticweb.owlapi.rdf.turtle.renderer.TurtleStorerFactory@5a101b1c, org.semanticweb.owlapi.rio.RioN3StorerFactory@26a94fa5, org.semanticweb.owlapi.rio.RioTurtleStorerFactory@757529a4, org.semanticweb.owlapi.rio.RioTrixStorerFactory@4303b7f0, org.semanticweb.owlapi.rio.RioRDFXMLStorerFactory@3bc735b3, org.semanticweb.owlapi.latex.renderer.LatexStorerFactory@3a60c416, org.semanticweb.owlapi.rio.RioTrigStorerFactory@577f9109, org.semanticweb.owlapi.oboformat.OBOFormatStorerFactory@6d64b553, org.semanticweb.owlapi.rio.RioBinaryRdfStorerFactory@53667cbe, org.semanticweb.owlapi.owlxml.renderer.OWLXMLStorerFactory@aed0151, org.semanticweb.owlapi.dlsyntax.renderer.DLSyntaxStorerFactory@e8fadb0]] on method public void uk.ac.manchester.cs.owl.owlapi.OWLOntologyManagerImpl.setOntologyStorers(java.util.Set).
Error occurred during bootstrapping: it.unibz.inf.ontop.exception.MetadataExtractionException: it.unibz.inf.ontop.dbschema.AttributeNotFoundException
Debugging information for developers:
it.unibz.inf.ontop.exception.MappingBootstrappingException: it.unibz.inf.ontop.exception.MetadataExtractionException: it.unibz.inf.ontop.dbschema.AttributeNotFoundException
at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.bootstrapMappingAndOntology(DirectMappingEngine.java:148)
at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.bootstrap(DirectMappingEngine.java:91)
at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DefaultDirectMappingBootstrapper.bootstrap(DefaultDirectMappingBootstrapper.java:16)
at it.unibz.inf.ontop.cli.OntopBootstrap.run(OntopBootstrap.java:56)
at it.unibz.inf.ontop.cli.Ontop.main(Ontop.java:20)
Caused by: it.unibz.inf.ontop.exception.MetadataExtractionException: it.unibz.inf.ontop.dbschema.AttributeNotFoundException
at it.unibz.inf.ontop.dbschema.impl.AbstractDBMetadataProvider.insertUniqueAttributes(AbstractDBMetadataProvider.java:243)
at it.unibz.inf.ontop.dbschema.impl.AbstractDBMetadataProvider.insertIntegrityConstraints(AbstractDBMetadataProvider.java:154)
at it.unibz.inf.ontop.dbschema.impl.CachingMetadataLookup.extractImmutableMetadata(CachingMetadataLookup.java:44)
at it.unibz.inf.ontop.dbschema.ImmutableMetadata.extractImmutableMetadata(ImmutableMetadata.java:17)
at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.extractPPMapping(DirectMappingEngine.java:166)
at it.unibz.inf.ontop.spec.mapping.bootstrap.impl.DirectMappingEngine.bootstrapMappingAndOntology(DirectMappingEngine.java:123)
... 4 more
Caused by: it.unibz.inf.ontop.dbschema.AttributeNotFoundException
at it.unibz.inf.ontop.dbschema.impl.AbstractRelationDefinition.getAttribute(AbstractRelationDefinition.java:49)
at it.unibz.inf.ontop.dbschema.impl.UniqueConstraintImpl$UniqueConstraintBuilder.addDeterminant(UniqueConstraintImpl.java:31)
at it.unibz.inf.ontop.dbschema.impl.AbstractDBMetadataProvider.insertUniqueAttributes(AbstractDBMetadataProvider.java:234)
... 9 more
```
### Steps to Reproduce
load this into a postgres database and then run the ontop bootstrap command against it:
```
--
-- PostgreSQL database dump
--
-- Dumped from database version 11.12 (Debian 11.12-1.pgdg90+1)
-- Dumped by pg_dump version 11.12 (Debian 11.12-1.pgdg90+1)
SET statement_timeout = 0;
SET lock_timeout = 0;
SET idle_in_transaction_session_timeout = 0;
SET client_encoding = 'UTF8';
SET standard_conforming_strings = on;
SELECT pg_catalog.set_config('search_path', '', false);
SET check_function_bodies = false;
SET xmloption = content;
SET client_min_messages = warning;
SET row_security = off;
SET default_tablespace = '';
SET default_with_oids = false;
--
-- Name: cat; Type: TABLE; Schema: public; Owner: postgres
--
CREATE TABLE public.cat (
id integer NOT NULL,
name character varying,
age integer
);
ALTER TABLE public.cat OWNER TO postgres;
--
-- Data for Name: cat; Type: TABLE DATA; Schema: public; Owner: postgres
--
COPY public.cat (id, name, age) FROM stdin;
5 freddo 12
\.
--
-- Name: indycar; Type: INDEX; Schema: public; Owner: postgres
--
CREATE UNIQUE INDEX indycar ON public.cat USING btree (age, md5((name)::text));
--
-- PostgreSQL database dump complete
--
```
**Expected behavior:** [What you expect to see]
i expect to get a mapping file from the bootstrap process
**Actual behavior:** [What actually happens]
an error is emitted. but if you remove this
```
CREATE UNIQUE INDEX indycar ON public.cat USING btree (age, md5((name)::text));
```
then the bootstrap is able to run successfully.
**Reproduces how often:** [Does it happen every time?]
always
### Versions
the HEAD of the version4 branch. and i tried some older versions of ontop and had the same result.
| defect | bootstrap fails when there is a unique index that uses of a column value description bootstrap fails when there is a unique index that uses of a column value with the following error debug o s owlapi utilities injector injecting values on method public void uk ac manchester cs owl owlapi owlontologymanagerimpl setontologystorers java util set error occurred during bootstrapping it unibz inf ontop exception metadataextractionexception it unibz inf ontop dbschema attributenotfoundexception debugging information for developers it unibz inf ontop exception mappingbootstrappingexception it unibz inf ontop exception metadataextractionexception it unibz inf ontop dbschema attributenotfoundexception at it unibz inf ontop spec mapping bootstrap impl directmappingengine bootstrapmappingandontology directmappingengine java at it unibz inf ontop spec mapping bootstrap impl directmappingengine bootstrap directmappingengine java at it unibz inf ontop spec mapping bootstrap impl defaultdirectmappingbootstrapper bootstrap defaultdirectmappingbootstrapper java at it unibz inf ontop cli ontopbootstrap run ontopbootstrap java at it unibz inf ontop cli ontop main ontop java caused by it unibz inf ontop exception metadataextractionexception it unibz inf ontop dbschema attributenotfoundexception at it unibz inf ontop dbschema impl abstractdbmetadataprovider insertuniqueattributes abstractdbmetadataprovider java at it unibz inf ontop dbschema impl abstractdbmetadataprovider insertintegrityconstraints abstractdbmetadataprovider java at it unibz inf ontop dbschema impl cachingmetadatalookup extractimmutablemetadata cachingmetadatalookup java at it unibz inf ontop dbschema immutablemetadata extractimmutablemetadata immutablemetadata java at it unibz inf ontop spec mapping bootstrap impl directmappingengine extractppmapping directmappingengine java at it unibz inf ontop spec mapping bootstrap impl directmappingengine bootstrapmappingandontology directmappingengine java more caused by it unibz inf ontop dbschema attributenotfoundexception at it unibz inf ontop dbschema impl abstractrelationdefinition getattribute abstractrelationdefinition java at it unibz inf ontop dbschema impl uniqueconstraintimpl uniqueconstraintbuilder adddeterminant uniqueconstraintimpl java at it unibz inf ontop dbschema impl abstractdbmetadataprovider insertuniqueattributes abstractdbmetadataprovider java more steps to reproduce load this into a postgres database and then run the ontop bootstrap command against it postgresql database dump dumped from database version debian dumped by pg dump version debian set statement timeout set lock timeout set idle in transaction session timeout set client encoding set standard conforming strings on select pg catalog set config search path false set check function bodies false set xmloption content set client min messages warning set row security off set default tablespace set default with oids false name cat type table schema public owner postgres create table public cat id integer not null name character varying age integer alter table public cat owner to postgres data for name cat type table data schema public owner postgres copy public cat id name age from stdin freddo name indycar type index schema public owner postgres create unique index indycar on public cat using btree age name text postgresql database dump complete expected behavior i expect to get a mapping file from the bootstrap process actual behavior an error is emitted but if you remove this create unique index indycar on public cat using btree age name text then the bootstrap is able to run successfully reproduces how often always versions the head of the branch and i tried some older versions of ontop and had the same result | 1 |
48,986 | 13,185,183,426 | IssuesEvent | 2020-08-12 20:53:20 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters (Trac #565) | Incomplete Migration Migrated from Trac defect tools/ports | <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/565
, reported by kislat and owned by cgils</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2009-06-30T14:16:42",
"description": "There is a bug in rootcint (ROOT 5.20.00 and later versions) that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable (after resolving symlinks) is longer than 128 characters.\n\nA patch for the file ROOT source file core/utils/src/rootcint.cxx for root_v5.20.00 is attached.",
"reporter": "kislat",
"cc": "",
"resolution": "fixed",
"_ts": "1246371402000000",
"component": "tools/ports",
"summary": "rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters",
"priority": "normal",
"keywords": "",
"time": "2009-06-30T12:29:54",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
</p>
</details>
| 1.0 | rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters (Trac #565) - <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/565
, reported by kislat and owned by cgils</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2009-06-30T14:16:42",
"description": "There is a bug in rootcint (ROOT 5.20.00 and later versions) that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable (after resolving symlinks) is longer than 128 characters.\n\nA patch for the file ROOT source file core/utils/src/rootcint.cxx for root_v5.20.00 is attached.",
"reporter": "kislat",
"cc": "",
"resolution": "fixed",
"_ts": "1246371402000000",
"component": "tools/ports",
"summary": "rootcint won't create dictionaries if $ROOTSYS is longer than 128 characters",
"priority": "normal",
"keywords": "",
"time": "2009-06-30T12:29:54",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
</p>
</details>
| defect | rootcint won t create dictionaries if rootsys is longer than characters trac migrated from reported by kislat and owned by cgils json status closed changetime description there is a bug in rootcint root and later versions that makes it impossible for rootcint to generate dictionaries if the path to the rootcint executable after resolving symlinks is longer than characters n na patch for the file root source file core utils src rootcint cxx for root is attached reporter kislat cc resolution fixed ts component tools ports summary rootcint won t create dictionaries if rootsys is longer than characters priority normal keywords time milestone owner cgils type defect | 1 |
16,253 | 2,882,833,250 | IssuesEvent | 2015-06-11 08:25:32 | ooskapenaar/awarenet | https://api.github.com/repos/ooskapenaar/awarenet | closed | chat doesn't work | auto-migrated Priority-Low Type-Defect | ```
What steps will reproduce the problem?
1. visit "http://awarenet.eu/chat/"
2. click on "Global Chat"
3.
What is the expected output? What do you see instead?
I see this error message (see attached pic). It happend to me while using
Firefox 31.0 (in VB), IE 11 and Opera 26. Its always the same error.
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `MrLordS...@gmail.com` on 16 Jan 2015 at 8:15
Attachments:
* [error.png](https://storage.googleapis.com/google-code-attachments/awarenet/issue-395/comment-0/error.png)
| 1.0 | chat doesn't work - ```
What steps will reproduce the problem?
1. visit "http://awarenet.eu/chat/"
2. click on "Global Chat"
3.
What is the expected output? What do you see instead?
I see this error message (see attached pic). It happend to me while using
Firefox 31.0 (in VB), IE 11 and Opera 26. Its always the same error.
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `MrLordS...@gmail.com` on 16 Jan 2015 at 8:15
Attachments:
* [error.png](https://storage.googleapis.com/google-code-attachments/awarenet/issue-395/comment-0/error.png)
| defect | chat doesn t work what steps will reproduce the problem visit click on global chat what is the expected output what do you see instead i see this error message see attached pic it happend to me while using firefox in vb ie and opera its always the same error please use labels and text to provide additional information original issue reported on code google com by mrlords gmail com on jan at attachments | 1 |
41,190 | 10,328,311,841 | IssuesEvent | 2019-09-02 09:14:11 | vector-im/riot-web | https://api.github.com/repos/vector-im/riot-web | closed | Handle the case of no IS in features that require IS to lookup | bug defect phase:1 privacy privacy-sprint type:identity-server | Since it's now possible to be disconnected from an identity server entirely (your session has no active identity server), it's unclear how features that require an IS to function should behave.
For the case of the Discovery section in Settings, it's easy enough: you have no IS, so we don't show any 3PIDs to control.
However, there are other IS-using features in Riot, like:
1. Inviting by 3PID via the address picker
<img width="792" alt="2019-08-23 at 16 02" src="https://user-images.githubusercontent.com/279572/63602530-8c9c7500-c5bf-11e9-8931-ec308e89de78.png">
2. Inviting by 3PID via `/invite bob@bob.com`
3. ~~Checking an email invite in the invite preview bar to ensure the invited email is for the active account (in case you have multiple Matrix accounts in different tabs, etc.)~~ (Extracted feature 3 to https://github.com/vector-im/riot-web/issues/10669.)
For features 1 and 2, I think we need some UX advice from @nadonomy on what they should do (communicate some kind of error to the user, etc.).
For feature 3, I think we are at an earlier stage where @lampholder needs to work out how the product should function for this case. | 1.0 | Handle the case of no IS in features that require IS to lookup - Since it's now possible to be disconnected from an identity server entirely (your session has no active identity server), it's unclear how features that require an IS to function should behave.
For the case of the Discovery section in Settings, it's easy enough: you have no IS, so we don't show any 3PIDs to control.
However, there are other IS-using features in Riot, like:
1. Inviting by 3PID via the address picker
<img width="792" alt="2019-08-23 at 16 02" src="https://user-images.githubusercontent.com/279572/63602530-8c9c7500-c5bf-11e9-8931-ec308e89de78.png">
2. Inviting by 3PID via `/invite bob@bob.com`
3. ~~Checking an email invite in the invite preview bar to ensure the invited email is for the active account (in case you have multiple Matrix accounts in different tabs, etc.)~~ (Extracted feature 3 to https://github.com/vector-im/riot-web/issues/10669.)
For features 1 and 2, I think we need some UX advice from @nadonomy on what they should do (communicate some kind of error to the user, etc.).
For feature 3, I think we are at an earlier stage where @lampholder needs to work out how the product should function for this case. | defect | handle the case of no is in features that require is to lookup since it s now possible to be disconnected from an identity server entirely your session has no active identity server it s unclear how features that require an is to function should behave for the case of the discovery section in settings it s easy enough you have no is so we don t show any to control however there are other is using features in riot like inviting by via the address picker img width alt at src inviting by via invite bob bob com checking an email invite in the invite preview bar to ensure the invited email is for the active account in case you have multiple matrix accounts in different tabs etc extracted feature to for features and i think we need some ux advice from nadonomy on what they should do communicate some kind of error to the user etc for feature i think we are at an earlier stage where lampholder needs to work out how the product should function for this case | 1 |
68,860 | 8,357,618,074 | IssuesEvent | 2018-10-02 22:15:00 | nextcloud/server | https://api.github.com/repos/nextcloud/server | closed | Improve backup codes UI/UX | 1. to develop Hacktoberfest design enhancement good first issue help wanted papercut | We recently added a new setting to the personal settings page – the ability to generate backup codes, which can be used in case users lose access to their second factor. At the moment, the user interface doesn't look that nice:




@nextcloud/designers any idea how we could make a prettier user interface out of that? Would be great if someone could create a mockup :art: :rocket:
ref https://github.com/nextcloud/server/pull/1171
| 1.0 | Improve backup codes UI/UX - We recently added a new setting to the personal settings page – the ability to generate backup codes, which can be used in case users lose access to their second factor. At the moment, the user interface doesn't look that nice:




@nextcloud/designers any idea how we could make a prettier user interface out of that? Would be great if someone could create a mockup :art: :rocket:
ref https://github.com/nextcloud/server/pull/1171
| non_defect | improve backup codes ui ux we recently added a new setting to the personal settings page – the ability to generate backup codes which can be used in case users lose access to their second factor at the moment the user interface doesn t look that nice nextcloud designers any idea how we could make a prettier user interface out of that would be great if someone could create a mockup art rocket ref | 0 |
1,683 | 2,812,799,119 | IssuesEvent | 2015-05-18 11:03:28 | facebook/fresco | https://api.github.com/repos/facebook/fresco | closed | Fresco + Realm.io library error | build duplicate needs-details | i use compile 'com.facebook.fresco:fresco:0.5.0+'
java.lang.UnsatisfiedLinkError: dalvik.system.PathClassLoader[DexPathList[[zip file "/system/framework/multiwindow.jar", zip file "/system/framework/android.test.runner.jar", zip file "/data/app/com.couchgram.privacycall-1/base.apk"],nativeLibraryDirectories=[/data/app/com.couchgram.privacycall-1/lib/arm64, /vendor/lib64, /system/lib64]]] couldn't find "libmemchunk.so"
at java.lang.Runtime.loadLibrary(Runtime.java:366)
at java.lang.System.loadLibrary(System.java:989)
at com.facebook.common.soloader.SoLoaderShim$DefaultHandler.loadLibrary(SoLoaderShim.java:34)
at com.facebook.common.soloader.SoLoaderShim.loadLibrary(SoLoaderShim.java:56)
at com.facebook.imagepipeline.memory.NativeMemoryChunk.<clinit>(NativeMemoryChunk.java:36)
at com.facebook.imagepipeline.memory.NativeMemoryChunkPool.alloc(NativeMemoryChunkPool.java:60)
at com.facebook.imagepipeline.memory.NativeMemoryChunkPool.alloc(NativeMemoryChunkPool.java:22)
at com.facebook.imagepipeline.memory.BasePool.get(BasePool.java:260)
at com.facebook.imagepipeline.memory.NativePooledByteBufferOutputStream.<init>(NativePooledByteBufferOutputStream.java:53)
at com.facebook.imagepipeline.memory.NativePooledByteBufferOutputStream.<init>(NativePooledByteBufferOutputStream.java:36)
at com.facebook.imagepipeline.memory.NativePooledByteBufferFactory.newByteBuffer(NativePooledByteBufferFactory.java:59)
at com.facebook.imagepipeline.memory.NativePooledByteBufferFactory.newByteBuffer(NativePooledByteBufferFactory.java:26)
at com.facebook.imagepipeline.producers.LocalFetchProducer$1.getResult(LocalFetchProducer.java:58)
at com.facebook.imagepipeline.producers.LocalFetchProducer$1.getResult(LocalFetchProducer.java:49)
at com.facebook.common.executors.StatefulRunnable.run(StatefulRunnable.java:45)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1112)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:587)
at java.lang.Thread.run(Thread.java:818) | 1.0 | Fresco + Realm.io library error - i use compile 'com.facebook.fresco:fresco:0.5.0+'
java.lang.UnsatisfiedLinkError: dalvik.system.PathClassLoader[DexPathList[[zip file "/system/framework/multiwindow.jar", zip file "/system/framework/android.test.runner.jar", zip file "/data/app/com.couchgram.privacycall-1/base.apk"],nativeLibraryDirectories=[/data/app/com.couchgram.privacycall-1/lib/arm64, /vendor/lib64, /system/lib64]]] couldn't find "libmemchunk.so"
at java.lang.Runtime.loadLibrary(Runtime.java:366)
at java.lang.System.loadLibrary(System.java:989)
at com.facebook.common.soloader.SoLoaderShim$DefaultHandler.loadLibrary(SoLoaderShim.java:34)
at com.facebook.common.soloader.SoLoaderShim.loadLibrary(SoLoaderShim.java:56)
at com.facebook.imagepipeline.memory.NativeMemoryChunk.<clinit>(NativeMemoryChunk.java:36)
at com.facebook.imagepipeline.memory.NativeMemoryChunkPool.alloc(NativeMemoryChunkPool.java:60)
at com.facebook.imagepipeline.memory.NativeMemoryChunkPool.alloc(NativeMemoryChunkPool.java:22)
at com.facebook.imagepipeline.memory.BasePool.get(BasePool.java:260)
at com.facebook.imagepipeline.memory.NativePooledByteBufferOutputStream.<init>(NativePooledByteBufferOutputStream.java:53)
at com.facebook.imagepipeline.memory.NativePooledByteBufferOutputStream.<init>(NativePooledByteBufferOutputStream.java:36)
at com.facebook.imagepipeline.memory.NativePooledByteBufferFactory.newByteBuffer(NativePooledByteBufferFactory.java:59)
at com.facebook.imagepipeline.memory.NativePooledByteBufferFactory.newByteBuffer(NativePooledByteBufferFactory.java:26)
at com.facebook.imagepipeline.producers.LocalFetchProducer$1.getResult(LocalFetchProducer.java:58)
at com.facebook.imagepipeline.producers.LocalFetchProducer$1.getResult(LocalFetchProducer.java:49)
at com.facebook.common.executors.StatefulRunnable.run(StatefulRunnable.java:45)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1112)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:587)
at java.lang.Thread.run(Thread.java:818) | non_defect | fresco realm io library error i use compile com facebook fresco fresco java lang unsatisfiedlinkerror dalvik system pathclassloader nativelibrarydirectories couldn t find libmemchunk so at java lang runtime loadlibrary runtime java at java lang system loadlibrary system java at com facebook common soloader soloadershim defaulthandler loadlibrary soloadershim java at com facebook common soloader soloadershim loadlibrary soloadershim java at com facebook imagepipeline memory nativememorychunk nativememorychunk java at com facebook imagepipeline memory nativememorychunkpool alloc nativememorychunkpool java at com facebook imagepipeline memory nativememorychunkpool alloc nativememorychunkpool java at com facebook imagepipeline memory basepool get basepool java at com facebook imagepipeline memory nativepooledbytebufferoutputstream nativepooledbytebufferoutputstream java at com facebook imagepipeline memory nativepooledbytebufferoutputstream nativepooledbytebufferoutputstream java at com facebook imagepipeline memory nativepooledbytebufferfactory newbytebuffer nativepooledbytebufferfactory java at com facebook imagepipeline memory nativepooledbytebufferfactory newbytebuffer nativepooledbytebufferfactory java at com facebook imagepipeline producers localfetchproducer getresult localfetchproducer java at com facebook imagepipeline producers localfetchproducer getresult localfetchproducer java at com facebook common executors statefulrunnable run statefulrunnable java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java | 0 |
15,993 | 2,870,250,090 | IssuesEvent | 2015-06-07 00:33:52 | pdelia/away3d | https://api.github.com/repos/pdelia/away3d | opened | Sprite3D Visual Glitches with Object3D when using Renderer.INTERSECTING_OBJECTS | auto-migrated Priority-Medium Type-Defect | #45 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:21Z
```
What steps will reproduce the problem?
1. view = new View3D({scene:scene, camera:camera,
renderer:Renderer.INTERSECTING_OBJECTS});
2. sprite = new Sprite2D(new Sprite_bmd(512, 512));
3. cube = new Cube({width:2.5, height:2.5, depth:2.5, y:-1.25});
What is the expected output? What do you see instead?
We should see a sprite cleanly cut by the cube at it's midway. Instead, the
entire cube mesh begins to breakdown. This may be related to Issue35
http://code.google.com/p/away3d/issues/detail?id=35.
Tests with deltaZ and ownSession will not solve the issue. The only
available measure I've discovered to remedy the issue is to switch to
Renderer.CORRECT_Z_ORDER, with some issues with Clipping. Perhaps Planes
are the solution for the time being.
What version of the product are you using? On what operating system?
I'm publishing with FP9 and haven't tested the results in FP10. This is Mac
OSX.
Please provide any additional information below.
```
Original issue reported on code.google.com by `robinlhu...@gmail.com` on 1 Jun 2009 at 9:03 | 1.0 | Sprite3D Visual Glitches with Object3D when using Renderer.INTERSECTING_OBJECTS - #45 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:21Z
```
What steps will reproduce the problem?
1. view = new View3D({scene:scene, camera:camera,
renderer:Renderer.INTERSECTING_OBJECTS});
2. sprite = new Sprite2D(new Sprite_bmd(512, 512));
3. cube = new Cube({width:2.5, height:2.5, depth:2.5, y:-1.25});
What is the expected output? What do you see instead?
We should see a sprite cleanly cut by the cube at it's midway. Instead, the
entire cube mesh begins to breakdown. This may be related to Issue35
http://code.google.com/p/away3d/issues/detail?id=35.
Tests with deltaZ and ownSession will not solve the issue. The only
available measure I've discovered to remedy the issue is to switch to
Renderer.CORRECT_Z_ORDER, with some issues with Clipping. Perhaps Planes
are the solution for the time being.
What version of the product are you using? On what operating system?
I'm publishing with FP9 and haven't tested the results in FP10. This is Mac
OSX.
Please provide any additional information below.
```
Original issue reported on code.google.com by `robinlhu...@gmail.com` on 1 Jun 2009 at 9:03 | defect | visual glitches with when using renderer intersecting objects issue by googlecodeexporter created on what steps will reproduce the problem view new scene scene camera camera renderer renderer intersecting objects sprite new new sprite bmd cube new cube width height depth y what is the expected output what do you see instead we should see a sprite cleanly cut by the cube at it s midway instead the entire cube mesh begins to breakdown this may be related to tests with deltaz and ownsession will not solve the issue the only available measure i ve discovered to remedy the issue is to switch to renderer correct z order with some issues with clipping perhaps planes are the solution for the time being what version of the product are you using on what operating system i m publishing with and haven t tested the results in this is mac osx please provide any additional information below original issue reported on code google com by robinlhu gmail com on jun at | 1 |
74,549 | 25,170,310,908 | IssuesEvent | 2022-11-11 02:14:00 | ascott18/TellMeWhen | https://api.github.com/repos/ascott18/TellMeWhen | closed | [Bug] Soul shard item count apparently doesn't work for Item Cooldown icons | T: defect V: classic | Reported by someone on Discord. Perhaps is due to different bag type for soul shard bags? | 1.0 | [Bug] Soul shard item count apparently doesn't work for Item Cooldown icons - Reported by someone on Discord. Perhaps is due to different bag type for soul shard bags? | defect | soul shard item count apparently doesn t work for item cooldown icons reported by someone on discord perhaps is due to different bag type for soul shard bags | 1 |
17,887 | 3,013,563,736 | IssuesEvent | 2015-07-29 09:45:03 | yawlfoundation/yawl | https://api.github.com/repos/yawlfoundation/yawl | closed | Can't save a specification with extended UTF-8 chars in data type definition | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Use for example the data type:
<xs:simpleType name="country">
<xs:restriction base="xs:string">
<xs:enumeration value="ÅLAND ISLANDS" />
<xs:enumeration value="TEST" />
</xs:restriction>
</xs:simpleType>
in your specification.
2. Try to save the specification leads to "Could not export Specification due
to missing or invalid user-definied datatype. Please see the problem list below
for details."
After rev 1755 there isn't an error in the data definition window anymore, but
the handling of UTF-8 seems not to be correct.
```
Original issue reported on code.google.com by `felix.mannhardt` on 12 Aug 2011 at 3:21 | 1.0 | Can't save a specification with extended UTF-8 chars in data type definition - ```
What steps will reproduce the problem?
1. Use for example the data type:
<xs:simpleType name="country">
<xs:restriction base="xs:string">
<xs:enumeration value="ÅLAND ISLANDS" />
<xs:enumeration value="TEST" />
</xs:restriction>
</xs:simpleType>
in your specification.
2. Try to save the specification leads to "Could not export Specification due
to missing or invalid user-definied datatype. Please see the problem list below
for details."
After rev 1755 there isn't an error in the data definition window anymore, but
the handling of UTF-8 seems not to be correct.
```
Original issue reported on code.google.com by `felix.mannhardt` on 12 Aug 2011 at 3:21 | defect | can t save a specification with extended utf chars in data type definition what steps will reproduce the problem use for example the data type in your specification try to save the specification leads to could not export specification due to missing or invalid user definied datatype please see the problem list below for details after rev there isn t an error in the data definition window anymore but the handling of utf seems not to be correct original issue reported on code google com by felix mannhardt on aug at | 1 |
55,174 | 14,257,032,075 | IssuesEvent | 2020-11-20 02:32:13 | naev/naev | https://api.github.com/repos/naev/naev | opened | make distcheck broken | Priority-High Type-Defect | ```
$ make distcheck
...
Making all in src
Making all in tk
Making all in widget
CC button.o
CC checkbox.o
CC cust.o
CC fader.o
CC image.o
CC imagearray.o
CC input.o
CC list.o
CC rect.o
CC tabwin.o
CC text.o
ar: `u' modifier ignored since `D' is the default (see `U')
CC ai.o
cc1: fatal error: ../src/attributes.h: No such file or directory
compilation terminated.
make[4]: *** [Makefile:925: ai.o] Error 1
make[3]: *** [Makefile:963: all-recursive] Error 1
make[2]: *** [Makefile:661: all] Error 2
make[1]: *** [Makefile:661: all-recursive] Error 1
make: *** [Makefile:879: distcheck] Error 1
``` | 1.0 | make distcheck broken - ```
$ make distcheck
...
Making all in src
Making all in tk
Making all in widget
CC button.o
CC checkbox.o
CC cust.o
CC fader.o
CC image.o
CC imagearray.o
CC input.o
CC list.o
CC rect.o
CC tabwin.o
CC text.o
ar: `u' modifier ignored since `D' is the default (see `U')
CC ai.o
cc1: fatal error: ../src/attributes.h: No such file or directory
compilation terminated.
make[4]: *** [Makefile:925: ai.o] Error 1
make[3]: *** [Makefile:963: all-recursive] Error 1
make[2]: *** [Makefile:661: all] Error 2
make[1]: *** [Makefile:661: all-recursive] Error 1
make: *** [Makefile:879: distcheck] Error 1
``` | defect | make distcheck broken make distcheck making all in src making all in tk making all in widget cc button o cc checkbox o cc cust o cc fader o cc image o cc imagearray o cc input o cc list o cc rect o cc tabwin o cc text o ar u modifier ignored since d is the default see u cc ai o fatal error src attributes h no such file or directory compilation terminated make error make error make error make error make error | 1 |
692,647 | 23,744,200,017 | IssuesEvent | 2022-08-31 14:42:51 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | ecerp01.ecil.gov.in - "Secure connection failed" when accessing the site | browser-firefox priority-normal severity-critical engine-gecko type-geolocation type-unsupported-tls | <!-- @browser: Firefox 78.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:78.0) Gecko/20100101 Firefox/78.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/53851 -->
**URL**: https://ecerp01.ecil.gov.in/ecilerec
**Browser / Version**: Firefox 78.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
sap pop up code come and every thing come begining
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/6/07f6c5dc-741a-4cfc-b894-9007fa7929ba.jpeg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200604213430</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/6/9ec93488-11a7-407b-9786-45ce0b11ddd2)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | ecerp01.ecil.gov.in - "Secure connection failed" when accessing the site - <!-- @browser: Firefox 78.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:78.0) Gecko/20100101 Firefox/78.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/53851 -->
**URL**: https://ecerp01.ecil.gov.in/ecilerec
**Browser / Version**: Firefox 78.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
sap pop up code come and every thing come begining
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/6/07f6c5dc-741a-4cfc-b894-9007fa7929ba.jpeg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200604213430</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/6/9ec93488-11a7-407b-9786-45ce0b11ddd2)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_defect | ecil gov in secure connection failed when accessing the site url browser version firefox operating system windows tested another browser yes edge problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce sap pop up code come and every thing come begining view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 0 |
22,783 | 3,698,519,087 | IssuesEvent | 2016-02-28 11:33:02 | dnaumenko/java-diff-utils | https://api.github.com/repos/dnaumenko/java-diff-utils | closed | Please include a LICENSE file in the source tree | auto-migrated Priority-Medium Type-Defect | ```
Including a LICENSE file makes it easier for downstream packagers and other
users to ensure that they are using the software in a manner consistent with
its license. Please include the Apache license in the source tree and in
source distributions: http://www.apache.org/licenses/LICENSE-2.0
```
Original issue reported on code.google.com by `willben...@gmail.com` on 21 Oct 2013 at 3:36 | 1.0 | Please include a LICENSE file in the source tree - ```
Including a LICENSE file makes it easier for downstream packagers and other
users to ensure that they are using the software in a manner consistent with
its license. Please include the Apache license in the source tree and in
source distributions: http://www.apache.org/licenses/LICENSE-2.0
```
Original issue reported on code.google.com by `willben...@gmail.com` on 21 Oct 2013 at 3:36 | defect | please include a license file in the source tree including a license file makes it easier for downstream packagers and other users to ensure that they are using the software in a manner consistent with its license please include the apache license in the source tree and in source distributions original issue reported on code google com by willben gmail com on oct at | 1 |
71,659 | 23,748,071,691 | IssuesEvent | 2022-08-31 17:50:00 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Panel layout broke across restart | T-Defect | ### Steps to reproduce
1. Minimise the spacepanel & leftpanel
2. Un-minimise them again
3. Restart the app (an upgrade, in this instance)
4. Observe that the app launches with minimised spacepanel & leftpanel
### Outcome
#### What did you expect?
Panel layout should persist across layouts
#### What happened instead?
Something failed to save when i expanded the panel layouts, it seems.
### Operating system
macOS
### Application version
nightly
### How did you install the app?
nightly
### Homeserver
matrix.org
### Will you send logs?
No | 1.0 | Panel layout broke across restart - ### Steps to reproduce
1. Minimise the spacepanel & leftpanel
2. Un-minimise them again
3. Restart the app (an upgrade, in this instance)
4. Observe that the app launches with minimised spacepanel & leftpanel
### Outcome
#### What did you expect?
Panel layout should persist across layouts
#### What happened instead?
Something failed to save when i expanded the panel layouts, it seems.
### Operating system
macOS
### Application version
nightly
### How did you install the app?
nightly
### Homeserver
matrix.org
### Will you send logs?
No | defect | panel layout broke across restart steps to reproduce minimise the spacepanel leftpanel un minimise them again restart the app an upgrade in this instance observe that the app launches with minimised spacepanel leftpanel outcome what did you expect panel layout should persist across layouts what happened instead something failed to save when i expanded the panel layouts it seems operating system macos application version nightly how did you install the app nightly homeserver matrix org will you send logs no | 1 |
42,332 | 10,966,177,776 | IssuesEvent | 2019-11-28 06:08:22 | vesoft-inc/nebula | https://api.github.com/repos/vesoft-inc/nebula | closed | Memory usage | defect-p3 |

nebula-storaged.conf:

partition:
CREATE SPACE test(partition_num=1024, replica_factor=1);
edge:
100 million
**Your Environments(__must be provided__)**
* OS: CentOS Linux release 7.4.1708
| 1.0 | Memory usage -

nebula-storaged.conf:

partition:
CREATE SPACE test(partition_num=1024, replica_factor=1);
edge:
100 million
**Your Environments(__must be provided__)**
* OS: CentOS Linux release 7.4.1708
| defect | memory usage nebula storaged conf: partition: create space test partition num replica factor edge: million your environments must be provided os centos linux release | 1 |
4,737 | 2,610,153,974 | IssuesEvent | 2015-02-26 18:49:01 | chrsmith/republic-at-war | https://api.github.com/repos/chrsmith/republic-at-war | closed | AT-TE | auto-migrated Priority-Medium Type-Defect | ```
ATTE glides sometimes
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:08 | 1.0 | AT-TE - ```
ATTE glides sometimes
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:08 | defect | at te atte glides sometimes original issue reported on code google com by gmail com on jan at | 1 |
448,756 | 31,811,398,823 | IssuesEvent | 2023-09-13 17:06:01 | fmi-faim/napari-psf-analysis | https://api.github.com/repos/fmi-faim/napari-psf-analysis | opened | Improve UI design | documentation enhancement | * Add guides to improve plugin flow (e.g. divider)
* Add plugin title, icon, link to documentation
* Add hint if layers are missing | 1.0 | Improve UI design - * Add guides to improve plugin flow (e.g. divider)
* Add plugin title, icon, link to documentation
* Add hint if layers are missing | non_defect | improve ui design add guides to improve plugin flow e g divider add plugin title icon link to documentation add hint if layers are missing | 0 |
255,823 | 8,126,543,129 | IssuesEvent | 2018-08-17 02:54:15 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Add ability to specify up vector in x ray image query. | Expected Use: 3 - Occasional Feature Impact: 3 - Medium Priority: Normal | There is currently no way to set the up vector on the view normal. It sets the up vector somewhat arbitrarily. The user should really be able to set it.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1191
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Add ability to specify up vector in x ray image query.
Assigned to: Kathleen Biagas
Category:
Target version: 2.6
Author: Eric Brugger
Start: 09/05/2012
Due date:
% Done: 0
Estimated time: 2.0
Created: 09/05/2012 11:40 am
Updated: 10/17/2012 05:59 pm
Likelihood:
Severity:
Found in version:
Impact: 3 - Medium
Expected Use: 3 - Occasional
OS: All
Support Group: Any
Description:
There is currently no way to set the up vector on the view normal. It sets the up vector somewhat arbitrarily. The user should really be able to set it.
Comments:
Added field to the XRay Image query widget. Added 'up_vector' parameter for setting via cli named arguments.M /src/avt/Filters/avtXRayFilter.hM /src/avt/Filters/avtXRayFilter.CM /src/avt/Queries/Queries/avtXRayImageQuery.CM /src/avt/Queries/Queries/avtXRayImageQuery.hM /src/gui/QvisXRayImageQueryWidget.CM /src/gui/QvisXRayImageQueryWidget.hM /src/visitpy/common/visitmodule.CM /src/resources/help/en_US/relnotes2.6.0.htmlM /test/tests/queries/xrayimage.pySVN Revisions 19382-19384.
| 1.0 | Add ability to specify up vector in x ray image query. - There is currently no way to set the up vector on the view normal. It sets the up vector somewhat arbitrarily. The user should really be able to set it.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1191
Status: Resolved
Project: VisIt
Tracker: Feature
Priority: Normal
Subject: Add ability to specify up vector in x ray image query.
Assigned to: Kathleen Biagas
Category:
Target version: 2.6
Author: Eric Brugger
Start: 09/05/2012
Due date:
% Done: 0
Estimated time: 2.0
Created: 09/05/2012 11:40 am
Updated: 10/17/2012 05:59 pm
Likelihood:
Severity:
Found in version:
Impact: 3 - Medium
Expected Use: 3 - Occasional
OS: All
Support Group: Any
Description:
There is currently no way to set the up vector on the view normal. It sets the up vector somewhat arbitrarily. The user should really be able to set it.
Comments:
Added field to the XRay Image query widget. Added 'up_vector' parameter for setting via cli named arguments.M /src/avt/Filters/avtXRayFilter.hM /src/avt/Filters/avtXRayFilter.CM /src/avt/Queries/Queries/avtXRayImageQuery.CM /src/avt/Queries/Queries/avtXRayImageQuery.hM /src/gui/QvisXRayImageQueryWidget.CM /src/gui/QvisXRayImageQueryWidget.hM /src/visitpy/common/visitmodule.CM /src/resources/help/en_US/relnotes2.6.0.htmlM /test/tests/queries/xrayimage.pySVN Revisions 19382-19384.
| non_defect | add ability to specify up vector in x ray image query there is currently no way to set the up vector on the view normal it sets the up vector somewhat arbitrarily the user should really be able to set it redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker feature priority normal subject add ability to specify up vector in x ray image query assigned to kathleen biagas category target version author eric brugger start due date done estimated time created am updated pm likelihood severity found in version impact medium expected use occasional os all support group any description there is currently no way to set the up vector on the view normal it sets the up vector somewhat arbitrarily the user should really be able to set it comments added field to the xray image query widget added up vector parameter for setting via cli named arguments m src avt filters avtxrayfilter hm src avt filters avtxrayfilter cm src avt queries queries avtxrayimagequery cm src avt queries queries avtxrayimagequery hm src gui qvisxrayimagequerywidget cm src gui qvisxrayimagequerywidget hm src visitpy common visitmodule cm src resources help en us htmlm test tests queries xrayimage pysvn revisions | 0 |
229,867 | 7,600,472,883 | IssuesEvent | 2018-04-28 02:33:49 | adventurerscodex/uat | https://api.github.com/repos/adventurerscodex/uat | closed | test_preview_armor fails | browser-firefox severity-high-priority type-bug | def test_preview_armor(player_wizard, browser): # noqa
"""As a player, I can select a row in the armor table and view the item
in the preview tab."""
print(('As a player, I can select a row in the armor table and view '
'the item in the preview tab'))
armor_add = armor.ArmorAddModal(browser)
armor_table = armor.ArmorTable(browser)
armor_preview = armor.ArmorPreviewModal(browser)
tabs = Tabs(browser)
tabs.equipment.click()
WebDriverWait(browser, 10).until(
EC.element_to_be_clickable(
(By.ID, armor_table.add_id)
)
)
armor_table.add.click()
ut.select_from_autocomplete(
armor_add,
'name',
browser,
has_search_term=False
)
armor_add.add.click()
WebDriverWait(browser, 10).until(
modal_finished_closing(armor_add.modal_div_id)
)
row = ut.get_table_row(armor_table, 'table', values=False)
row[0].click()
WebDriverWait(browser, 10).until(
EC.text_to_be_present_in_element(
> (By.ID, armor_preview.name_id), 'Breastplate'
)
)
TIMEOUT Exception | 1.0 | test_preview_armor fails - def test_preview_armor(player_wizard, browser): # noqa
"""As a player, I can select a row in the armor table and view the item
in the preview tab."""
print(('As a player, I can select a row in the armor table and view '
'the item in the preview tab'))
armor_add = armor.ArmorAddModal(browser)
armor_table = armor.ArmorTable(browser)
armor_preview = armor.ArmorPreviewModal(browser)
tabs = Tabs(browser)
tabs.equipment.click()
WebDriverWait(browser, 10).until(
EC.element_to_be_clickable(
(By.ID, armor_table.add_id)
)
)
armor_table.add.click()
ut.select_from_autocomplete(
armor_add,
'name',
browser,
has_search_term=False
)
armor_add.add.click()
WebDriverWait(browser, 10).until(
modal_finished_closing(armor_add.modal_div_id)
)
row = ut.get_table_row(armor_table, 'table', values=False)
row[0].click()
WebDriverWait(browser, 10).until(
EC.text_to_be_present_in_element(
> (By.ID, armor_preview.name_id), 'Breastplate'
)
)
TIMEOUT Exception | non_defect | test preview armor fails def test preview armor player wizard browser noqa as a player i can select a row in the armor table and view the item in the preview tab print as a player i can select a row in the armor table and view the item in the preview tab armor add armor armoraddmodal browser armor table armor armortable browser armor preview armor armorpreviewmodal browser tabs tabs browser tabs equipment click webdriverwait browser until ec element to be clickable by id armor table add id armor table add click ut select from autocomplete armor add name browser has search term false armor add add click webdriverwait browser until modal finished closing armor add modal div id row ut get table row armor table table values false row click webdriverwait browser until ec text to be present in element by id armor preview name id breastplate timeout exception | 0 |
35,371 | 7,721,055,227 | IssuesEvent | 2018-05-24 02:52:50 | AtlasOfLivingAustralia/layers-service | https://api.github.com/repos/AtlasOfLivingAustralia/layers-service | closed | No download and no thumbnails on http://spatial.ala.org.au/layers# | priority-critical type-defect | Looks like neither of the downloads is working on http://spatial.ala.org.au/layers# and the list doesn't include the thumbnail.
| 1.0 | No download and no thumbnails on http://spatial.ala.org.au/layers# - Looks like neither of the downloads is working on http://spatial.ala.org.au/layers# and the list doesn't include the thumbnail.
| defect | no download and no thumbnails on looks like neither of the downloads is working on and the list doesn t include the thumbnail | 1 |
30,106 | 6,025,660,231 | IssuesEvent | 2017-06-08 09:15:43 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | SelectOneButton: wrong unseletable behavior | defect | Currently if you click on a selected item, it will be unselected - also if unselectable is set to true. | 1.0 | SelectOneButton: wrong unseletable behavior - Currently if you click on a selected item, it will be unselected - also if unselectable is set to true. | defect | selectonebutton wrong unseletable behavior currently if you click on a selected item it will be unselected also if unselectable is set to true | 1 |
70,124 | 22,956,984,080 | IssuesEvent | 2022-07-19 12:28:46 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | FileDownload/CommandButton inside facet of a datatable, not working at all | :lady_beetle: defect :bangbang: needs-triage | ### Describe the bug
In a dataTable I have facets for input and output. In a Blob field, for documents, I render a column with fileupload for input, all working there, and for output I render a commandButton with a fileDownload, no way to make it work.
If I go for the actionListener route to prepare the Stream prior to be called by the filedownload component, it doesnt fire at all the actionlistener. Using process="@this" makes the actionlistenerr work, but crashes without throwing any error.
If I try to process the blob and create de stream direct in the method called by the filedownload, nothing happens.
This is the relevant sourc excerpt:
```
<p:column headerText="Documento">
<p:cellEditor id="conteudo">
<f:facet name="output">
<p:commandButton
value="Baixar o documento"
icon="pi pi-arrow-down"
styleClass="ui-button-outlined"
ajax="false">
<p:fileDownload value="#{cadastroEventoController.arquivo}"/>
</p:commandButton>
</f:facet>
<f:facet name="input">
<p:fileUpload listener="#{cadastroEventoController.trataUploadDeDocumento}"
mode="advanced" dragDropSupport="false"
update="dtDocumentos msgs" sizeLimit="100000000"
value="#{cadastroEventoController.arquivo}"
fileLimit="1" >
</p:fileUpload>
</f:facet>
</p:cellEditor>
</p:column>
```
Using ActionListener
```
<p:commandButton
value="Baixar o documento"
icon="pi pi-arrow-down"
styleClass="ui-button-outlined"
process="@this"
actionListener="#{cadastroEventoController.preparaArquivo()}"
ajax="false">
<p:fileDownload value="#{cadastroEventoController.arquivo}"/>
</p:commandButton>
```
Relevant methods:
```
// In Case direct use by filedownload
public StreamedContent getArquivo() {
FacesContext context = FacesContext.getCurrentInstance();
Integer linha = context.getApplication().evaluateExpressionGet(context, "#{linha}",Integer.class);
DocumentoEntity doc = eventoDTO.getObjeto().getDocumentos().get(linha);
ByteArrayInputStream stream = new ByteArrayInputStream(doc.getConteudo());
return new DefaultStreamedContent(stream, doc.getMime(), doc.getLegenda() + "." + doc.getExtensao());
}
// In Case ActionListener works
public void preparaArquivo() {
FacesContext context = FacesContext.getCurrentInstance();
Integer linha = context.getApplication().evaluateExpressionGet(context, "#{linha}",Integer.class);
DocumentoEntity doc = eventoDTO.getObjeto().getDocumentos().get(linha);
ByteArrayInputStream stream = new ByteArrayInputStream(doc.getConteudo());
setArquivo(new DefaultStreamedContent(stream, doc.getMime(), doc.getLegenda() + "." + doc.getExtensao()));
}
```
Extracting the CommandButton from the dataTable, but still inside the form, passing the index, makes any approach works to download the correct blob.
### Reproducer
_No response_
### Expected behavior
I would expect the FileDownload, inside a CommandButton, inside a dataTable, to try to get de BLOB contents through the provided attribute and initiate the download. Behavior that works perfectly outside the dataTable.
### PrimeFaces edition
Elite
### PrimeFaces version
8.0.16
### Theme
_No response_
### JSF implementation
_No response_
### JSF version
2.3
### Browser(s)
_No response_ | 1.0 | FileDownload/CommandButton inside facet of a datatable, not working at all - ### Describe the bug
In a dataTable I have facets for input and output. In a Blob field, for documents, I render a column with fileupload for input, all working there, and for output I render a commandButton with a fileDownload, no way to make it work.
If I go for the actionListener route to prepare the Stream prior to be called by the filedownload component, it doesnt fire at all the actionlistener. Using process="@this" makes the actionlistenerr work, but crashes without throwing any error.
If I try to process the blob and create de stream direct in the method called by the filedownload, nothing happens.
This is the relevant sourc excerpt:
```
<p:column headerText="Documento">
<p:cellEditor id="conteudo">
<f:facet name="output">
<p:commandButton
value="Baixar o documento"
icon="pi pi-arrow-down"
styleClass="ui-button-outlined"
ajax="false">
<p:fileDownload value="#{cadastroEventoController.arquivo}"/>
</p:commandButton>
</f:facet>
<f:facet name="input">
<p:fileUpload listener="#{cadastroEventoController.trataUploadDeDocumento}"
mode="advanced" dragDropSupport="false"
update="dtDocumentos msgs" sizeLimit="100000000"
value="#{cadastroEventoController.arquivo}"
fileLimit="1" >
</p:fileUpload>
</f:facet>
</p:cellEditor>
</p:column>
```
Using ActionListener
```
<p:commandButton
value="Baixar o documento"
icon="pi pi-arrow-down"
styleClass="ui-button-outlined"
process="@this"
actionListener="#{cadastroEventoController.preparaArquivo()}"
ajax="false">
<p:fileDownload value="#{cadastroEventoController.arquivo}"/>
</p:commandButton>
```
Relevant methods:
```
// In Case direct use by filedownload
public StreamedContent getArquivo() {
FacesContext context = FacesContext.getCurrentInstance();
Integer linha = context.getApplication().evaluateExpressionGet(context, "#{linha}",Integer.class);
DocumentoEntity doc = eventoDTO.getObjeto().getDocumentos().get(linha);
ByteArrayInputStream stream = new ByteArrayInputStream(doc.getConteudo());
return new DefaultStreamedContent(stream, doc.getMime(), doc.getLegenda() + "." + doc.getExtensao());
}
// In Case ActionListener works
public void preparaArquivo() {
FacesContext context = FacesContext.getCurrentInstance();
Integer linha = context.getApplication().evaluateExpressionGet(context, "#{linha}",Integer.class);
DocumentoEntity doc = eventoDTO.getObjeto().getDocumentos().get(linha);
ByteArrayInputStream stream = new ByteArrayInputStream(doc.getConteudo());
setArquivo(new DefaultStreamedContent(stream, doc.getMime(), doc.getLegenda() + "." + doc.getExtensao()));
}
```
Extracting the CommandButton from the dataTable, but still inside the form, passing the index, makes any approach works to download the correct blob.
### Reproducer
_No response_
### Expected behavior
I would expect the FileDownload, inside a CommandButton, inside a dataTable, to try to get de BLOB contents through the provided attribute and initiate the download. Behavior that works perfectly outside the dataTable.
### PrimeFaces edition
Elite
### PrimeFaces version
8.0.16
### Theme
_No response_
### JSF implementation
_No response_
### JSF version
2.3
### Browser(s)
_No response_ | defect | filedownload commandbutton inside facet of a datatable not working at all describe the bug in a datatable i have facets for input and output in a blob field for documents i render a column with fileupload for input all working there and for output i render a commandbutton with a filedownload no way to make it work if i go for the actionlistener route to prepare the stream prior to be called by the filedownload component it doesnt fire at all the actionlistener using process this makes the actionlistenerr work but crashes without throwing any error if i try to process the blob and create de stream direct in the method called by the filedownload nothing happens this is the relevant sourc excerpt p commandbutton value baixar o documento icon pi pi arrow down styleclass ui button outlined ajax false p fileupload listener cadastroeventocontroller tratauploaddedocumento mode advanced dragdropsupport false update dtdocumentos msgs sizelimit value cadastroeventocontroller arquivo filelimit using actionlistener p commandbutton value baixar o documento icon pi pi arrow down styleclass ui button outlined process this actionlistener cadastroeventocontroller preparaarquivo ajax false relevant methods in case direct use by filedownload public streamedcontent getarquivo facescontext context facescontext getcurrentinstance integer linha context getapplication evaluateexpressionget context linha integer class documentoentity doc eventodto getobjeto getdocumentos get linha bytearrayinputstream stream new bytearrayinputstream doc getconteudo return new defaultstreamedcontent stream doc getmime doc getlegenda doc getextensao in case actionlistener works public void preparaarquivo facescontext context facescontext getcurrentinstance integer linha context getapplication evaluateexpressionget context linha integer class documentoentity doc eventodto getobjeto getdocumentos get linha bytearrayinputstream stream new bytearrayinputstream doc getconteudo setarquivo new defaultstreamedcontent stream doc getmime doc getlegenda doc getextensao extracting the commandbutton from the datatable but still inside the form passing the index makes any approach works to download the correct blob reproducer no response expected behavior i would expect the filedownload inside a commandbutton inside a datatable to try to get de blob contents through the provided attribute and initiate the download behavior that works perfectly outside the datatable primefaces edition elite primefaces version theme no response jsf implementation no response jsf version browser s no response | 1 |
276,960 | 30,581,322,092 | IssuesEvent | 2023-07-21 09:54:23 | ministryofjustice/hmpps-probation-integration-services | https://api.github.com/repos/ministryofjustice/hmpps-probation-integration-services | closed | CVE-2023-34035 (pre-sentence-reports-to-delius) | dependencies security | Spring Security's authorization rules can be misconfigured when using multiple servlets
* Project: pre-sentence-reports-to-delius
* Package: `org.springframework.security:spring-security-config:6.1.1`
* Location: `app/libs/spring-security-config-6.1.1.jar`
>Spring Security versions 5.8 prior to 5.8.5, 6.0 prior to 6.0.5, and 6.1 prior to 6.1.2 could be susceptible to authorization rule misconfiguration if the application uses requestMatchers(String) and multiple servlets, one of them being Spring MVC’s DispatcherServlet. (DispatcherServlet is a Spring MVC component that maps HTTP endpoints to methods on @Controller-annotated classes.)
Specifically, an application is vulnerable when all of the following are true:
* Spring MVC is on the classpath
* Spring Security is securing more than one servlet in a single application (one of them being Spring MVC’s DispatcherServlet)
* The application uses requestMatchers(String) to refer to endpoints that are not Spring MVC endpoints
An application is not vulnerable if any of the following is true:
* The application does not have Spring MVC on the classpath
* The application secures no servlets other than Spring MVC’s DispatcherServlet
* The application uses requestMatchers(String) only for Spring MVC endpoints
https://avd.aquasec.com/nvd/cve-2023-34035
If the vulnerability does not impact the `pre-sentence-reports-to-delius` project, you can suppress this alert by adding a comment starting with `Suppress`. For example, "Suppressed because we do not process any untrusted XML content". | True | CVE-2023-34035 (pre-sentence-reports-to-delius) - Spring Security's authorization rules can be misconfigured when using multiple servlets
* Project: pre-sentence-reports-to-delius
* Package: `org.springframework.security:spring-security-config:6.1.1`
* Location: `app/libs/spring-security-config-6.1.1.jar`
>Spring Security versions 5.8 prior to 5.8.5, 6.0 prior to 6.0.5, and 6.1 prior to 6.1.2 could be susceptible to authorization rule misconfiguration if the application uses requestMatchers(String) and multiple servlets, one of them being Spring MVC’s DispatcherServlet. (DispatcherServlet is a Spring MVC component that maps HTTP endpoints to methods on @Controller-annotated classes.)
Specifically, an application is vulnerable when all of the following are true:
* Spring MVC is on the classpath
* Spring Security is securing more than one servlet in a single application (one of them being Spring MVC’s DispatcherServlet)
* The application uses requestMatchers(String) to refer to endpoints that are not Spring MVC endpoints
An application is not vulnerable if any of the following is true:
* The application does not have Spring MVC on the classpath
* The application secures no servlets other than Spring MVC’s DispatcherServlet
* The application uses requestMatchers(String) only for Spring MVC endpoints
https://avd.aquasec.com/nvd/cve-2023-34035
If the vulnerability does not impact the `pre-sentence-reports-to-delius` project, you can suppress this alert by adding a comment starting with `Suppress`. For example, "Suppressed because we do not process any untrusted XML content". | non_defect | cve pre sentence reports to delius spring security s authorization rules can be misconfigured when using multiple servlets project pre sentence reports to delius package org springframework security spring security config location app libs spring security config jar spring security versions prior to prior to and prior to could be susceptible to authorization rule misconfiguration if the application uses requestmatchers string and multiple servlets one of them being spring mvc’s dispatcherservlet dispatcherservlet is a spring mvc component that maps http endpoints to methods on controller annotated classes specifically an application is vulnerable when all of the following are true spring mvc is on the classpath spring security is securing more than one servlet in a single application one of them being spring mvc’s dispatcherservlet the application uses requestmatchers string to refer to endpoints that are not spring mvc endpoints an application is not vulnerable if any of the following is true the application does not have spring mvc on the classpath the application secures no servlets other than spring mvc’s dispatcherservlet the application uses requestmatchers string only for spring mvc endpoints if the vulnerability does not impact the pre sentence reports to delius project you can suppress this alert by adding a comment starting with suppress for example suppressed because we do not process any untrusted xml content | 0 |
434,990 | 30,480,764,155 | IssuesEvent | 2023-07-17 20:09:25 | openxla/iree | https://api.github.com/repos/openxla/iree | closed | [New documentation] Write docs for "TensorFlow Lite bindings" | documentation bindings/java bindings/tflite | * Live preview of this page: https://scotttodd.github.io/iree/bindings/tensorflow-lite/, which will eventually live on https://google.github.io/iree/ after we switch over the docs.
* ~~Source currently lives on the `staging-docs-rework` branch here: https://github.com/google/iree/blob/staging-docs-rework/mkdocs/docs/bindings/tensorflow-lite.md . PRs or direct commits to the branch are okay. We'll want to be careful about force pushes / rebases though.~~
Source lives in https://github.com/google/iree/blob/main/docs/website/docs/bindings/tensorflow-lite.md
* See the README here for instructions on previewing changes locally: ~~https://github.com/google/iree/tree/staging-docs-rework/mkdocs~~ https://github.com/google/iree/tree/main/docs/website
This will connect in some places with the corresponding frontend documentation (tracking issue: https://github.com/google/iree/issues/5455) and could discuss both the C API [bindings/tflite](https://github.com/google/iree/tree/main/bindings/tflite) and the Java API (https://github.com/google/iree/pull/5361).
We should specifically speak to the pros/cons of using these bindings over the alternatives.
If this is still being built / is unstable, we could omit the pages from the documentation site at first. | 1.0 | [New documentation] Write docs for "TensorFlow Lite bindings" - * Live preview of this page: https://scotttodd.github.io/iree/bindings/tensorflow-lite/, which will eventually live on https://google.github.io/iree/ after we switch over the docs.
* ~~Source currently lives on the `staging-docs-rework` branch here: https://github.com/google/iree/blob/staging-docs-rework/mkdocs/docs/bindings/tensorflow-lite.md . PRs or direct commits to the branch are okay. We'll want to be careful about force pushes / rebases though.~~
Source lives in https://github.com/google/iree/blob/main/docs/website/docs/bindings/tensorflow-lite.md
* See the README here for instructions on previewing changes locally: ~~https://github.com/google/iree/tree/staging-docs-rework/mkdocs~~ https://github.com/google/iree/tree/main/docs/website
This will connect in some places with the corresponding frontend documentation (tracking issue: https://github.com/google/iree/issues/5455) and could discuss both the C API [bindings/tflite](https://github.com/google/iree/tree/main/bindings/tflite) and the Java API (https://github.com/google/iree/pull/5361).
We should specifically speak to the pros/cons of using these bindings over the alternatives.
If this is still being built / is unstable, we could omit the pages from the documentation site at first. | non_defect | write docs for tensorflow lite bindings live preview of this page which will eventually live on after we switch over the docs source currently lives on the staging docs rework branch here prs or direct commits to the branch are okay we ll want to be careful about force pushes rebases though source lives in see the readme here for instructions on previewing changes locally this will connect in some places with the corresponding frontend documentation tracking issue and could discuss both the c api and the java api we should specifically speak to the pros cons of using these bindings over the alternatives if this is still being built is unstable we could omit the pages from the documentation site at first | 0 |
192,944 | 14,633,704,230 | IssuesEvent | 2020-12-24 02:50:33 | MarlinFirmware/Marlin | https://api.github.com/repos/MarlinFirmware/Marlin | closed | [BUG] TMC2209 UART In dual Z mode Z2 (E1) changes direction randomly | Needs: Testing | ### Bug Description
In dual Z mode for MKS Robin Nano V2 with TMC2209 controlled over UART second Z motor attached to E1 from time to time changes its direction when new move command send (no matter if over LCD or via PROTERFACE gcommand)
### Configuration Files
[ConfigHeaders.zip](https://github.com/MarlinFirmware/Marlin/files/5573281/ConfigHeaders.zip)
I have tested it in CoreXY mode and regular (separated motor for x and y)
### Steps to Reproduce
1. Start pronterface / LCD menu
2. Send Z move command (up or down) - multiple time
**Expected behavior:**
Both motor rotate the same direction or opposite (if inverted) but stable.
**Actual behavior:**
From time to time second Z motor changes it direction (randomly)
| 1.0 | [BUG] TMC2209 UART In dual Z mode Z2 (E1) changes direction randomly - ### Bug Description
In dual Z mode for MKS Robin Nano V2 with TMC2209 controlled over UART second Z motor attached to E1 from time to time changes its direction when new move command send (no matter if over LCD or via PROTERFACE gcommand)
### Configuration Files
[ConfigHeaders.zip](https://github.com/MarlinFirmware/Marlin/files/5573281/ConfigHeaders.zip)
I have tested it in CoreXY mode and regular (separated motor for x and y)
### Steps to Reproduce
1. Start pronterface / LCD menu
2. Send Z move command (up or down) - multiple time
**Expected behavior:**
Both motor rotate the same direction or opposite (if inverted) but stable.
**Actual behavior:**
From time to time second Z motor changes it direction (randomly)
| non_defect | uart in dual z mode changes direction randomly bug description in dual z mode for mks robin nano with controlled over uart second z motor attached to from time to time changes its direction when new move command send no matter if over lcd or via proterface gcommand configuration files i have tested it in corexy mode and regular separated motor for x and y steps to reproduce start pronterface lcd menu send z move command up or down multiple time expected behavior both motor rotate the same direction or opposite if inverted but stable actual behavior from time to time second z motor changes it direction randomly | 0 |
179,816 | 21,581,943,932 | IssuesEvent | 2022-05-02 19:44:15 | temporalio/temporal-ecommerce | https://api.github.com/repos/temporalio/temporal-ecommerce | closed | github.com/temporalio/sdk-go-v1.9.0: 2 vulnerabilities (highest severity is: 7.5) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/temporalio/sdk-go-v1.9.0</b></p></summary>
<p></p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/temporalio/temporal-ecommerce/commit/91f5d7a169fd359bad9e33b954d19e6f8f7adac5">91f5d7a169fd359bad9e33b954d19e6f8f7adac5</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2019-0205](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | github.com/uber-go/tally-v3.3.17 | Transitive | N/A | ❌ |
| [CVE-2019-0210](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0210) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | github.com/uber-go/tally-v3.3.17 | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-0205</summary>
### Vulnerable Library - <b>github.com/uber-go/tally-v3.3.17</b></p>
<p>A Go metrics interface with fast buffered metrics and third party reporters</p>
<p>
Dependency Hierarchy:
- github.com/temporalio/sdk-go-v1.9.0 (Root Library)
- :x: **github.com/uber-go/tally-v3.3.17** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/temporal-ecommerce/commit/91f5d7a169fd359bad9e33b954d19e6f8f7adac5">91f5d7a169fd359bad9e33b954d19e6f8f7adac5</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Apache Thrift all versions up to and including 0.12.0, a server or client may run into an endless loop when feed with specific input data. Because the issue had already been partially fixed in version 0.11.0, depending on the installed version it affects only certain language bindings.
<p>Publish Date: 2019-10-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205>CVE-2019-0205</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205</a></p>
<p>Release Date: 2019-10-29</p>
<p>Fix Resolution: org.apache.thrift:libthrift:0.13.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-0210</summary>
### Vulnerable Library - <b>github.com/uber-go/tally-v3.3.17</b></p>
<p>A Go metrics interface with fast buffered metrics and third party reporters</p>
<p>
Dependency Hierarchy:
- github.com/temporalio/sdk-go-v1.9.0 (Root Library)
- :x: **github.com/uber-go/tally-v3.3.17** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/temporal-ecommerce/commit/91f5d7a169fd359bad9e33b954d19e6f8f7adac5">91f5d7a169fd359bad9e33b954d19e6f8f7adac5</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Apache Thrift 0.9.3 to 0.12.0, a server implemented in Go using TJSONProtocol or TSimpleJSONProtocol may panic when feed with invalid input data.
<p>Publish Date: 2019-10-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0210>CVE-2019-0210</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://mail-archives.apache.org/mod_mbox/thrift-dev/201910.mbox/%3C277A46CA87494176B1BBCF5D72624A2A%40HAGGIS%3E">http://mail-archives.apache.org/mod_mbox/thrift-dev/201910.mbox/%3C277A46CA87494176B1BBCF5D72624A2A%40HAGGIS%3E</a></p>
<p>Release Date: 2019-10-29</p>
<p>Fix Resolution: 0.13.0</p>
</p>
<p></p>
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/uber-go/tally","packageVersion":"v3.3.17","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"github.com/temporalio/sdk-go:v1.9.0;github.com/uber-go/tally:v3.3.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.thrift:libthrift:0.13.0","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-0205","vulnerabilityDetails":"In Apache Thrift all versions up to and including 0.12.0, a server or client may run into an endless loop when feed with specific input data. Because the issue had already been partially fixed in version 0.11.0, depending on the installed version it affects only certain language bindings.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/uber-go/tally","packageVersion":"v3.3.17","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"github.com/temporalio/sdk-go:v1.9.0;github.com/uber-go/tally:v3.3.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.13.0","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-0210","vulnerabilityDetails":"In Apache Thrift 0.9.3 to 0.12.0, a server implemented in Go using TJSONProtocol or TSimpleJSONProtocol may panic when feed with invalid input data.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0210","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> --> | True | github.com/temporalio/sdk-go-v1.9.0: 2 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/temporalio/sdk-go-v1.9.0</b></p></summary>
<p></p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/temporalio/temporal-ecommerce/commit/91f5d7a169fd359bad9e33b954d19e6f8f7adac5">91f5d7a169fd359bad9e33b954d19e6f8f7adac5</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2019-0205](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | github.com/uber-go/tally-v3.3.17 | Transitive | N/A | ❌ |
| [CVE-2019-0210](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0210) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | github.com/uber-go/tally-v3.3.17 | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-0205</summary>
### Vulnerable Library - <b>github.com/uber-go/tally-v3.3.17</b></p>
<p>A Go metrics interface with fast buffered metrics and third party reporters</p>
<p>
Dependency Hierarchy:
- github.com/temporalio/sdk-go-v1.9.0 (Root Library)
- :x: **github.com/uber-go/tally-v3.3.17** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/temporal-ecommerce/commit/91f5d7a169fd359bad9e33b954d19e6f8f7adac5">91f5d7a169fd359bad9e33b954d19e6f8f7adac5</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Apache Thrift all versions up to and including 0.12.0, a server or client may run into an endless loop when feed with specific input data. Because the issue had already been partially fixed in version 0.11.0, depending on the installed version it affects only certain language bindings.
<p>Publish Date: 2019-10-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205>CVE-2019-0205</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205</a></p>
<p>Release Date: 2019-10-29</p>
<p>Fix Resolution: org.apache.thrift:libthrift:0.13.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-0210</summary>
### Vulnerable Library - <b>github.com/uber-go/tally-v3.3.17</b></p>
<p>A Go metrics interface with fast buffered metrics and third party reporters</p>
<p>
Dependency Hierarchy:
- github.com/temporalio/sdk-go-v1.9.0 (Root Library)
- :x: **github.com/uber-go/tally-v3.3.17** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/temporal-ecommerce/commit/91f5d7a169fd359bad9e33b954d19e6f8f7adac5">91f5d7a169fd359bad9e33b954d19e6f8f7adac5</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Apache Thrift 0.9.3 to 0.12.0, a server implemented in Go using TJSONProtocol or TSimpleJSONProtocol may panic when feed with invalid input data.
<p>Publish Date: 2019-10-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0210>CVE-2019-0210</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://mail-archives.apache.org/mod_mbox/thrift-dev/201910.mbox/%3C277A46CA87494176B1BBCF5D72624A2A%40HAGGIS%3E">http://mail-archives.apache.org/mod_mbox/thrift-dev/201910.mbox/%3C277A46CA87494176B1BBCF5D72624A2A%40HAGGIS%3E</a></p>
<p>Release Date: 2019-10-29</p>
<p>Fix Resolution: 0.13.0</p>
</p>
<p></p>
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/uber-go/tally","packageVersion":"v3.3.17","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"github.com/temporalio/sdk-go:v1.9.0;github.com/uber-go/tally:v3.3.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.thrift:libthrift:0.13.0","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-0205","vulnerabilityDetails":"In Apache Thrift all versions up to and including 0.12.0, a server or client may run into an endless loop when feed with specific input data. Because the issue had already been partially fixed in version 0.11.0, depending on the installed version it affects only certain language bindings.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/uber-go/tally","packageVersion":"v3.3.17","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"github.com/temporalio/sdk-go:v1.9.0;github.com/uber-go/tally:v3.3.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.13.0","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-0210","vulnerabilityDetails":"In Apache Thrift 0.9.3 to 0.12.0, a server implemented in Go using TJSONProtocol or TSimpleJSONProtocol may panic when feed with invalid input data.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0210","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> --> | non_defect | github com temporalio sdk go vulnerabilities highest severity is autoclosed vulnerable library github com temporalio sdk go found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high github com uber go tally transitive n a high github com uber go tally transitive n a details cve vulnerable library github com uber go tally a go metrics interface with fast buffered metrics and third party reporters dependency hierarchy github com temporalio sdk go root library x github com uber go tally vulnerable library found in head commit a href found in base branch main vulnerability details in apache thrift all versions up to and including a server or client may run into an endless loop when feed with specific input data because the issue had already been partially fixed in version depending on the installed version it affects only certain language bindings publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache thrift libthrift cve vulnerable library github com uber go tally a go metrics interface with fast buffered metrics and third party reporters dependency hierarchy github com temporalio sdk go root library x github com uber go tally vulnerable library found in head commit a href found in base branch main vulnerability details in apache thrift to a server implemented in go using tjsonprotocol or tsimplejsonprotocol may panic when feed with invalid input data publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution istransitivedependency true dependencytree github com temporalio sdk go github com uber go tally isminimumfixversionavailable true minimumfixversion org apache thrift libthrift isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails in apache thrift all versions up to and including a server or client may run into an endless loop when feed with specific input data because the issue had already been partially fixed in version depending on the installed version it affects only certain language bindings vulnerabilityurl istransitivedependency true dependencytree github com temporalio sdk go github com uber go tally isminimumfixversionavailable true minimumfixversion isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails in apache thrift to a server implemented in go using tjsonprotocol or tsimplejsonprotocol may panic when feed with invalid input data vulnerabilityurl | 0 |
14,790 | 25,732,942,363 | IssuesEvent | 2022-12-07 21:50:40 | pypa/packaging | https://api.github.com/repos/pypa/packaging | closed | InvalidRequirement: name@ git+...@branch; (python_version=="...") and extra == "..." | bug packaging.requirements | Hello. I found a requirement that packaging deems invalid but I am not sure why.
This is happening on 20.4, 20.9 and main branch @ 73507464faca545d83b748929e60ec67c0cb50b1.
I see what I believe is a proper requirement generated by https://github.com/sqlobject/sqlobject/blob/3.9.1/setup.py#L118-L121 as:
```
[oursql:python_version=="2.7"]
oursql@ git+https://github.com/sqlobject/oursql.git@master
[oursql:python_version>="3.4"]
oursql3@ git+https://github.com/sqlobject/oursql.git@py3k
```
Parsed by `importlib.metadata.PathDistribution.requires` as:
```
oursql@ git+https://github.com/sqlobject/oursql.git@master; (python_version=="2.7") and extra == "oursql"
oursql3@ git+https://github.com/sqlobject/oursql.git@py3k; (python_version>="3.4") and extra == "oursql"
```
Failing with:
```
>>> from packaging.requirements import Requirement
>>> Requirement('oursql@ git+https://github.com/sqlobject/oursql.git@master; (python_version=="2.7") and extra == "oursql"')
Traceback (most recent call last):
File "/home/churchyard/Dokumenty/pypa/packaging/packaging/requirements.py", line 102, in __init__
req = REQUIREMENT.parseString(requirement_string)
File "/usr/lib/python3.9/site-packages/pyparsing.py", line 1955, in parseString
raise exc
File "/usr/lib/python3.9/site-packages/pyparsing.py", line 3814, in parseImpl
raise ParseException(instring, loc, self.errmsg, self)
pyparsing.ParseException: Expected stringEnd, found '(' (at char 60), (line:1, col:61)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/churchyard/Dokumenty/pypa/packaging/packaging/requirements.py", line 104, in __init__
raise InvalidRequirement(
packaging.requirements.InvalidRequirement: Parse error at "'(python_'": Expected stringEnd
```
What's wrong with this requirement? | 1.0 | InvalidRequirement: name@ git+...@branch; (python_version=="...") and extra == "..." - Hello. I found a requirement that packaging deems invalid but I am not sure why.
This is happening on 20.4, 20.9 and main branch @ 73507464faca545d83b748929e60ec67c0cb50b1.
I see what I believe is a proper requirement generated by https://github.com/sqlobject/sqlobject/blob/3.9.1/setup.py#L118-L121 as:
```
[oursql:python_version=="2.7"]
oursql@ git+https://github.com/sqlobject/oursql.git@master
[oursql:python_version>="3.4"]
oursql3@ git+https://github.com/sqlobject/oursql.git@py3k
```
Parsed by `importlib.metadata.PathDistribution.requires` as:
```
oursql@ git+https://github.com/sqlobject/oursql.git@master; (python_version=="2.7") and extra == "oursql"
oursql3@ git+https://github.com/sqlobject/oursql.git@py3k; (python_version>="3.4") and extra == "oursql"
```
Failing with:
```
>>> from packaging.requirements import Requirement
>>> Requirement('oursql@ git+https://github.com/sqlobject/oursql.git@master; (python_version=="2.7") and extra == "oursql"')
Traceback (most recent call last):
File "/home/churchyard/Dokumenty/pypa/packaging/packaging/requirements.py", line 102, in __init__
req = REQUIREMENT.parseString(requirement_string)
File "/usr/lib/python3.9/site-packages/pyparsing.py", line 1955, in parseString
raise exc
File "/usr/lib/python3.9/site-packages/pyparsing.py", line 3814, in parseImpl
raise ParseException(instring, loc, self.errmsg, self)
pyparsing.ParseException: Expected stringEnd, found '(' (at char 60), (line:1, col:61)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/churchyard/Dokumenty/pypa/packaging/packaging/requirements.py", line 104, in __init__
raise InvalidRequirement(
packaging.requirements.InvalidRequirement: Parse error at "'(python_'": Expected stringEnd
```
What's wrong with this requirement? | non_defect | invalidrequirement name git branch python version and extra hello i found a requirement that packaging deems invalid but i am not sure why this is happening on and main branch i see what i believe is a proper requirement generated by as oursql git git parsed by importlib metadata pathdistribution requires as oursql git python version and extra oursql git python version and extra oursql failing with from packaging requirements import requirement requirement oursql git python version and extra oursql traceback most recent call last file home churchyard dokumenty pypa packaging packaging requirements py line in init req requirement parsestring requirement string file usr lib site packages pyparsing py line in parsestring raise exc file usr lib site packages pyparsing py line in parseimpl raise parseexception instring loc self errmsg self pyparsing parseexception expected stringend found at char line col during handling of the above exception another exception occurred traceback most recent call last file line in file home churchyard dokumenty pypa packaging packaging requirements py line in init raise invalidrequirement packaging requirements invalidrequirement parse error at python expected stringend what s wrong with this requirement | 0 |
186,758 | 15,083,338,636 | IssuesEvent | 2021-02-05 15:42:37 | CyVerse-learning-materials/foss | https://api.github.com/repos/CyVerse-learning-materials/foss | opened | Fix nav bar to match course schedule from top to bottom | documentation | Related to issue #29
- [ ] Resolve Issue #29 links to remove
- [ ] Reorder pages to match syntax of agenda
- [ ] Check github repo link at the bottom of the page. | 1.0 | Fix nav bar to match course schedule from top to bottom - Related to issue #29
- [ ] Resolve Issue #29 links to remove
- [ ] Reorder pages to match syntax of agenda
- [ ] Check github repo link at the bottom of the page. | non_defect | fix nav bar to match course schedule from top to bottom related to issue resolve issue links to remove reorder pages to match syntax of agenda check github repo link at the bottom of the page | 0 |
33,664 | 7,195,147,657 | IssuesEvent | 2018-02-04 14:15:44 | ganeti/ganeti | https://api.github.com/repos/ganeti/ganeti | closed | gnt-backup does not attach ext-based snapshot | AutoMigrated Component-storage Status:Fixed Type-Defect | Originally reported of Google Code with ID 1198.
```
What software version are you running? Please provide the output of "gnt-
cluster --version", "gnt-cluster version", and "hspace --version".
gnt-cluster (ganeti 2.15.2-99ds) 2.15.2
hspace (ganeti) version 2.15.2-99ds
compiled with ghc 7.6
running on linux x86_64
<b>What distribution are you using?</b>
Debian Jessie 8.6
<b>What steps will reproduce the problem?</b>
1. gnt-instance start INSTANCE
2. wait until instance is fully started.
3. gnt-backup -n NODE --noshutdown INSTANCE
<b>What is the expected output? What do you see instead?</b>
Exporting...finished.
<b>Please provide any additional information below.</b>
I use my own extstorage provider which I released last year to public here...
https://code.google.com/p/ganeti/wiki/ExtStorageProviders (see bottom > cLVM and "Shared" LVM).
Below is the error I get.
Fri Dec 16 09:52:47 2016 Creating a snapshot of disk/0 on node xen01.rkt-zentral.local
Fri Dec 16 09:52:48 2016 Exporting snapshot/0 from xen01.rkt-zentral.local to xen01.rkt-zentral.local
Fri Dec 16 09:52:51 2016 snapshot/0 is now listening, starting export
Fri Dec 16 09:52:52 2016 - WARNING: export on xen01.rkt-zentral.local failed: Failed to start export on xen01.rkt-zentral.local: Block device '<Disk(type=ext, logical_id=('lvm', '1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap'), children=None, visible as /dev/disk/0, size=10240m)>' is not set up
Fri Dec 16 09:52:52 2016 snapshot/0 failed to send data: Failed to start export on xen01.rkt-zentral.local: Block device '<Disk(type=ext, logical_id=('lvm', '1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap'), children=None, visible as /dev/disk/0, size=10240m)>' is not set up (recent output: None)
Fri Dec 16 09:52:52 2016 Removing snapshot of disk/0 on node xen01.rkt-zentral.local
Fri Dec 16 09:52:53 2016 - WARNING: Aborting import 'import-disk0-2016-12-16_09_52_48-WYeFav' on b661ad19-f12d-4631-9d8e-b913a68494cf
Fri Dec 16 09:52:53 2016 - WARNING: import 'import-disk0-2016-12-16_09_52_48-WYeFav' on xen01.rkt-zentral.local failed: Exited due to signal 15
Fri Dec 16 09:52:53 2016 snapshot/0 failed to receive data: Exited due to signal 15 (recent output: socat: W exiting on signal 15)
Fri Dec 16 09:52:53 2016 - WARNING: Some disk exports have failed; there may be leftover data for instance debian.rkt-zentral.local on node xen01.rkt-zentral.local
All executables in this provider write into a log file /usr/share/ganeti/extstorage/lvm/var.log.
There you can see, that the snapshot is created successfully, but then the snapshot cannot be found because gnt-backup seems not to export the VolumegroupName to the attach executable.
Here is a snippet...
>>>>
--- attach
rktstorage8
1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0
--- snapshot
rktstorage8
1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0
10240
1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap
--- attach
rktstorage8
1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0
--- attach
1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap
--- attach
1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap
<<<<<<
Here you can see that when it comes to attaching the snapshot "1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap" the parameter "vgname" is empty.
Using gnt-backup without "--noshutdown" works, because no snapshot has to be taken.
Please help.
Kind regards
Darius
```
Originally added on 2016-12-16 10:14:20 +0000 UTC. | 1.0 | gnt-backup does not attach ext-based snapshot - Originally reported of Google Code with ID 1198.
```
What software version are you running? Please provide the output of "gnt-
cluster --version", "gnt-cluster version", and "hspace --version".
gnt-cluster (ganeti 2.15.2-99ds) 2.15.2
hspace (ganeti) version 2.15.2-99ds
compiled with ghc 7.6
running on linux x86_64
<b>What distribution are you using?</b>
Debian Jessie 8.6
<b>What steps will reproduce the problem?</b>
1. gnt-instance start INSTANCE
2. wait until instance is fully started.
3. gnt-backup -n NODE --noshutdown INSTANCE
<b>What is the expected output? What do you see instead?</b>
Exporting...finished.
<b>Please provide any additional information below.</b>
I use my own extstorage provider which I released last year to public here...
https://code.google.com/p/ganeti/wiki/ExtStorageProviders (see bottom > cLVM and "Shared" LVM).
Below is the error I get.
Fri Dec 16 09:52:47 2016 Creating a snapshot of disk/0 on node xen01.rkt-zentral.local
Fri Dec 16 09:52:48 2016 Exporting snapshot/0 from xen01.rkt-zentral.local to xen01.rkt-zentral.local
Fri Dec 16 09:52:51 2016 snapshot/0 is now listening, starting export
Fri Dec 16 09:52:52 2016 - WARNING: export on xen01.rkt-zentral.local failed: Failed to start export on xen01.rkt-zentral.local: Block device '<Disk(type=ext, logical_id=('lvm', '1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap'), children=None, visible as /dev/disk/0, size=10240m)>' is not set up
Fri Dec 16 09:52:52 2016 snapshot/0 failed to send data: Failed to start export on xen01.rkt-zentral.local: Block device '<Disk(type=ext, logical_id=('lvm', '1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap'), children=None, visible as /dev/disk/0, size=10240m)>' is not set up (recent output: None)
Fri Dec 16 09:52:52 2016 Removing snapshot of disk/0 on node xen01.rkt-zentral.local
Fri Dec 16 09:52:53 2016 - WARNING: Aborting import 'import-disk0-2016-12-16_09_52_48-WYeFav' on b661ad19-f12d-4631-9d8e-b913a68494cf
Fri Dec 16 09:52:53 2016 - WARNING: import 'import-disk0-2016-12-16_09_52_48-WYeFav' on xen01.rkt-zentral.local failed: Exited due to signal 15
Fri Dec 16 09:52:53 2016 snapshot/0 failed to receive data: Exited due to signal 15 (recent output: socat: W exiting on signal 15)
Fri Dec 16 09:52:53 2016 - WARNING: Some disk exports have failed; there may be leftover data for instance debian.rkt-zentral.local on node xen01.rkt-zentral.local
All executables in this provider write into a log file /usr/share/ganeti/extstorage/lvm/var.log.
There you can see, that the snapshot is created successfully, but then the snapshot cannot be found because gnt-backup seems not to export the VolumegroupName to the attach executable.
Here is a snippet...
>>>>
--- attach
rktstorage8
1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0
--- snapshot
rktstorage8
1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0
10240
1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap
--- attach
rktstorage8
1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0
--- attach
1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap
--- attach
1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap
<<<<<<
Here you can see that when it comes to attaching the snapshot "1a3374d5-5070-4fe0-922b-6423c6f8a6c5.ext.disk0.snap" the parameter "vgname" is empty.
Using gnt-backup without "--noshutdown" works, because no snapshot has to be taken.
Please help.
Kind regards
Darius
```
Originally added on 2016-12-16 10:14:20 +0000 UTC. | defect | gnt backup does not attach ext based snapshot originally reported of google code with id what software version are you running please provide the output of gnt cluster version gnt cluster version and hspace version gnt cluster ganeti hspace ganeti version compiled with ghc running on linux what distribution are you using debian jessie what steps will reproduce the problem gnt instance start instance wait until instance is fully started gnt backup n node noshutdown instance what is the expected output what do you see instead exporting finished please provide any additional information below i use my own extstorage provider which i released last year to public here see bottom clvm and shared lvm below is the error i get fri dec creating a snapshot of disk on node rkt zentral local fri dec exporting snapshot from rkt zentral local to rkt zentral local fri dec snapshot is now listening starting export fri dec warning export on rkt zentral local failed failed to start export on rkt zentral local block device is not set up fri dec snapshot failed to send data failed to start export on rkt zentral local block device is not set up recent output none fri dec removing snapshot of disk on node rkt zentral local fri dec warning aborting import import wyefav on fri dec warning import import wyefav on rkt zentral local failed exited due to signal fri dec snapshot failed to receive data exited due to signal recent output socat w exiting on signal fri dec warning some disk exports have failed there may be leftover data for instance debian rkt zentral local on node rkt zentral local all executables in this provider write into a log file usr share ganeti extstorage lvm var log there you can see that the snapshot is created successfully but then the snapshot cannot be found because gnt backup seems not to export the volumegroupname to the attach executable here is a snippet attach ext snapshot ext ext snap attach ext attach ext snap attach ext snap here you can see that when it comes to attaching the snapshot ext snap the parameter vgname is empty using gnt backup without noshutdown works because no snapshot has to be taken please help kind regards darius originally added on utc | 1 |
49,643 | 13,187,244,688 | IssuesEvent | 2020-08-13 02:48:19 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | [icetray] uninitialized value (Trac #1799) | Incomplete Migration Migrated from Trac combo core defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1799">https://code.icecube.wisc.edu/ticket/1799</a>, reported by kjmeagher and owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:38",
"description": "found by static analysis: http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-581265.html#EndPath",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067158057333",
"component": "combo core",
"summary": "[icetray] uninitialized value",
"priority": "normal",
"keywords": "",
"time": "2016-07-27T08:01:06",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [icetray] uninitialized value (Trac #1799) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1799">https://code.icecube.wisc.edu/ticket/1799</a>, reported by kjmeagher and owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:38",
"description": "found by static analysis: http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-581265.html#EndPath",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067158057333",
"component": "combo core",
"summary": "[icetray] uninitialized value",
"priority": "normal",
"keywords": "",
"time": "2016-07-27T08:01:06",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| defect | uninitialized value trac migrated from json status closed changetime description found by static analysis reporter kjmeagher cc resolution fixed ts component combo core summary uninitialized value priority normal keywords time milestone owner olivas type defect | 1 |
49,600 | 13,187,238,627 | IssuesEvent | 2020-08-13 02:47:11 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | [tableio] named argument in class declaration (Trac #1734) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1734">https://code.icecube.wisc.edu/ticket/1734</a>, reported by kjmeagher and owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-06-10T08:01:03",
"description": "The sphinx build gives the following error:\n{{{\nTraceback (most recent call last):\n File \"/private/var/folders/rc/g_4_lyp9039cj1586zzg88f40000gn/T/pip-build-A327aa/sphinx/sphinx/ext/autodoc.py\", line 385, in import_object\n File \"/Users/kmeagher/icecube/combo/release/lib/icecube/tableio/enum3.py\", line 43\n class enum(baseEnum, metaclass=metaEnum):\n ^\nSyntaxError: invalid syntax\n}}}\n",
"reporter": "kjmeagher",
"cc": "",
"resolution": "wontfix",
"_ts": "1465545663829039",
"component": "cmake",
"summary": "[tableio] named argument in class declaration",
"priority": "normal",
"keywords": "",
"time": "2016-06-10T07:25:29",
"milestone": "Long-Term Future",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [tableio] named argument in class declaration (Trac #1734) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1734">https://code.icecube.wisc.edu/ticket/1734</a>, reported by kjmeagher and owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-06-10T08:01:03",
"description": "The sphinx build gives the following error:\n{{{\nTraceback (most recent call last):\n File \"/private/var/folders/rc/g_4_lyp9039cj1586zzg88f40000gn/T/pip-build-A327aa/sphinx/sphinx/ext/autodoc.py\", line 385, in import_object\n File \"/Users/kmeagher/icecube/combo/release/lib/icecube/tableio/enum3.py\", line 43\n class enum(baseEnum, metaclass=metaEnum):\n ^\nSyntaxError: invalid syntax\n}}}\n",
"reporter": "kjmeagher",
"cc": "",
"resolution": "wontfix",
"_ts": "1465545663829039",
"component": "cmake",
"summary": "[tableio] named argument in class declaration",
"priority": "normal",
"keywords": "",
"time": "2016-06-10T07:25:29",
"milestone": "Long-Term Future",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
| defect | named argument in class declaration trac migrated from json status closed changetime description the sphinx build gives the following error n ntraceback most recent call last n file private var folders rc g t pip build sphinx sphinx ext autodoc py line in import object n file users kmeagher icecube combo release lib icecube tableio py line n class enum baseenum metaclass metaenum n nsyntaxerror invalid syntax n n reporter kjmeagher cc resolution wontfix ts component cmake summary named argument in class declaration priority normal keywords time milestone long term future owner jvansanten type defect | 1 |
12,381 | 2,694,263,997 | IssuesEvent | 2015-04-01 19:16:21 | google/google-api-go-client | https://api.github.com/repos/google/google-api-go-client | opened | Support for Standard Query Parameters (fields, quotaUser, userIp, etc) in Google APIs | new priority-medium type-defect |
**shazow** on 5 Sep 2013 at 4:40:
```
Most (if not all) Google API's support standard query parameters, such as ones
documented here:
https://developers.google.com/discovery/v1/parameters
https://developers.google.com/drive/query-parameters
https://developers.google.com/analytics/devguides/reporting/realtime/v3/paramete
rs
Etc.
There doesn't seem to be a way to set these using the google-api-go-client
library.
Would be great if there was native support for these fields. Perhaps it would
make sense letting the user set arbitrary fields to the opt_ field map, for
future workarounds to such limitations?
Related bugs:
#5 https://code.google.com/p/google-api-go-client/issues/detail?id=5
#38 https://code.google.com/p/google-api-go-client/issues/detail?id=38
I would consider this bug to be a superset of the other two.
Thanks!
```
| 1.0 | Support for Standard Query Parameters (fields, quotaUser, userIp, etc) in Google APIs -
**shazow** on 5 Sep 2013 at 4:40:
```
Most (if not all) Google API's support standard query parameters, such as ones
documented here:
https://developers.google.com/discovery/v1/parameters
https://developers.google.com/drive/query-parameters
https://developers.google.com/analytics/devguides/reporting/realtime/v3/paramete
rs
Etc.
There doesn't seem to be a way to set these using the google-api-go-client
library.
Would be great if there was native support for these fields. Perhaps it would
make sense letting the user set arbitrary fields to the opt_ field map, for
future workarounds to such limitations?
Related bugs:
#5 https://code.google.com/p/google-api-go-client/issues/detail?id=5
#38 https://code.google.com/p/google-api-go-client/issues/detail?id=38
I would consider this bug to be a superset of the other two.
Thanks!
```
| defect | support for standard query parameters fields quotauser userip etc in google apis shazow on sep at most if not all google api s support standard query parameters such as ones documented here rs etc there doesn t seem to be a way to set these using the google api go client library would be great if there was native support for these fields perhaps it would make sense letting the user set arbitrary fields to the opt field map for future workarounds to such limitations related bugs i would consider this bug to be a superset of the other two thanks | 1 |
81,941 | 31,825,594,527 | IssuesEvent | 2023-09-14 07:17:27 | DependencyTrack/dependency-track | https://api.github.com/repos/DependencyTrack/dependency-track | opened | Unable to upload sbom file | defect in triage | ### Current Behavior
When I try to upload the SBOM file via UI (Upload BOM), i get a successfully updated message. However i don't see any component on the project. While analysing the API log i found following error message.
`2023-09-14 07:07:52,138 ERROR [BomUploadProcessingTask] Error while processing bom
org.cyclonedx.exception.ParseException: Unable to parse BOM from byte array
at org.cyclonedx.parsers.JsonParser.parse(JsonParser.java:70)
at org.dependencytrack.tasks.BomUploadProcessingTask.inform(BomUploadProcessingTask.java:105)
at alpine.event.framework.BaseEventService.lambda$publish$0(BaseEventService.java:101)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.base/java.lang.Thread.run(Unknown Source)
Caused by: com.fasterxml.jackson.databind.exc.MismatchedInputException: Cannot deserialize value of type `java.util.ArrayList<org.cyclonedx.model.Tool>` from Object value (token `JsonToken.START_OBJECT`)
at [Source: (byte[])"{
"bomFormat": "CycloneDX",
"specVersion": "1.5",
"serialNumber": "urn:uuid:a4da8087-25bb-4b89-98b9-f2b6c368d6a1",
"version": 1,
"metadata": {
"timestamp": "2023-09-14T05:24:09.218Z",
"tools": {
"components": [
{
"group": "@cyclonedx",
"name": "cdxgen",
"version": "9.3.1",
"purl": "pkg:npm/%40cyclonedx/cdxgen@9.3.1",
"type": "application",
"bom-ref": "pkg:npm/@cyclonedx/cdxgen@9.3.1"
}
]
"[truncated 46201 bytes]; line: 8, column: 14] (through reference chain: org.cyclonedx.model.Bom["metadata"]->org.cyclonedx.model.Metadata["tools"])
at com.fasterxml.jackson.databind.exc.MismatchedInputException.from(MismatchedInputException.java:59)
at com.fasterxml.jackson.databind.DeserializationContext.reportInputMismatch(DeserializationContext.java:1746)
at com.fasterxml.jackson.databind.DeserializationContext.handleUnexpectedToken(DeserializationContext.java:1520)
at com.fasterxml.jackson.databind.DeserializationContext.handleUnexpectedToken(DeserializationContext.java:1467)
at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.handleNonArray(CollectionDeserializer.java:396)
at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.deserialize(CollectionDeserializer.java:252)
at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.deserialize(CollectionDeserializer.java:28)
at com.fasterxml.jackson.databind.deser.impl.MethodProperty.deserializeAndSet(MethodProperty.java:129)
at com.fasterxml.jackson.databind.deser.BeanDeserializer.vanillaDeserialize(BeanDeserializer.java:314)
at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:177)
at com.fasterxml.jackson.databind.deser.impl.MethodProperty.deserializeAndSet(MethodProperty.java:129)
at com.fasterxml.jackson.databind.deser.BeanDeserializer.vanillaDeserialize(BeanDeserializer.java:314)
at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:177)
at com.fasterxml.jackson.databind.deser.DefaultDeserializationContext.readRootValue(DefaultDeserializationContext.java:323)
at com.fasterxml.jackson.databind.ObjectMapper._readMapAndClose(ObjectMapper.java:4730)
at com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3738)
at org.cyclonedx.parsers.JsonParser.parse(JsonParser.java:68)
... 5 common frames omitted
2023-09-14 07:07:52,175 DEBUG [NotificationService] Dispatching notification: class alpine.notification.Notification
2023-09-14 07:07:52,175 DEBUG [NotificationService] Alerting subscriber org.dependencytrack.notification.NotificationRouter
2023-09-14 07:07:52,175 DEBUG [EventService] Dispatching event: class org.dependencytrack.event.IndexEvent
2023-09-14 07:07:52,175 DEBUG [EventService] Alerting subscriber org.dependencytrack.tasks.IndexTask
2023-09-14 07:07:52,179 DEBUG [EventService] Dispatching event: class org.dependencytrack.event.IndexEvent
2023-09-14 07:07:52,179 DEBUG [EventService] Alerting subscriber org.dependencytrack.tasks.IndexTask
2023-09-14 07:07:52,179 DEBUG [EventService] Calling onSuccess
2023-09-14 07:07:52,179 DEBUG [EventService] Calling onSuccess
2023-09-14 07:07:52,185 DEBUG [EventService] Calling onSuccess`
### Steps to Reproduce
1. Login into DependencyTrack UI
2. Navigate to the Project > Components
3. Click on Upload BOM
4. Select the BOM file (file [attached)](url)
[bom.zip](https://github.com/DependencyTrack/dependency-track/files/12605879/bom.zip)
5. Click Upload, a success message will appear on the UI but the upload will fail at the api level
### Expected Behavior
BOM file need to uploaded and processed successfully
### Dependency-Track Version
4.8.2
### Dependency-Track Distribution
Container Image
### Database Server
PostgreSQL
### Database Server Version
_No response_
### Browser
Google Chrome
### Checklist
- [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues)
- [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported | 1.0 | Unable to upload sbom file - ### Current Behavior
When I try to upload the SBOM file via UI (Upload BOM), i get a successfully updated message. However i don't see any component on the project. While analysing the API log i found following error message.
`2023-09-14 07:07:52,138 ERROR [BomUploadProcessingTask] Error while processing bom
org.cyclonedx.exception.ParseException: Unable to parse BOM from byte array
at org.cyclonedx.parsers.JsonParser.parse(JsonParser.java:70)
at org.dependencytrack.tasks.BomUploadProcessingTask.inform(BomUploadProcessingTask.java:105)
at alpine.event.framework.BaseEventService.lambda$publish$0(BaseEventService.java:101)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.base/java.lang.Thread.run(Unknown Source)
Caused by: com.fasterxml.jackson.databind.exc.MismatchedInputException: Cannot deserialize value of type `java.util.ArrayList<org.cyclonedx.model.Tool>` from Object value (token `JsonToken.START_OBJECT`)
at [Source: (byte[])"{
"bomFormat": "CycloneDX",
"specVersion": "1.5",
"serialNumber": "urn:uuid:a4da8087-25bb-4b89-98b9-f2b6c368d6a1",
"version": 1,
"metadata": {
"timestamp": "2023-09-14T05:24:09.218Z",
"tools": {
"components": [
{
"group": "@cyclonedx",
"name": "cdxgen",
"version": "9.3.1",
"purl": "pkg:npm/%40cyclonedx/cdxgen@9.3.1",
"type": "application",
"bom-ref": "pkg:npm/@cyclonedx/cdxgen@9.3.1"
}
]
"[truncated 46201 bytes]; line: 8, column: 14] (through reference chain: org.cyclonedx.model.Bom["metadata"]->org.cyclonedx.model.Metadata["tools"])
at com.fasterxml.jackson.databind.exc.MismatchedInputException.from(MismatchedInputException.java:59)
at com.fasterxml.jackson.databind.DeserializationContext.reportInputMismatch(DeserializationContext.java:1746)
at com.fasterxml.jackson.databind.DeserializationContext.handleUnexpectedToken(DeserializationContext.java:1520)
at com.fasterxml.jackson.databind.DeserializationContext.handleUnexpectedToken(DeserializationContext.java:1467)
at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.handleNonArray(CollectionDeserializer.java:396)
at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.deserialize(CollectionDeserializer.java:252)
at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.deserialize(CollectionDeserializer.java:28)
at com.fasterxml.jackson.databind.deser.impl.MethodProperty.deserializeAndSet(MethodProperty.java:129)
at com.fasterxml.jackson.databind.deser.BeanDeserializer.vanillaDeserialize(BeanDeserializer.java:314)
at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:177)
at com.fasterxml.jackson.databind.deser.impl.MethodProperty.deserializeAndSet(MethodProperty.java:129)
at com.fasterxml.jackson.databind.deser.BeanDeserializer.vanillaDeserialize(BeanDeserializer.java:314)
at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:177)
at com.fasterxml.jackson.databind.deser.DefaultDeserializationContext.readRootValue(DefaultDeserializationContext.java:323)
at com.fasterxml.jackson.databind.ObjectMapper._readMapAndClose(ObjectMapper.java:4730)
at com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3738)
at org.cyclonedx.parsers.JsonParser.parse(JsonParser.java:68)
... 5 common frames omitted
2023-09-14 07:07:52,175 DEBUG [NotificationService] Dispatching notification: class alpine.notification.Notification
2023-09-14 07:07:52,175 DEBUG [NotificationService] Alerting subscriber org.dependencytrack.notification.NotificationRouter
2023-09-14 07:07:52,175 DEBUG [EventService] Dispatching event: class org.dependencytrack.event.IndexEvent
2023-09-14 07:07:52,175 DEBUG [EventService] Alerting subscriber org.dependencytrack.tasks.IndexTask
2023-09-14 07:07:52,179 DEBUG [EventService] Dispatching event: class org.dependencytrack.event.IndexEvent
2023-09-14 07:07:52,179 DEBUG [EventService] Alerting subscriber org.dependencytrack.tasks.IndexTask
2023-09-14 07:07:52,179 DEBUG [EventService] Calling onSuccess
2023-09-14 07:07:52,179 DEBUG [EventService] Calling onSuccess
2023-09-14 07:07:52,185 DEBUG [EventService] Calling onSuccess`
### Steps to Reproduce
1. Login into DependencyTrack UI
2. Navigate to the Project > Components
3. Click on Upload BOM
4. Select the BOM file (file [attached)](url)
[bom.zip](https://github.com/DependencyTrack/dependency-track/files/12605879/bom.zip)
5. Click Upload, a success message will appear on the UI but the upload will fail at the api level
### Expected Behavior
BOM file need to uploaded and processed successfully
### Dependency-Track Version
4.8.2
### Dependency-Track Distribution
Container Image
### Database Server
PostgreSQL
### Database Server Version
_No response_
### Browser
Google Chrome
### Checklist
- [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues)
- [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported | defect | unable to upload sbom file current behavior when i try to upload the sbom file via ui upload bom i get a successfully updated message however i don t see any component on the project while analysing the api log i found following error message error error while processing bom org cyclonedx exception parseexception unable to parse bom from byte array at org cyclonedx parsers jsonparser parse jsonparser java at org dependencytrack tasks bomuploadprocessingtask inform bomuploadprocessingtask java at alpine event framework baseeventservice lambda publish baseeventservice java at java base java util concurrent threadpoolexecutor runworker unknown source at java base java util concurrent threadpoolexecutor worker run unknown source at java base java lang thread run unknown source caused by com fasterxml jackson databind exc mismatchedinputexception cannot deserialize value of type java util arraylist from object value token jsontoken start object at bomformat cyclonedx specversion serialnumber urn uuid version metadata timestamp tools components group cyclonedx name cdxgen version purl pkg npm cdxgen type application bom ref pkg npm cyclonedx cdxgen line column through reference chain org cyclonedx model bom org cyclonedx model metadata at com fasterxml jackson databind exc mismatchedinputexception from mismatchedinputexception java at com fasterxml jackson databind deserializationcontext reportinputmismatch deserializationcontext java at com fasterxml jackson databind deserializationcontext handleunexpectedtoken deserializationcontext java at com fasterxml jackson databind deserializationcontext handleunexpectedtoken deserializationcontext java at com fasterxml jackson databind deser std collectiondeserializer handlenonarray collectiondeserializer java at com fasterxml jackson databind deser std collectiondeserializer deserialize collectiondeserializer java at com fasterxml jackson databind deser std collectiondeserializer deserialize collectiondeserializer java at com fasterxml jackson databind deser impl methodproperty deserializeandset methodproperty java at com fasterxml jackson databind deser beandeserializer vanilladeserialize beandeserializer java at com fasterxml jackson databind deser beandeserializer deserialize beandeserializer java at com fasterxml jackson databind deser impl methodproperty deserializeandset methodproperty java at com fasterxml jackson databind deser beandeserializer vanilladeserialize beandeserializer java at com fasterxml jackson databind deser beandeserializer deserialize beandeserializer java at com fasterxml jackson databind deser defaultdeserializationcontext readrootvalue defaultdeserializationcontext java at com fasterxml jackson databind objectmapper readmapandclose objectmapper java at com fasterxml jackson databind objectmapper readvalue objectmapper java at org cyclonedx parsers jsonparser parse jsonparser java common frames omitted debug dispatching notification class alpine notification notification debug alerting subscriber org dependencytrack notification notificationrouter debug dispatching event class org dependencytrack event indexevent debug alerting subscriber org dependencytrack tasks indextask debug dispatching event class org dependencytrack event indexevent debug alerting subscriber org dependencytrack tasks indextask debug calling onsuccess debug calling onsuccess debug calling onsuccess steps to reproduce login into dependencytrack ui navigate to the project components click on upload bom select the bom file file url click upload a success message will appear on the ui but the upload will fail at the api level expected behavior bom file need to uploaded and processed successfully dependency track version dependency track distribution container image database server postgresql database server version no response browser google chrome checklist i have read and understand the i have checked the for whether this defect was already reported | 1 |
66,143 | 20,016,796,631 | IssuesEvent | 2022-02-01 12:55:18 | BOINC/boinc | https://api.github.com/repos/BOINC/boinc | closed | Problem with <max_concurrent>0</max_concurrent> | C: Client - Scheduler Policy P: Undetermined R: invalid T: Defect | **Describe the bug**
I have set this app_config, but boinc still computes rosetta app in addition to rosetta_python_projects
[app_config.txt](https://github.com/BOINC/boinc/files/7966612/app_config.txt)
**Steps To Reproduce**
1.
2.
**Expected behavior**
<max_concurrent>0</max_concurrent> should disallow computing of given app
**Screenshots**
**System Information**
- OS: Microsoft Windows [Version 10.0.22000.469]
- BOINC Version: 7.16.20
**Additional context**
| 1.0 | Problem with <max_concurrent>0</max_concurrent> - **Describe the bug**
I have set this app_config, but boinc still computes rosetta app in addition to rosetta_python_projects
[app_config.txt](https://github.com/BOINC/boinc/files/7966612/app_config.txt)
**Steps To Reproduce**
1.
2.
**Expected behavior**
<max_concurrent>0</max_concurrent> should disallow computing of given app
**Screenshots**
**System Information**
- OS: Microsoft Windows [Version 10.0.22000.469]
- BOINC Version: 7.16.20
**Additional context**
| defect | problem with describe the bug i have set this app config but boinc still computes rosetta app in addition to rosetta python projects steps to reproduce expected behavior should disallow computing of given app screenshots system information os microsoft windows boinc version additional context | 1 |
308,109 | 26,577,984,845 | IssuesEvent | 2023-01-22 03:18:30 | pola-rs/polars | https://api.github.com/repos/pola-rs/polars | closed | Use pytest markers to distinguish between test types | python test | ### Problem description
I think the [custom marker functionality](https://docs.pytest.org/en/7.1.x/example/markers.html#marking-test-functions-and-selecting-them-for-a-run) from pytest could be nice for us. We could mark parametric/slow tests explicitly, rather than relying on folder structure.
You can apply a mark to [all tests](https://docs.pytest.org/en/7.1.x/example/markers.html#marking-whole-classes-or-modules) in a module easily.
Might be a worthwhile improvement. | 1.0 | Use pytest markers to distinguish between test types - ### Problem description
I think the [custom marker functionality](https://docs.pytest.org/en/7.1.x/example/markers.html#marking-test-functions-and-selecting-them-for-a-run) from pytest could be nice for us. We could mark parametric/slow tests explicitly, rather than relying on folder structure.
You can apply a mark to [all tests](https://docs.pytest.org/en/7.1.x/example/markers.html#marking-whole-classes-or-modules) in a module easily.
Might be a worthwhile improvement. | non_defect | use pytest markers to distinguish between test types problem description i think the from pytest could be nice for us we could mark parametric slow tests explicitly rather than relying on folder structure you can apply a mark to in a module easily might be a worthwhile improvement | 0 |
17,020 | 2,966,744,197 | IssuesEvent | 2015-07-12 06:45:59 | ElvisChiang/welly | https://api.github.com/repos/ElvisChiang/welly | closed | 中文介面下書籤功能出問題 | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1.將作業系統介面改成繁體中文
What is the expected output? What do you see instead?
如圖所示
英文介面:https://www.dropbox.com/s/cz0ac9oj2io30jg/Screenshot%202014-07-28
%2009.47.45.png
中文介面:https://www.dropbox.com/s/vhcjws0hq2tovia/%E8%9E%A2%E5%B9%95%E6%8
8%AA%E5%9C%96%202014-07-28%2009.49.41.png
What version of the product are you using? On what operating system?
Welly版本從2.6一直用到最新版2.7.fix都有同樣問題
作業系統也是從10.6.8到目前最新的10.9.4
Please provide any additional information below.
```
Original issue reported on code.google.com by `swn...@gmail.com` on 28 Jul 2014 at 7:55 | 1.0 | 中文介面下書籤功能出問題 - ```
What steps will reproduce the problem?
1.將作業系統介面改成繁體中文
What is the expected output? What do you see instead?
如圖所示
英文介面:https://www.dropbox.com/s/cz0ac9oj2io30jg/Screenshot%202014-07-28
%2009.47.45.png
中文介面:https://www.dropbox.com/s/vhcjws0hq2tovia/%E8%9E%A2%E5%B9%95%E6%8
8%AA%E5%9C%96%202014-07-28%2009.49.41.png
What version of the product are you using? On what operating system?
Welly版本從2.6一直用到最新版2.7.fix都有同樣問題
作業系統也是從10.6.8到目前最新的10.9.4
Please provide any additional information below.
```
Original issue reported on code.google.com by `swn...@gmail.com` on 28 Jul 2014 at 7:55 | defect | 中文介面下書籤功能出問題 what steps will reproduce the problem 將作業系統介面改成繁體中文 what is the expected output what do you see instead 如圖所示 英文介面: png 中文介面: aa png what version of the product are you using on what operating system fix都有同樣問題 please provide any additional information below original issue reported on code google com by swn gmail com on jul at | 1 |
18,048 | 3,022,104,717 | IssuesEvent | 2015-07-31 18:22:32 | rolly32/support-tools | https://api.github.com/repos/rolly32/support-tools | reopened | Convert googlecode-issues-exporter.py to use python3 | auto-migrated Priority-Low Type-Defect | ```
python3 googlecode-issues-exporter/github_issue_converter.py <a lot of options>
File "googlecode-issues-exporter/github_issue_converter.py", line 423
print "\nDone!\n"
^
SyntaxError: Missing parentheses in call to 'print'
---
Reported on google-code-shutdown@
```
Original issue reported on code.google.com by `chrsm...@google.com` on 13 Mar 2015 at 10:07 | 1.0 | Convert googlecode-issues-exporter.py to use python3 - ```
python3 googlecode-issues-exporter/github_issue_converter.py <a lot of options>
File "googlecode-issues-exporter/github_issue_converter.py", line 423
print "\nDone!\n"
^
SyntaxError: Missing parentheses in call to 'print'
---
Reported on google-code-shutdown@
```
Original issue reported on code.google.com by `chrsm...@google.com` on 13 Mar 2015 at 10:07 | defect | convert googlecode issues exporter py to use googlecode issues exporter github issue converter py file googlecode issues exporter github issue converter py line print ndone n syntaxerror missing parentheses in call to print reported on google code shutdown original issue reported on code google com by chrsm google com on mar at | 1 |
48,675 | 20,204,819,718 | IssuesEvent | 2022-02-11 19:02:05 | codeoverflow-org/nodecg-io | https://api.github.com/repos/codeoverflow-org/nodecg-io | closed | Bundles using Reddit service fail to compile | bug minor service | ### Description
Newly generated bundles that use the reddit service currently fail to compile.
The odd thing is that it does not affect the sample bundle which still builds fine.
### How to reproduce
1. Create a fresh nodecg install and install nodecg-io (current `main` or `0.2`)
2. Generate a new bundle that uses the reddit service using `nodecg-io generate`
3. Bundle fails to compile with the error below
### Error Log
Compile failure:
<!-- If the console contains some information that could
help identifying the problem paste them in the next empty line -->
```
> tsc -b
node_modules/reddit-ts/src/RedditAPI.ts:185:25 - error TS2345: Argument of type 'Child | undefined' is not assignable to parameter of type 'Child'.
Type 'undefined' is not assignable to type 'Child'.
185 return map_search(data.children[0])
~~~~~~~~~~~~~~~~
Found 1 error.
```
| 1.0 | Bundles using Reddit service fail to compile - ### Description
Newly generated bundles that use the reddit service currently fail to compile.
The odd thing is that it does not affect the sample bundle which still builds fine.
### How to reproduce
1. Create a fresh nodecg install and install nodecg-io (current `main` or `0.2`)
2. Generate a new bundle that uses the reddit service using `nodecg-io generate`
3. Bundle fails to compile with the error below
### Error Log
Compile failure:
<!-- If the console contains some information that could
help identifying the problem paste them in the next empty line -->
```
> tsc -b
node_modules/reddit-ts/src/RedditAPI.ts:185:25 - error TS2345: Argument of type 'Child | undefined' is not assignable to parameter of type 'Child'.
Type 'undefined' is not assignable to type 'Child'.
185 return map_search(data.children[0])
~~~~~~~~~~~~~~~~
Found 1 error.
```
| non_defect | bundles using reddit service fail to compile description newly generated bundles that use the reddit service currently fail to compile the odd thing is that it does not affect the sample bundle which still builds fine how to reproduce create a fresh nodecg install and install nodecg io current main or generate a new bundle that uses the reddit service using nodecg io generate bundle fails to compile with the error below error log compile failure if the console contains some information that could help identifying the problem paste them in the next empty line tsc b node modules reddit ts src redditapi ts error argument of type child undefined is not assignable to parameter of type child type undefined is not assignable to type child return map search data children found error | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.