Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
82,035
| 31,865,584,590
|
IssuesEvent
|
2023-09-15 13:56:22
|
dotCMS/core
|
https://api.github.com/repos/dotCMS/core
|
reopened
|
[Containers] : Fix UI Issues in the Portlet
|
Type : Defect Team : Lunik OKR : Core Features Priority : 3 Average
|
### Parent
#23140
### Problem Statement
The UI for containers portlets in dotCMS is not properly displaying in some cases. The portlet displays as a blank box, and the content stored in the container is not visible or accessible.
```[tasklist]
### Tasks
- [x] Prevent the jump when click the "Pre post loop" button, [see video](https://github.com/dotCMS/core/assets/751424/79743a40-e060-4ae2-93f9-3964405fb63c)
- [ ] Prevent the jump to edit the title, [see image](https://github.com/dotCMS/core/assets/751424/f9b3a354-82c5-4285-981f-a81c179c4de6)
- [x] Make the content type selector wider to avoid some of the ellipsis, [see image](https://github.com/dotCMS/core/assets/751424/1bd4a0b3-0d32-40c1-a927-b331147e7e35)
- [x] Fix the buttons size and position, [see image](https://github.com/dotCMS/core/assets/751424/4384333a-a613-4da8-a8b4-843e33dca1c4)
- [x] Make the add container button primary, [see image](https://github.com/dotCMS/core/assets/751424/bb3e5ba1-dd94-4595-bd18-bbcb1ae12662)
- [x] Fix "Add variables" buttons, [see image](https://github.com/dotCMS/core/assets/751424/29d66caf-e70e-4244-aa80-b575a0a4d209)
- [x] Fix scrolling tabs, [see images](https://github.com/dotCMS/core/assets/751424/84337a1a-8390-4ecd-a5d7-13486c9224ed)
- [x] Avoid the jump in the tabs from loading to loaded, [see video](https://github.com/dotCMS/core/assets/751424/735ec4d7-af10-4872-83c4-0246f51f4605)
```
### Acceptance Criteria
- All UI issues attached to this ticket should be fixed
- The content stored in the containers portlets should be visible and accessible
### External Links
N/A
### Assumptions & Initiation Needs
- The developer should have knowledge of dotCMS and the DotCMS UI
- The developer should have access to the existing codebase and the support ticket
### Quality Assurance Notes & Workarounds
- The QA team should test the changes to ensure that all UI issues are fixed and the content is visible and accessible
- As a workaround, the content in the containers portlets can be manually accessed by the developer in the meantime
|
1.0
|
[Containers] : Fix UI Issues in the Portlet - ### Parent
#23140
### Problem Statement
The UI for containers portlets in dotCMS is not properly displaying in some cases. The portlet displays as a blank box, and the content stored in the container is not visible or accessible.
```[tasklist]
### Tasks
- [x] Prevent the jump when click the "Pre post loop" button, [see video](https://github.com/dotCMS/core/assets/751424/79743a40-e060-4ae2-93f9-3964405fb63c)
- [ ] Prevent the jump to edit the title, [see image](https://github.com/dotCMS/core/assets/751424/f9b3a354-82c5-4285-981f-a81c179c4de6)
- [x] Make the content type selector wider to avoid some of the ellipsis, [see image](https://github.com/dotCMS/core/assets/751424/1bd4a0b3-0d32-40c1-a927-b331147e7e35)
- [x] Fix the buttons size and position, [see image](https://github.com/dotCMS/core/assets/751424/4384333a-a613-4da8-a8b4-843e33dca1c4)
- [x] Make the add container button primary, [see image](https://github.com/dotCMS/core/assets/751424/bb3e5ba1-dd94-4595-bd18-bbcb1ae12662)
- [x] Fix "Add variables" buttons, [see image](https://github.com/dotCMS/core/assets/751424/29d66caf-e70e-4244-aa80-b575a0a4d209)
- [x] Fix scrolling tabs, [see images](https://github.com/dotCMS/core/assets/751424/84337a1a-8390-4ecd-a5d7-13486c9224ed)
- [x] Avoid the jump in the tabs from loading to loaded, [see video](https://github.com/dotCMS/core/assets/751424/735ec4d7-af10-4872-83c4-0246f51f4605)
```
### Acceptance Criteria
- All UI issues attached to this ticket should be fixed
- The content stored in the containers portlets should be visible and accessible
### External Links
N/A
### Assumptions & Initiation Needs
- The developer should have knowledge of dotCMS and the DotCMS UI
- The developer should have access to the existing codebase and the support ticket
### Quality Assurance Notes & Workarounds
- The QA team should test the changes to ensure that all UI issues are fixed and the content is visible and accessible
- As a workaround, the content in the containers portlets can be manually accessed by the developer in the meantime
|
defect
|
fix ui issues in the portlet parent problem statement the ui for containers portlets in dotcms is not properly displaying in some cases the portlet displays as a blank box and the content stored in the container is not visible or accessible tasks prevent the jump when click the pre post loop button prevent the jump to edit the title make the content type selector wider to avoid some of the ellipsis fix the buttons size and position make the add container button primary fix add variables buttons fix scrolling tabs avoid the jump in the tabs from loading to loaded acceptance criteria all ui issues attached to this ticket should be fixed the content stored in the containers portlets should be visible and accessible external links n a assumptions initiation needs the developer should have knowledge of dotcms and the dotcms ui the developer should have access to the existing codebase and the support ticket quality assurance notes workarounds the qa team should test the changes to ensure that all ui issues are fixed and the content is visible and accessible as a workaround the content in the containers portlets can be manually accessed by the developer in the meantime
| 1
|
33,874
| 7,286,415,962
|
IssuesEvent
|
2018-02-23 09:40:03
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
Cannot display faces messages bound to iterated elements with p:dataTable
|
defect
|
When using a `p:dataTable` (PF 5.3) like this:
```
<p:dataTable values="..." var="..." id="current">
...
<h:messages for="current" />
```
No messages will be displayed for `current` although I do have faces messages attached to client ids (current:0, current:1...). However, messages for sub-components (current:0:property) can be displayed.
When rewriting the same example with `h:dataTable` or `ui:repeat`, all messages can be displayed.
More details on [stackoverflow](http://stackoverflow.com/questions/34681454/displaying-faces-messages-bound-to-elements-being-iterated-by-pdatatable).
If needed, I can provide a complete example.
|
1.0
|
Cannot display faces messages bound to iterated elements with p:dataTable - When using a `p:dataTable` (PF 5.3) like this:
```
<p:dataTable values="..." var="..." id="current">
...
<h:messages for="current" />
```
No messages will be displayed for `current` although I do have faces messages attached to client ids (current:0, current:1...). However, messages for sub-components (current:0:property) can be displayed.
When rewriting the same example with `h:dataTable` or `ui:repeat`, all messages can be displayed.
More details on [stackoverflow](http://stackoverflow.com/questions/34681454/displaying-faces-messages-bound-to-elements-being-iterated-by-pdatatable).
If needed, I can provide a complete example.
|
defect
|
cannot display faces messages bound to iterated elements with p datatable when using a p datatable pf like this no messages will be displayed for current although i do have faces messages attached to client ids current current however messages for sub components current property can be displayed when rewriting the same example with h datatable or ui repeat all messages can be displayed more details on if needed i can provide a complete example
| 1
|
22,430
| 7,175,256,553
|
IssuesEvent
|
2018-01-31 04:14:28
|
Pterodactyl/Panel
|
https://api.github.com/repos/Pterodactyl/Panel
|
closed
|
FTP upload/permissions
|
Beta/RC Build Bug
|
* Panel or Daemon: Both
* Version of Panel/Daemon: Panel (dev 0.7.0) Daemon v0.5.0-beta.5
* Server's OS: Ubuntu 16.04
* Your Computer's OS & Browser: Windows 10, fillezilla
## Add Details Below:
Created a CSGO server, used the SFTP settings to upload the files for metamod and sourcemod.
Booted the server, errors with permissions/crashed when player joins.
checked the files in /srv/daemon-data/SERVERID and noticed that instead of pterodactyl:ssh being the file owner (like the rest of the files) for what I just uploaded they were root:root
https://gyazo.com/91d50168e8e59e59bb856a3ab3ae1e4a
|
1.0
|
FTP upload/permissions - * Panel or Daemon: Both
* Version of Panel/Daemon: Panel (dev 0.7.0) Daemon v0.5.0-beta.5
* Server's OS: Ubuntu 16.04
* Your Computer's OS & Browser: Windows 10, fillezilla
## Add Details Below:
Created a CSGO server, used the SFTP settings to upload the files for metamod and sourcemod.
Booted the server, errors with permissions/crashed when player joins.
checked the files in /srv/daemon-data/SERVERID and noticed that instead of pterodactyl:ssh being the file owner (like the rest of the files) for what I just uploaded they were root:root
https://gyazo.com/91d50168e8e59e59bb856a3ab3ae1e4a
|
non_defect
|
ftp upload permissions panel or daemon both version of panel daemon panel dev daemon beta server s os ubuntu your computer s os browser windows fillezilla add details below created a csgo server used the sftp settings to upload the files for metamod and sourcemod booted the server errors with permissions crashed when player joins checked the files in srv daemon data serverid and noticed that instead of pterodactyl ssh being the file owner like the rest of the files for what i just uploaded they were root root
| 0
|
5,203
| 2,610,183,131
|
IssuesEvent
|
2015-02-26 18:58:19
|
chrsmith/quchuseban
|
https://api.github.com/repos/chrsmith/quchuseban
|
opened
|
提示怎么祛脸上的色斑
|
auto-migrated Priority-Medium Type-Defect
|
```
《摘要》
在日常生活中色斑越来越普遍,色斑虽然不像其他疾病那样��
�世迅猛,但带来的危害也是不容小视。据市场调查发现,一�
��女性平均每年花在去色斑上面的费用占到整个家庭收入的五
分之一,可见女性对色斑的忌惮。这也印证了脸上长色斑了��
�么办,什么产品去色斑效果好这一系列问题。盲目祛斑是现�
��普遍存在的现象,想要知道什么产品去色斑效果好,首先要
了解色斑的原因是什么,为什么会出现色斑?怎么祛脸上的色�
��,
《客户案例》
我和黄褐斑长了大概十年了,最早是星星点点的,没有��
�成块。<br>
在怀孕期间我没有采取任何措施,孩子出生后,我也在��
�待那些斑块快点消失,可是,我这一等就是四五个月,脸上�
��还是有好多斑,我也急了,好友说的话咋就不灵呢。估计是
因为之前我就有斑的缘故吧,始终找不到长斑的根源。我又��
�始了病急乱投医,这其中的过程我就不啰嗦了,想必长斑的�
��友们都有大同小异的遭遇。尝试过无数的祛斑方法、偏方后
我把最后的希望寄托在高科技的彩光上。在我历时5个月话费�
��6、7千元的一个周期后,结果还是很让我失望。<br>
我老公看到我天天被这些斑块折磨着,很是心疼,就上��
�找了很多关于祛斑的产品,最终他选定了「黛芙薇尔精华液�
��,他说不能再用那些激素产品了,植物产品可以试试,而且
还可以调理内分泌,对我身体恢复健康也有好处。就给我订��
�了两个周期的。一般老公的意见我都会听从,再说了我也想�
��点好。使用了一个周期后,黄褐斑就淡化了,特别是斑块变
小了。第二个周期见效更快,用完之后脸颊两块就都淡得接��
�肤色了,额头上也没有了。我老公看到效果不错,就又给我�
��购了一个周期的,使用完之后,天哪,我整张脸发生了天翻
地覆的变化啊,从一个斑脸婆就变成了一个水嫩白净的女人��
�应该说是漂亮的母亲。呵呵,这一点都不夸张。老公现在看�
��的眼神都总是含情脉脉的,还老说我一点不像生过孩子的样
子。
阅读了怎么祛脸上的色斑,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
怎么祛脸上的色斑,同时为您分享祛斑小方法
山药:山药的黏质液是合成雌激素的基础物质,对工作和学��
�较大压力的女性具有很好的健身功效。山药所含的精氨酸还�
��健脾补肺的功效。山药含有一种可溶性纤维,有助于控制饮
食,改善消化系统。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:00
|
1.0
|
提示怎么祛脸上的色斑 - ```
《摘要》
在日常生活中色斑越来越普遍,色斑虽然不像其他疾病那样��
�世迅猛,但带来的危害也是不容小视。据市场调查发现,一�
��女性平均每年花在去色斑上面的费用占到整个家庭收入的五
分之一,可见女性对色斑的忌惮。这也印证了脸上长色斑了��
�么办,什么产品去色斑效果好这一系列问题。盲目祛斑是现�
��普遍存在的现象,想要知道什么产品去色斑效果好,首先要
了解色斑的原因是什么,为什么会出现色斑?怎么祛脸上的色�
��,
《客户案例》
我和黄褐斑长了大概十年了,最早是星星点点的,没有��
�成块。<br>
在怀孕期间我没有采取任何措施,孩子出生后,我也在��
�待那些斑块快点消失,可是,我这一等就是四五个月,脸上�
��还是有好多斑,我也急了,好友说的话咋就不灵呢。估计是
因为之前我就有斑的缘故吧,始终找不到长斑的根源。我又��
�始了病急乱投医,这其中的过程我就不啰嗦了,想必长斑的�
��友们都有大同小异的遭遇。尝试过无数的祛斑方法、偏方后
我把最后的希望寄托在高科技的彩光上。在我历时5个月话费�
��6、7千元的一个周期后,结果还是很让我失望。<br>
我老公看到我天天被这些斑块折磨着,很是心疼,就上��
�找了很多关于祛斑的产品,最终他选定了「黛芙薇尔精华液�
��,他说不能再用那些激素产品了,植物产品可以试试,而且
还可以调理内分泌,对我身体恢复健康也有好处。就给我订��
�了两个周期的。一般老公的意见我都会听从,再说了我也想�
��点好。使用了一个周期后,黄褐斑就淡化了,特别是斑块变
小了。第二个周期见效更快,用完之后脸颊两块就都淡得接��
�肤色了,额头上也没有了。我老公看到效果不错,就又给我�
��购了一个周期的,使用完之后,天哪,我整张脸发生了天翻
地覆的变化啊,从一个斑脸婆就变成了一个水嫩白净的女人��
�应该说是漂亮的母亲。呵呵,这一点都不夸张。老公现在看�
��的眼神都总是含情脉脉的,还老说我一点不像生过孩子的样
子。
阅读了怎么祛脸上的色斑,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
怎么祛脸上的色斑,同时为您分享祛斑小方法
山药:山药的黏质液是合成雌激素的基础物质,对工作和学��
�较大压力的女性具有很好的健身功效。山药所含的精氨酸还�
��健脾补肺的功效。山药含有一种可溶性纤维,有助于控制饮
食,改善消化系统。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:00
|
defect
|
提示怎么祛脸上的色斑 《摘要》 在日常生活中色斑越来越普遍,色斑虽然不像其他疾病那样�� �世迅猛,但带来的危害也是不容小视。据市场调查发现,一� ��女性平均每年花在去色斑上面的费用占到整个家庭收入的五 分之一,可见女性对色斑的忌惮。这也印证了脸上长色斑了�� �么办,什么产品去色斑效果好这一系列问题。盲目祛斑是现� ��普遍存在的现象,想要知道什么产品去色斑效果好,首先要 了解色斑的原因是什么,为什么会出现色斑 怎么祛脸上的色� ��, 《客户案例》 我和黄褐斑长了大概十年了,最早是星星点点的,没有�� �成块。 在怀孕期间我没有采取任何措施,孩子出生后,我也在�� �待那些斑块快点消失,可是,我这一等就是四五个月,脸上� ��还是有好多斑,我也急了,好友说的话咋就不灵呢。估计是 因为之前我就有斑的缘故吧,始终找不到长斑的根源。我又�� �始了病急乱投医,这其中的过程我就不啰嗦了,想必长斑的� ��友们都有大同小异的遭遇。尝试过无数的祛斑方法、偏方后 我把最后的希望寄托在高科技的彩光上。 � �� 、 ,结果还是很让我失望。 我老公看到我天天被这些斑块折磨着,很是心疼,就上�� �找了很多关于祛斑的产品,最终他选定了「黛芙薇尔精华液� ��,他说不能再用那些激素产品了,植物产品可以试试,而且 还可以调理内分泌,对我身体恢复健康也有好处。就给我订�� �了两个周期的。一般老公的意见我都会听从,再说了我也想� ��点好。使用了一个周期后,黄褐斑就淡化了,特别是斑块变 小了。第二个周期见效更快,用完之后脸颊两块就都淡得接�� �肤色了,额头上也没有了。我老公看到效果不错,就又给我� ��购了一个周期的,使用完之后,天哪,我整张脸发生了天翻 地覆的变化啊,从一个斑脸婆就变成了一个水嫩白净的女人�� �应该说是漂亮的母亲。呵呵,这一点都不夸张。老公现在看� ��的眼神都总是含情脉脉的,还老说我一点不像生过孩子的样 子。 阅读了怎么祛脸上的色斑,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 怎么祛脸上的色斑,同时为您分享祛斑小方法 山药:山药的黏质液是合成雌激素的基础物质,对工作和学�� �较大压力的女性具有很好的健身功效。山药所含的精氨酸还� ��健脾补肺的功效。山药含有一种可溶性纤维,有助于控制饮 食,改善消化系统。 original issue reported on code google com by additive gmail com on jul at
| 1
|
7,321
| 2,610,363,360
|
IssuesEvent
|
2015-02-26 19:57:26
|
chrsmith/scribefire-chrome
|
https://api.github.com/repos/chrsmith/scribefire-chrome
|
closed
|
Firefox Portable with Scribefire Next cannot add a blog
|
auto-migrated Priority-Medium Type-Defect
|
```
What's the problem?
I installed the ScribeFire Next add-on to my Firefox Portable and I tried to
add a blog. I filled in my blog address and clicked on the securely authorize
ScribeFire button. It takes me to a wordpress.com login page to authorize
ScribeFire. I entered my login details and clicked "authorize" then ScribeFire
give me the error message:
Well, this is embarrassing...
ScribeFire couldn't get the information it needed about your blog. Helpfully,
your blog returned this message: Bad login/pass combination.
What browser are you using?
Firefox Portable version 15 from PortableApps.com running on Windows XP.
What version of ScribeFire are you running?
ScribeFire Next 4.0
```
-----
Original issue reported on code.google.com by `talk...@gmail.com` on 2 Sep 2012 at 5:18
|
1.0
|
Firefox Portable with Scribefire Next cannot add a blog - ```
What's the problem?
I installed the ScribeFire Next add-on to my Firefox Portable and I tried to
add a blog. I filled in my blog address and clicked on the securely authorize
ScribeFire button. It takes me to a wordpress.com login page to authorize
ScribeFire. I entered my login details and clicked "authorize" then ScribeFire
give me the error message:
Well, this is embarrassing...
ScribeFire couldn't get the information it needed about your blog. Helpfully,
your blog returned this message: Bad login/pass combination.
What browser are you using?
Firefox Portable version 15 from PortableApps.com running on Windows XP.
What version of ScribeFire are you running?
ScribeFire Next 4.0
```
-----
Original issue reported on code.google.com by `talk...@gmail.com` on 2 Sep 2012 at 5:18
|
defect
|
firefox portable with scribefire next cannot add a blog what s the problem i installed the scribefire next add on to my firefox portable and i tried to add a blog i filled in my blog address and clicked on the securely authorize scribefire button it takes me to a wordpress com login page to authorize scribefire i entered my login details and clicked authorize then scribefire give me the error message well this is embarrassing scribefire couldn t get the information it needed about your blog helpfully your blog returned this message bad login pass combination what browser are you using firefox portable version from portableapps com running on windows xp what version of scribefire are you running scribefire next original issue reported on code google com by talk gmail com on sep at
| 1
|
56,711
| 15,308,029,370
|
IssuesEvent
|
2021-02-24 21:48:12
|
ontop/ontop
|
https://api.github.com/repos/ontop/ontop
|
closed
|
use of R2RML predicateMaps throws error
|
type: defect
|
My R2RML mapping works as long as I do not include predicateMaps.
When adding the second predicateObjectMap with a predicateMap I get the error below.
```
<ActivityMapping>
a rr:TriplesMap;
rr:logicalTable [ rr:tableName "activity" ];
rr:subjectMap [
rr:template "http://data.kbodata.be/organisation/{EntityNumber}#id";
rr:class org:FormalOrganization;
rr:termType rr:IRI
];
rr:predicateObjectMap [
rr:predicate rov:orgActivity;
rr:objectMap [
rr:template "http://id.fedstats.be/nace{NaceVersion}/{NaceCode}#id";
rr:termType rr:IRI
]
];
rr:predicateObjectMap [
rr:predicateMap
[
rr:template "http://data.kbodata.be/def#{Classification}";
rr:termType rr:IRI
];
rr:objectMap [
rr:template "http://id.fedstats.be/nace{NaceVersion}/{NaceCode}#id";
rr:termType rr:IRI
]
];
.
```
Error
```
Error during r2rml import.
java.lang.RuntimeException: java.lang.NullPointerException
at it.unibz.krdb.obda.r2rml.R2RMLManager.getMappings(R2RMLManager.java:131)
at it.unibz.krdb.obda.r2rml.R2RMLReader.readMappings(R2RMLReader.java:110)
at org.semanticweb.ontop.protege.gui.action.R2RMLImportAction.actionPerformed(R2RMLImportAction.java:95)
at javax.swing.AbstractButton.fireActionPerformed(AbstractButton.java:2022)
at javax.swing.AbstractButton$Handler.actionPerformed(AbstractButton.java:2346)
at javax.swing.DefaultButtonModel.fireActionPerformed(DefaultButtonModel.java:402)
at javax.swing.DefaultButtonModel.setPressed(DefaultButtonModel.java:259)
at javax.swing.AbstractButton.doClick(AbstractButton.java:376)
at com.apple.laf.ScreenMenuItem.actionPerformed(ScreenMenuItem.java:125)
at java.awt.MenuItem.processActionEvent(MenuItem.java:669)
at java.awt.MenuItem.processEvent(MenuItem.java:628)
at java.awt.MenuComponent.dispatchEventImpl(MenuComponent.java:351)
at java.awt.MenuComponent.dispatchEvent(MenuComponent.java:339)
at java.awt.EventQueue.dispatchEventImpl(EventQueue.java:754)
at java.awt.EventQueue.access$500(EventQueue.java:97)
at java.awt.EventQueue$3.run(EventQueue.java:702)
at java.awt.EventQueue$3.run(EventQueue.java:696)
at java.security.AccessController.doPrivileged(Native Method)
at java.security.ProtectionDomain$1.doIntersectionPrivilege(ProtectionDomain.java:75)
at java.security.ProtectionDomain$1.doIntersectionPrivilege(ProtectionDomain.java:86)
at java.awt.EventQueue$4.run(EventQueue.java:724)
at java.awt.EventQueue$4.run(EventQueue.java:722)
at java.security.AccessController.doPrivileged(Native Method)
at java.security.ProtectionDomain$1.doIntersectionPrivilege(ProtectionDomain.java:75)
at java.awt.EventQueue.dispatchEvent(EventQueue.java:721)
at java.awt.EventDispatchThread.pumpOneEventForFilters(EventDispatchThread.java:201)
at java.awt.EventDispatchThread.pumpEventsForFilter(EventDispatchThread.java:116)
at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThread.java:105)
at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:101)
at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:93)
at java.awt.EventDispatchThread.run(EventDispatchThread.java:82)
Caused by: java.lang.NullPointerException
at it.unibz.krdb.obda.model.impl.PredicateImpl.<init>(PredicateImpl.java:36)
at it.unibz.krdb.obda.model.impl.OBDADataFactoryImpl.getPredicate(OBDADataFactoryImpl.java:92)
at it.unibz.krdb.obda.model.impl.OBDADataFactoryImpl.getPredicate(OBDADataFactoryImpl.java:37)
at it.unibz.krdb.obda.r2rml.R2RMLParser.getBodyPredicates(R2RMLParser.java:221)
at it.unibz.krdb.obda.r2rml.R2RMLManager.getMappingTripleAtoms(R2RMLManager.java:254)
at it.unibz.krdb.obda.r2rml.R2RMLManager.getMapping(R2RMLManager.java:144)
at it.unibz.krdb.obda.r2rml.R2RMLManager.getMappings(R2RMLManager.java:117)
... 30 more
```
|
1.0
|
use of R2RML predicateMaps throws error - My R2RML mapping works as long as I do not include predicateMaps.
When adding the second predicateObjectMap with a predicateMap I get the error below.
```
<ActivityMapping>
a rr:TriplesMap;
rr:logicalTable [ rr:tableName "activity" ];
rr:subjectMap [
rr:template "http://data.kbodata.be/organisation/{EntityNumber}#id";
rr:class org:FormalOrganization;
rr:termType rr:IRI
];
rr:predicateObjectMap [
rr:predicate rov:orgActivity;
rr:objectMap [
rr:template "http://id.fedstats.be/nace{NaceVersion}/{NaceCode}#id";
rr:termType rr:IRI
]
];
rr:predicateObjectMap [
rr:predicateMap
[
rr:template "http://data.kbodata.be/def#{Classification}";
rr:termType rr:IRI
];
rr:objectMap [
rr:template "http://id.fedstats.be/nace{NaceVersion}/{NaceCode}#id";
rr:termType rr:IRI
]
];
.
```
Error
```
Error during r2rml import.
java.lang.RuntimeException: java.lang.NullPointerException
at it.unibz.krdb.obda.r2rml.R2RMLManager.getMappings(R2RMLManager.java:131)
at it.unibz.krdb.obda.r2rml.R2RMLReader.readMappings(R2RMLReader.java:110)
at org.semanticweb.ontop.protege.gui.action.R2RMLImportAction.actionPerformed(R2RMLImportAction.java:95)
at javax.swing.AbstractButton.fireActionPerformed(AbstractButton.java:2022)
at javax.swing.AbstractButton$Handler.actionPerformed(AbstractButton.java:2346)
at javax.swing.DefaultButtonModel.fireActionPerformed(DefaultButtonModel.java:402)
at javax.swing.DefaultButtonModel.setPressed(DefaultButtonModel.java:259)
at javax.swing.AbstractButton.doClick(AbstractButton.java:376)
at com.apple.laf.ScreenMenuItem.actionPerformed(ScreenMenuItem.java:125)
at java.awt.MenuItem.processActionEvent(MenuItem.java:669)
at java.awt.MenuItem.processEvent(MenuItem.java:628)
at java.awt.MenuComponent.dispatchEventImpl(MenuComponent.java:351)
at java.awt.MenuComponent.dispatchEvent(MenuComponent.java:339)
at java.awt.EventQueue.dispatchEventImpl(EventQueue.java:754)
at java.awt.EventQueue.access$500(EventQueue.java:97)
at java.awt.EventQueue$3.run(EventQueue.java:702)
at java.awt.EventQueue$3.run(EventQueue.java:696)
at java.security.AccessController.doPrivileged(Native Method)
at java.security.ProtectionDomain$1.doIntersectionPrivilege(ProtectionDomain.java:75)
at java.security.ProtectionDomain$1.doIntersectionPrivilege(ProtectionDomain.java:86)
at java.awt.EventQueue$4.run(EventQueue.java:724)
at java.awt.EventQueue$4.run(EventQueue.java:722)
at java.security.AccessController.doPrivileged(Native Method)
at java.security.ProtectionDomain$1.doIntersectionPrivilege(ProtectionDomain.java:75)
at java.awt.EventQueue.dispatchEvent(EventQueue.java:721)
at java.awt.EventDispatchThread.pumpOneEventForFilters(EventDispatchThread.java:201)
at java.awt.EventDispatchThread.pumpEventsForFilter(EventDispatchThread.java:116)
at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThread.java:105)
at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:101)
at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:93)
at java.awt.EventDispatchThread.run(EventDispatchThread.java:82)
Caused by: java.lang.NullPointerException
at it.unibz.krdb.obda.model.impl.PredicateImpl.<init>(PredicateImpl.java:36)
at it.unibz.krdb.obda.model.impl.OBDADataFactoryImpl.getPredicate(OBDADataFactoryImpl.java:92)
at it.unibz.krdb.obda.model.impl.OBDADataFactoryImpl.getPredicate(OBDADataFactoryImpl.java:37)
at it.unibz.krdb.obda.r2rml.R2RMLParser.getBodyPredicates(R2RMLParser.java:221)
at it.unibz.krdb.obda.r2rml.R2RMLManager.getMappingTripleAtoms(R2RMLManager.java:254)
at it.unibz.krdb.obda.r2rml.R2RMLManager.getMapping(R2RMLManager.java:144)
at it.unibz.krdb.obda.r2rml.R2RMLManager.getMappings(R2RMLManager.java:117)
... 30 more
```
|
defect
|
use of predicatemaps throws error my mapping works as long as i do not include predicatemaps when adding the second predicateobjectmap with a predicatemap i get the error below a rr triplesmap rr logicaltable rr subjectmap rr template rr class org formalorganization rr termtype rr iri rr predicateobjectmap rr predicate rov orgactivity rr objectmap rr template rr termtype rr iri rr predicateobjectmap rr predicatemap rr template rr termtype rr iri rr objectmap rr template rr termtype rr iri error error during import java lang runtimeexception java lang nullpointerexception at it unibz krdb obda getmappings java at it unibz krdb obda readmappings java at org semanticweb ontop protege gui action actionperformed java at javax swing abstractbutton fireactionperformed abstractbutton java at javax swing abstractbutton handler actionperformed abstractbutton java at javax swing defaultbuttonmodel fireactionperformed defaultbuttonmodel java at javax swing defaultbuttonmodel setpressed defaultbuttonmodel java at javax swing abstractbutton doclick abstractbutton java at com apple laf screenmenuitem actionperformed screenmenuitem java at java awt menuitem processactionevent menuitem java at java awt menuitem processevent menuitem java at java awt menucomponent dispatcheventimpl menucomponent java at java awt menucomponent dispatchevent menucomponent java at java awt eventqueue dispatcheventimpl eventqueue java at java awt eventqueue access eventqueue java at java awt eventqueue run eventqueue java at java awt eventqueue run eventqueue java at java security accesscontroller doprivileged native method at java security protectiondomain dointersectionprivilege protectiondomain java at java security protectiondomain dointersectionprivilege protectiondomain java at java awt eventqueue run eventqueue java at java awt eventqueue run eventqueue java at java security accesscontroller doprivileged native method at java security protectiondomain dointersectionprivilege protectiondomain java at java awt eventqueue dispatchevent eventqueue java at java awt eventdispatchthread pumponeeventforfilters eventdispatchthread java at java awt eventdispatchthread pumpeventsforfilter eventdispatchthread java at java awt eventdispatchthread pumpeventsforhierarchy eventdispatchthread java at java awt eventdispatchthread pumpevents eventdispatchthread java at java awt eventdispatchthread pumpevents eventdispatchthread java at java awt eventdispatchthread run eventdispatchthread java caused by java lang nullpointerexception at it unibz krdb obda model impl predicateimpl predicateimpl java at it unibz krdb obda model impl obdadatafactoryimpl getpredicate obdadatafactoryimpl java at it unibz krdb obda model impl obdadatafactoryimpl getpredicate obdadatafactoryimpl java at it unibz krdb obda getbodypredicates java at it unibz krdb obda getmappingtripleatoms java at it unibz krdb obda getmapping java at it unibz krdb obda getmappings java more
| 1
|
53,156
| 6,691,286,309
|
IssuesEvent
|
2017-10-09 12:37:24
|
desktop/desktop
|
https://api.github.com/repos/desktop/desktop
|
closed
|
Contrast - Red & Green
|
needs-design-input
|
<!--
Have you read GitHub Desktop's Code of Conduct? By filing an Issue, you are
expected to comply with it, including treating everyone with respect:
https://github.com/desktop/desktop/blob/master/CODE_OF_CONDUCT.md
-->
<!--
Are you encountering an issue where the “Minimize” tooltip stays visible
when you click the minimize button in the window? If so, that is an issue
with Electron, the framework the app uses. Please subscribe to the issue
at this link for updates on the issue:
https://github.com/electron/electron/issues/9943
-->
<!--
Please summarize the issue in the title, and then use the template below to
fill out the details so we can reproduce the issue on our end.
-->
### Description
The desktop client uses very light colored red and green to indicated deletes and additions, I'm not sure if I've always been color blind, but trying to see the difference between the two colors is nearly impossible for me. I really like the client, it's simple to use and it is helping me learn source control, but this color thing is becoming a real hurdle to get passed. Would it be possible to get settings to adjust colors or contrast, or to have a dark mode so that the reds and greens stand out more? Thank you so much for all you do in providing this website and control and collaboration.
### Version
<!--
What version of GitHub Desktop are you running? This is displayed under the
`About GitHub Desktop` menu item. If you are running from source, include
the commit by running `git rev-parse HEAD` from your local repository.
-->
**GitHub Desktop version:** 1.0.3
<!--
The operating system you are running on may also help with reproducing the
issue:
- If you are on macOS, launch `About This Mac` and write down the OS version
listed.
- If you are on Windows, open `Command Prompt` and attach the output of this
command: `cmd /c ver`
-->
**OS version:** Windows 10
### Steps to Reproduce
1. open repo, make adds and deletes, hard to see if colorblind
<!--
If the issue involves a specific public repository, including the information
about that repository will make it is easier to recreate the issue.
If you think screenshots or a GIF recording will help demonstrate the issue
better, feel free to add them here.
-->
**Expected behavior:** able to see difference between green and red on screen
**Actual behavior:** everything looks white
**Reproduces how often:** always, I think I'm colorblind
### Logs
<!--
There may be some relevant information in log files generated by GitHub
Desktop:
- If you are on macOS, attach the most recent log file from:
`~/Library/Application Support/GitHub Desktop/logs/*.desktop.production.log`
- If you are on Windows, attach the most recent log file from:
`%APPDATA%\GitHub Desktop\logs\*.desktop.production.log`
The log files are organized by date, so see if anything was generated for
today's date.
-->
#### Additional Information
<!--
Any additional information, configuration or data that might be necessary to
reproduce the issue.
If you are dealing with a performance issue or regression, attaching a
[Timeline profile](https://github.com/desktop/desktop/blob/master/docs/contributing/timeline-profile.md)
of the task will help the developers understand the runtime behavior of the
application on your machine.
-->
|
1.0
|
Contrast - Red & Green - <!--
Have you read GitHub Desktop's Code of Conduct? By filing an Issue, you are
expected to comply with it, including treating everyone with respect:
https://github.com/desktop/desktop/blob/master/CODE_OF_CONDUCT.md
-->
<!--
Are you encountering an issue where the “Minimize” tooltip stays visible
when you click the minimize button in the window? If so, that is an issue
with Electron, the framework the app uses. Please subscribe to the issue
at this link for updates on the issue:
https://github.com/electron/electron/issues/9943
-->
<!--
Please summarize the issue in the title, and then use the template below to
fill out the details so we can reproduce the issue on our end.
-->
### Description
The desktop client uses very light colored red and green to indicated deletes and additions, I'm not sure if I've always been color blind, but trying to see the difference between the two colors is nearly impossible for me. I really like the client, it's simple to use and it is helping me learn source control, but this color thing is becoming a real hurdle to get passed. Would it be possible to get settings to adjust colors or contrast, or to have a dark mode so that the reds and greens stand out more? Thank you so much for all you do in providing this website and control and collaboration.
### Version
<!--
What version of GitHub Desktop are you running? This is displayed under the
`About GitHub Desktop` menu item. If you are running from source, include
the commit by running `git rev-parse HEAD` from your local repository.
-->
**GitHub Desktop version:** 1.0.3
<!--
The operating system you are running on may also help with reproducing the
issue:
- If you are on macOS, launch `About This Mac` and write down the OS version
listed.
- If you are on Windows, open `Command Prompt` and attach the output of this
command: `cmd /c ver`
-->
**OS version:** Windows 10
### Steps to Reproduce
1. open repo, make adds and deletes, hard to see if colorblind
<!--
If the issue involves a specific public repository, including the information
about that repository will make it is easier to recreate the issue.
If you think screenshots or a GIF recording will help demonstrate the issue
better, feel free to add them here.
-->
**Expected behavior:** able to see difference between green and red on screen
**Actual behavior:** everything looks white
**Reproduces how often:** always, I think I'm colorblind
### Logs
<!--
There may be some relevant information in log files generated by GitHub
Desktop:
- If you are on macOS, attach the most recent log file from:
`~/Library/Application Support/GitHub Desktop/logs/*.desktop.production.log`
- If you are on Windows, attach the most recent log file from:
`%APPDATA%\GitHub Desktop\logs\*.desktop.production.log`
The log files are organized by date, so see if anything was generated for
today's date.
-->
#### Additional Information
<!--
Any additional information, configuration or data that might be necessary to
reproduce the issue.
If you are dealing with a performance issue or regression, attaching a
[Timeline profile](https://github.com/desktop/desktop/blob/master/docs/contributing/timeline-profile.md)
of the task will help the developers understand the runtime behavior of the
application on your machine.
-->
|
non_defect
|
contrast red green have you read github desktop s code of conduct by filing an issue you are expected to comply with it including treating everyone with respect are you encountering an issue where the “minimize” tooltip stays visible when you click the minimize button in the window if so that is an issue with electron the framework the app uses please subscribe to the issue at this link for updates on the issue please summarize the issue in the title and then use the template below to fill out the details so we can reproduce the issue on our end description the desktop client uses very light colored red and green to indicated deletes and additions i m not sure if i ve always been color blind but trying to see the difference between the two colors is nearly impossible for me i really like the client it s simple to use and it is helping me learn source control but this color thing is becoming a real hurdle to get passed would it be possible to get settings to adjust colors or contrast or to have a dark mode so that the reds and greens stand out more thank you so much for all you do in providing this website and control and collaboration version what version of github desktop are you running this is displayed under the about github desktop menu item if you are running from source include the commit by running git rev parse head from your local repository github desktop version the operating system you are running on may also help with reproducing the issue if you are on macos launch about this mac and write down the os version listed if you are on windows open command prompt and attach the output of this command cmd c ver os version windows steps to reproduce open repo make adds and deletes hard to see if colorblind if the issue involves a specific public repository including the information about that repository will make it is easier to recreate the issue if you think screenshots or a gif recording will help demonstrate the issue better feel free to add them here expected behavior able to see difference between green and red on screen actual behavior everything looks white reproduces how often always i think i m colorblind logs there may be some relevant information in log files generated by github desktop if you are on macos attach the most recent log file from library application support github desktop logs desktop production log if you are on windows attach the most recent log file from appdata github desktop logs desktop production log the log files are organized by date so see if anything was generated for today s date additional information any additional information configuration or data that might be necessary to reproduce the issue if you are dealing with a performance issue or regression attaching a of the task will help the developers understand the runtime behavior of the application on your machine
| 0
|
78,034
| 27,284,991,567
|
IssuesEvent
|
2023-02-23 12:53:54
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Translator duplicates comment only content when retaining comments
|
T: Defect P: Medium E: All Editions C: Translator
|
When translating comment only content, such as e.g.
```sql
/* a */
-- b
```
Then the translator duplicates it:
<img width="768" alt="image" src="https://user-images.githubusercontent.com/734593/220898060-5a4f24c2-0568-4ffa-9bbb-467215e56c90.png">
The problem doesn't appear when there's an actual query in the input:
<img width="776" alt="image" src="https://user-images.githubusercontent.com/734593/220898144-b74626e5-bf2b-426c-9a41-fe43222286e5.png">
|
1.0
|
Translator duplicates comment only content when retaining comments - When translating comment only content, such as e.g.
```sql
/* a */
-- b
```
Then the translator duplicates it:
<img width="768" alt="image" src="https://user-images.githubusercontent.com/734593/220898060-5a4f24c2-0568-4ffa-9bbb-467215e56c90.png">
The problem doesn't appear when there's an actual query in the input:
<img width="776" alt="image" src="https://user-images.githubusercontent.com/734593/220898144-b74626e5-bf2b-426c-9a41-fe43222286e5.png">
|
defect
|
translator duplicates comment only content when retaining comments when translating comment only content such as e g sql a b then the translator duplicates it img width alt image src the problem doesn t appear when there s an actual query in the input img width alt image src
| 1
|
51,218
| 13,207,396,526
|
IssuesEvent
|
2020-08-14 22:56:50
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
libPyROOT on a mac not loading in python properly (Trac #57)
|
Incomplete Migration Migrated from Trac cmake defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/57">https://code.icecube.wisc.edu/projects/icecube/ticket/57</a>, reported by blaufussand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-09T22:34:06",
"_ts": "1194647646000000",
"description": "trying to import libPyROOT on a mac gives issues:\n\n(root-v5.10.00)\n\nblaufuss@teufel[~/..work/V01-11-02/build]% ./env-shell.sh\nblaufuss@teufel[~/..work/V01-11-02/build](I3)% python\nPython 2.3.5 (#1, Jun 9 2007, 18:47:11)\n[GCC 4.0.1 (Apple Computer, Inc. build 5363)] on darwin\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\n>>> from libPyROOT import *\nTraceback (most recent call last):\n File \"<stdin>\", line 1, in ?\nImportError: Inappropriate file type for dynamic loading\n>>>\n\nThis seems to work fine on Linux*\n\nThis thread seems to indicate that this is a python/root linking issue:\nhttp://root.cern.ch/phpBB2/viewtopic.php?p=13401&sid=17c3a761d528ae2926c35f1f6b0afc4d\n\n--Erik",
"reporter": "blaufuss",
"cc": "",
"resolution": "duplicate",
"time": "2007-06-11T13:53:13",
"component": "cmake",
"summary": "libPyROOT on a mac not loading in python properly",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
libPyROOT on a mac not loading in python properly (Trac #57) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/57">https://code.icecube.wisc.edu/projects/icecube/ticket/57</a>, reported by blaufussand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-09T22:34:06",
"_ts": "1194647646000000",
"description": "trying to import libPyROOT on a mac gives issues:\n\n(root-v5.10.00)\n\nblaufuss@teufel[~/..work/V01-11-02/build]% ./env-shell.sh\nblaufuss@teufel[~/..work/V01-11-02/build](I3)% python\nPython 2.3.5 (#1, Jun 9 2007, 18:47:11)\n[GCC 4.0.1 (Apple Computer, Inc. build 5363)] on darwin\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\n>>> from libPyROOT import *\nTraceback (most recent call last):\n File \"<stdin>\", line 1, in ?\nImportError: Inappropriate file type for dynamic loading\n>>>\n\nThis seems to work fine on Linux*\n\nThis thread seems to indicate that this is a python/root linking issue:\nhttp://root.cern.ch/phpBB2/viewtopic.php?p=13401&sid=17c3a761d528ae2926c35f1f6b0afc4d\n\n--Erik",
"reporter": "blaufuss",
"cc": "",
"resolution": "duplicate",
"time": "2007-06-11T13:53:13",
"component": "cmake",
"summary": "libPyROOT on a mac not loading in python properly",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
|
defect
|
libpyroot on a mac not loading in python properly trac migrated from json status closed changetime ts description trying to import libpyroot on a mac gives issues n n root n nblaufuss teufel env shell sh nblaufuss teufel python npython jun n on darwin ntype help copyright credits or license for more information n from libpyroot import ntraceback most recent call last n file line in nimporterror inappropriate file type for dynamic loading n n nthis seems to work fine on linux n nthis thread seems to indicate that this is a python root linking issue n reporter blaufuss cc resolution duplicate time component cmake summary libpyroot on a mac not loading in python properly priority normal keywords milestone owner troy type defect
| 1
|
2,387
| 2,607,899,878
|
IssuesEvent
|
2015-02-26 00:13:07
|
chrsmithdemos/zen-coding
|
https://api.github.com/repos/chrsmithdemos/zen-coding
|
closed
|
add '<' operator
|
auto-migrated Priority-Medium Type-Defect
|
```
using '<' operator parentNode :)
```
-----
Original issue reported on code.google.com by `tangoboy...@gmail.com` on 22 Dec 2009 at 4:41
|
1.0
|
add '<' operator - ```
using '<' operator parentNode :)
```
-----
Original issue reported on code.google.com by `tangoboy...@gmail.com` on 22 Dec 2009 at 4:41
|
defect
|
add operator using operator parentnode original issue reported on code google com by tangoboy gmail com on dec at
| 1
|
130,007
| 12,421,897,879
|
IssuesEvent
|
2020-05-23 19:09:57
|
mpls-landlord-db/landlord-lookup
|
https://api.github.com/repos/mpls-landlord-db/landlord-lookup
|
opened
|
Create glossary for relevant columns in `active_rental_licenses` data
|
documentation feature
|
The meanings of the columns `category` and `milestone` are unclear. When users lookup information on the properties they rent, they should be able to understand the data.
|
1.0
|
Create glossary for relevant columns in `active_rental_licenses` data - The meanings of the columns `category` and `milestone` are unclear. When users lookup information on the properties they rent, they should be able to understand the data.
|
non_defect
|
create glossary for relevant columns in active rental licenses data the meanings of the columns category and milestone are unclear when users lookup information on the properties they rent they should be able to understand the data
| 0
|
64,011
| 6,890,552,142
|
IssuesEvent
|
2017-11-22 14:23:42
|
Shadowss/TravianZ
|
https://api.github.com/repos/Shadowss/TravianZ
|
closed
|
Account activation issues
|
bug needs testing
|
During the registration of the account, everything goes well. The letter comes to the post office. And then there are problems with activation of both the link and the activation code. Records that the wrong password or account has been deleted. There is everything in the database. The new user is in `s1_activate`.

This is most likely due to encryption of the password. Because without an activated account I can not log into my account. An incorrect password is displayed.

|
1.0
|
Account activation issues - During the registration of the account, everything goes well. The letter comes to the post office. And then there are problems with activation of both the link and the activation code. Records that the wrong password or account has been deleted. There is everything in the database. The new user is in `s1_activate`.

This is most likely due to encryption of the password. Because without an activated account I can not log into my account. An incorrect password is displayed.

|
non_defect
|
account activation issues during the registration of the account everything goes well the letter comes to the post office and then there are problems with activation of both the link and the activation code records that the wrong password or account has been deleted there is everything in the database the new user is in activate this is most likely due to encryption of the password because without an activated account i can not log into my account an incorrect password is displayed
| 0
|
51,380
| 13,207,459,348
|
IssuesEvent
|
2020-08-14 23:11:03
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
TriggerType and TriggerMode/Situation (the DOMLaunch enums) not set (Trac #311)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/311">https://code.icecube.wisc.edu/projects/icecube/ticket/311</a>, reported by icecubeand owned by olivas, sflis</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-22T18:26:26",
"_ts": "1416680786877026",
"description": "The enums TriggerType and TriggerMode from DOMLaunch.h are not set in current MC. This would be very useful to understand trigger behaviour (like HLC/SLC/...).\n\nShould go in the old and upcoming new DOMsimulator, maybe?\n",
"reporter": "icecube",
"cc": "",
"resolution": "wontfix",
"time": "2011-09-27T16:18:15",
"component": "combo simulation",
"summary": "TriggerType and TriggerMode/Situation (the DOMLaunch enums) not set",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "olivas, sflis",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
TriggerType and TriggerMode/Situation (the DOMLaunch enums) not set (Trac #311) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/311">https://code.icecube.wisc.edu/projects/icecube/ticket/311</a>, reported by icecubeand owned by olivas, sflis</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-22T18:26:26",
"_ts": "1416680786877026",
"description": "The enums TriggerType and TriggerMode from DOMLaunch.h are not set in current MC. This would be very useful to understand trigger behaviour (like HLC/SLC/...).\n\nShould go in the old and upcoming new DOMsimulator, maybe?\n",
"reporter": "icecube",
"cc": "",
"resolution": "wontfix",
"time": "2011-09-27T16:18:15",
"component": "combo simulation",
"summary": "TriggerType and TriggerMode/Situation (the DOMLaunch enums) not set",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "olivas, sflis",
"type": "defect"
}
```
</p>
</details>
|
defect
|
triggertype and triggermode situation the domlaunch enums not set trac migrated from json status closed changetime ts description the enums triggertype and triggermode from domlaunch h are not set in current mc this would be very useful to understand trigger behaviour like hlc slc n nshould go in the old and upcoming new domsimulator maybe n reporter icecube cc resolution wontfix time component combo simulation summary triggertype and triggermode situation the domlaunch enums not set priority normal keywords milestone owner olivas sflis type defect
| 1
|
348,293
| 31,495,245,190
|
IssuesEvent
|
2023-08-31 01:26:31
|
WordPress/gutenberg
|
https://api.github.com/repos/WordPress/gutenberg
|
closed
|
`TreeGrid`: write unit tests after recent a11y improvements
|
[Type] Enhancement [Type] Automated Testing [Package] Components [Feature] Component System
|
Hey folks, I noticed a few great accessibility fixes / improvements on `TreeGrid` recently!
It would be great if we could also add a few unit tests to make sure we don't introduce any regressions around these areas in the future— would that be possible?
_Originally posted by @ciampo in https://github.com/WordPress/gutenberg/issues/38679#issuecomment-1034744487_
|
1.0
|
`TreeGrid`: write unit tests after recent a11y improvements - Hey folks, I noticed a few great accessibility fixes / improvements on `TreeGrid` recently!
It would be great if we could also add a few unit tests to make sure we don't introduce any regressions around these areas in the future— would that be possible?
_Originally posted by @ciampo in https://github.com/WordPress/gutenberg/issues/38679#issuecomment-1034744487_
|
non_defect
|
treegrid write unit tests after recent improvements hey folks i noticed a few great accessibility fixes improvements on treegrid recently it would be great if we could also add a few unit tests to make sure we don t introduce any regressions around these areas in the future— would that be possible originally posted by ciampo in
| 0
|
4,599
| 2,610,121,475
|
IssuesEvent
|
2015-02-26 18:37:43
|
chrsmith/scribefire-chrome
|
https://api.github.com/repos/chrsmith/scribefire-chrome
|
closed
|
Red bar error
|
auto-migrated Priority-Medium Type-Defect
|
```
What's the problem?
-> When I open the extension I get a huge red bar saying some code: "<!DOCTYPE
html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN"..."
What browser are you using?
-> Google Chrome
What version of ScribeFire are you running?
-> 1.4.2.0
```
-----
Original issue reported on code.google.com by `thomazm...@gmail.com` on 26 Oct 2010 at 2:10
|
1.0
|
Red bar error - ```
What's the problem?
-> When I open the extension I get a huge red bar saying some code: "<!DOCTYPE
html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN"..."
What browser are you using?
-> Google Chrome
What version of ScribeFire are you running?
-> 1.4.2.0
```
-----
Original issue reported on code.google.com by `thomazm...@gmail.com` on 26 Oct 2010 at 2:10
|
defect
|
red bar error what s the problem when i open the extension i get a huge red bar saying some code doctype html public dtd xhtml transitional en what browser are you using google chrome what version of scribefire are you running original issue reported on code google com by thomazm gmail com on oct at
| 1
|
17,460
| 3,006,534,529
|
IssuesEvent
|
2015-07-27 11:04:27
|
bridgedotnet/Bridge
|
https://api.github.com/repos/bridgedotnet/Bridge
|
closed
|
[Array] Should implement ICollection, IEnumerable, IList, ICloneable
|
defect
|
```C#
using Bridge;
using Bridge.Html5;
using System;
using System.Collections.Generic;
using System.Linq;
namespace Demo
{
public class App
{
[Ready]
public static void Main()
{
Html5.Console.Log("Is IEnumerable: " + (arr is IEnumerable));
Html5.Console.Log("Is ICollection: " + (arr is ICollection));
Html5.Console.Log("Is ICloneable: " + (arr is ICloneable));
Html5.Console.Log("Is ICollection<int>: " + (arr is ICollection<int>));
Html5.Console.Log("Is IEnumerable<int>: " + (arr is IEnumerable<int>));
Html5.Console.Log("Is IList<int>: " + (arr is IList<int>));
}
}
}
```
Actual result: implements IEnumerable only.
Code snippet to test .Net code in LINQPad:
```C#
public static void Main(params object[] p)
{
object arr = new[] { 1, 2, 3 };
(arr is Array).Dump();
(arr is int[]).Dump();
(arr is ICollection).Dump();
(arr is IEnumerable).Dump();
(arr is ICloneable).Dump();
(arr is IList).Dump();
(arr is ICollection<int>).Dump();
(arr is IEnumerable<int>).Dump();
(arr is IList<int>).Dump();
}
```
|
1.0
|
[Array] Should implement ICollection, IEnumerable, IList, ICloneable - ```C#
using Bridge;
using Bridge.Html5;
using System;
using System.Collections.Generic;
using System.Linq;
namespace Demo
{
public class App
{
[Ready]
public static void Main()
{
Html5.Console.Log("Is IEnumerable: " + (arr is IEnumerable));
Html5.Console.Log("Is ICollection: " + (arr is ICollection));
Html5.Console.Log("Is ICloneable: " + (arr is ICloneable));
Html5.Console.Log("Is ICollection<int>: " + (arr is ICollection<int>));
Html5.Console.Log("Is IEnumerable<int>: " + (arr is IEnumerable<int>));
Html5.Console.Log("Is IList<int>: " + (arr is IList<int>));
}
}
}
```
Actual result: implements IEnumerable only.
Code snippet to test .Net code in LINQPad:
```C#
public static void Main(params object[] p)
{
object arr = new[] { 1, 2, 3 };
(arr is Array).Dump();
(arr is int[]).Dump();
(arr is ICollection).Dump();
(arr is IEnumerable).Dump();
(arr is ICloneable).Dump();
(arr is IList).Dump();
(arr is ICollection<int>).Dump();
(arr is IEnumerable<int>).Dump();
(arr is IList<int>).Dump();
}
```
|
defect
|
should implement icollection ienumerable ilist icloneable c using bridge using bridge using system using system collections generic using system linq namespace demo public class app public static void main console log is ienumerable arr is ienumerable console log is icollection arr is icollection console log is icloneable arr is icloneable console log is icollection arr is icollection console log is ienumerable arr is ienumerable console log is ilist arr is ilist actual result implements ienumerable only code snippet to test net code in linqpad c public static void main params object p object arr new arr is array dump arr is int dump arr is icollection dump arr is ienumerable dump arr is icloneable dump arr is ilist dump arr is icollection dump arr is ienumerable dump arr is ilist dump
| 1
|
49,229
| 13,445,715,013
|
IssuesEvent
|
2020-09-08 11:52:14
|
chaitanya00/aem-wknd
|
https://api.github.com/repos/chaitanya00/aem-wknd
|
opened
|
WS-2018-0224 (Medium) detected in mpath-0.1.1.tgz
|
security vulnerability
|
## WS-2018-0224 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mpath-0.1.1.tgz</b></p></summary>
<p>{G,S}et object values using MongoDB path notation</p>
<p>Library home page: <a href="https://registry.npmjs.org/mpath/-/mpath-0.1.1.tgz">https://registry.npmjs.org/mpath/-/mpath-0.1.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/aem-wknd/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/aem-wknd/node_modules/mpath/package.json</p>
<p>
Dependency Hierarchy:
- mongoose-4.2.4.tgz (Root Library)
- :x: **mpath-0.1.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/chaitanya00/aem-wknd/commit/3f4c2902a45eb04bc7915c408df14545aa90511c">3f4c2902a45eb04bc7915c408df14545aa90511c</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Mpath, versions 0.0.1--0.0.5, have a Prototype Pollution Vulnerability. An attacker can specify a path that include the prototype object.
<p>Publish Date: 2018-08-30
<p>URL: <a href=https://github.com/aheckmann/mpath/commit/fe732eb05adebe29d1d741bdf3981afbae8ea94d>WS-2018-0224</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/390860">https://hackerone.com/reports/390860</a></p>
<p>Release Date: 2018-12-13</p>
<p>Fix Resolution: 0.5.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2018-0224 (Medium) detected in mpath-0.1.1.tgz - ## WS-2018-0224 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mpath-0.1.1.tgz</b></p></summary>
<p>{G,S}et object values using MongoDB path notation</p>
<p>Library home page: <a href="https://registry.npmjs.org/mpath/-/mpath-0.1.1.tgz">https://registry.npmjs.org/mpath/-/mpath-0.1.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/aem-wknd/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/aem-wknd/node_modules/mpath/package.json</p>
<p>
Dependency Hierarchy:
- mongoose-4.2.4.tgz (Root Library)
- :x: **mpath-0.1.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/chaitanya00/aem-wknd/commit/3f4c2902a45eb04bc7915c408df14545aa90511c">3f4c2902a45eb04bc7915c408df14545aa90511c</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Mpath, versions 0.0.1--0.0.5, have a Prototype Pollution Vulnerability. An attacker can specify a path that include the prototype object.
<p>Publish Date: 2018-08-30
<p>URL: <a href=https://github.com/aheckmann/mpath/commit/fe732eb05adebe29d1d741bdf3981afbae8ea94d>WS-2018-0224</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/390860">https://hackerone.com/reports/390860</a></p>
<p>Release Date: 2018-12-13</p>
<p>Fix Resolution: 0.5.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
ws medium detected in mpath tgz ws medium severity vulnerability vulnerable library mpath tgz g s et object values using mongodb path notation library home page a href path to dependency file tmp ws scm aem wknd package json path to vulnerable library tmp ws scm aem wknd node modules mpath package json dependency hierarchy mongoose tgz root library x mpath tgz vulnerable library found in head commit a href vulnerability details mpath versions have a prototype pollution vulnerability an attacker can specify a path that include the prototype object publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
1,785
| 2,603,971,379
|
IssuesEvent
|
2015-02-24 19:00:20
|
chrsmith/nishazi6
|
https://api.github.com/repos/chrsmith/nishazi6
|
opened
|
沈阳龟头上长了一圈肉芽
|
auto-migrated Priority-Medium Type-Defect
|
```
沈阳龟头上长了一圈肉芽〓沈陽軍區政治部醫院性病〓TEL:02
4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療�
��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝�
��的歷史悠久、設備精良、技術權威、專家云集,是預防、保
健、醫療、科研康復為一體的綜合性醫院。是國家首批公立��
�等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學�
��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍
空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集��
�二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:33
|
1.0
|
沈阳龟头上长了一圈肉芽 - ```
沈阳龟头上长了一圈肉芽〓沈陽軍區政治部醫院性病〓TEL:02
4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療�
��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝�
��的歷史悠久、設備精良、技術權威、專家云集,是預防、保
健、醫療、科研康復為一體的綜合性醫院。是國家首批公立��
�等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學�
��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍
空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集��
�二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:33
|
defect
|
沈阳龟头上长了一圈肉芽 沈阳龟头上长了一圈肉芽〓沈陽軍區政治部醫院性病〓tel: 〓 , � �� 。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 original issue reported on code google com by gmail com on jun at
| 1
|
1,696
| 2,603,969,690
|
IssuesEvent
|
2015-02-24 18:59:56
|
chrsmith/nishazi6
|
https://api.github.com/repos/chrsmith/nishazi6
|
opened
|
沈阳包皮上起小疙瘩
|
auto-migrated Priority-Medium Type-Defect
|
```
沈阳包皮上起小疙瘩〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:24
|
1.0
|
沈阳包皮上起小疙瘩 - ```
沈阳包皮上起小疙瘩〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:24
|
defect
|
沈阳包皮上起小疙瘩 沈阳包皮上起小疙瘩〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位� �� 。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 original issue reported on code google com by gmail com on jun at
| 1
|
31,686
| 13,616,160,783
|
IssuesEvent
|
2020-09-23 15:17:30
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
closed
|
Map Projects for eCapris - Manuel Gallegos
|
Need: 1-Must Have Service: Geo Type: Data Workgroup: TED
|
> Please disregard my previous email, see below updated list with TED projects that need to be mapped in eCapris:
>
> • 6598.056 - 2018 Bond Congress/Ramble Intersection Safety Improvements – 2018 Bond
> • 11580.030 – Safety and Mobility Improvements on West Gate Blvd. (from Willian Cannon to Cameron Loop) – Quarter Cent
> • 11580.058 – Intersection Improvements at N Lamar and Morrow St. – Quarter Cent
> • 11899.010 - Barton Springs Rd - South 1st St Intersection Safety Improvements – 2016 Bond
> • 11899.011 - IH-35 SR (NB) - 7th St Intersection Safety Improvements – 2016 Bond
> • 11899.013 - 8th St - IH-35 Intersection Safety Improvements – 2016 Bond
> • 11899.015 - US 183 SR (NB)/Lakeline Blvd, Intersection Safety Improvements – 2016 Bond
> • 11899.016 - IH 35 / Rundberg Ln, Intersection Safety Improvements – 2016 Bond
>
> CPE website shown a pin for the following locations. Can we do an intersection as the other projects? (If take too much time or is more complicated we can leave them as they are).
>
> • 11899.009 - Braker Ln./Stonelake Blvd. Intersection Safety Improvements – 2016 Bond
> • 11899.014 - Lamar/St Johns Intersection Safety Improvements – 2016 Bond
|
1.0
|
Map Projects for eCapris - Manuel Gallegos - > Please disregard my previous email, see below updated list with TED projects that need to be mapped in eCapris:
>
> • 6598.056 - 2018 Bond Congress/Ramble Intersection Safety Improvements – 2018 Bond
> • 11580.030 – Safety and Mobility Improvements on West Gate Blvd. (from Willian Cannon to Cameron Loop) – Quarter Cent
> • 11580.058 – Intersection Improvements at N Lamar and Morrow St. – Quarter Cent
> • 11899.010 - Barton Springs Rd - South 1st St Intersection Safety Improvements – 2016 Bond
> • 11899.011 - IH-35 SR (NB) - 7th St Intersection Safety Improvements – 2016 Bond
> • 11899.013 - 8th St - IH-35 Intersection Safety Improvements – 2016 Bond
> • 11899.015 - US 183 SR (NB)/Lakeline Blvd, Intersection Safety Improvements – 2016 Bond
> • 11899.016 - IH 35 / Rundberg Ln, Intersection Safety Improvements – 2016 Bond
>
> CPE website shown a pin for the following locations. Can we do an intersection as the other projects? (If take too much time or is more complicated we can leave them as they are).
>
> • 11899.009 - Braker Ln./Stonelake Blvd. Intersection Safety Improvements – 2016 Bond
> • 11899.014 - Lamar/St Johns Intersection Safety Improvements – 2016 Bond
|
non_defect
|
map projects for ecapris manuel gallegos please disregard my previous email see below updated list with ted projects that need to be mapped in ecapris • bond congress ramble intersection safety improvements – bond • – safety and mobility improvements on west gate blvd from willian cannon to cameron loop – quarter cent • – intersection improvements at n lamar and morrow st – quarter cent • barton springs rd south st intersection safety improvements – bond • ih sr nb st intersection safety improvements – bond • st ih intersection safety improvements – bond • us sr nb lakeline blvd intersection safety improvements – bond • ih rundberg ln intersection safety improvements – bond cpe website shown a pin for the following locations can we do an intersection as the other projects if take too much time or is more complicated we can leave them as they are • braker ln stonelake blvd intersection safety improvements – bond • lamar st johns intersection safety improvements – bond
| 0
|
86,468
| 17,009,823,603
|
IssuesEvent
|
2021-07-02 01:29:20
|
microsoft/vscode-dev-containers
|
https://api.github.com/repos/microsoft/vscode-dev-containers
|
closed
|
Switch to using Linux package managers to acquire kubectl, helm, GitHub CLI
|
codespaces feature-request script-library
|
There are a few spots in scripts in this repository that use curl to download a utility even when native Linux package manager support is available. Using apt-get has the added benefit of improved security due to signing and easier and simplified upgrades to the packages.
In particular, the following can be switched:
- [kubectl](https://kubernetes.io/docs/tasks/tools/install-kubectl/#install-using-native-package-management)
- [helm](https://helm.sh/docs/intro/install/#from-apt-debianubuntu)
- [The GitHub CLI](https://github.com/cli/cli/blob/trunk/docs/install_linux.md#debian-ubuntu-linux-apt)
- [Terraform](https://www.terraform.io/docs/cli/install/apt.html)
|
1.0
|
Switch to using Linux package managers to acquire kubectl, helm, GitHub CLI - There are a few spots in scripts in this repository that use curl to download a utility even when native Linux package manager support is available. Using apt-get has the added benefit of improved security due to signing and easier and simplified upgrades to the packages.
In particular, the following can be switched:
- [kubectl](https://kubernetes.io/docs/tasks/tools/install-kubectl/#install-using-native-package-management)
- [helm](https://helm.sh/docs/intro/install/#from-apt-debianubuntu)
- [The GitHub CLI](https://github.com/cli/cli/blob/trunk/docs/install_linux.md#debian-ubuntu-linux-apt)
- [Terraform](https://www.terraform.io/docs/cli/install/apt.html)
|
non_defect
|
switch to using linux package managers to acquire kubectl helm github cli there are a few spots in scripts in this repository that use curl to download a utility even when native linux package manager support is available using apt get has the added benefit of improved security due to signing and easier and simplified upgrades to the packages in particular the following can be switched
| 0
|
31,523
| 6,544,159,136
|
IssuesEvent
|
2017-09-03 12:31:02
|
bridgedotnet/Bridge
|
https://api.github.com/repos/bridgedotnet/Bridge
|
closed
|
[Template] with implicit operators broke compilation
|
defect
|
### Steps To Reproduce
https://deck.net/efa24e49e859f6ebb14dfd44234a74c4
```c#
public class C
{
[Template("({0})")]
[Script("return c")]
static public implicit operator C(string c)
{
return null;
}
}
public class Program
{
public static void Main()
{
var a = new C();
C b = method();
}
static public string method()
{
return "";
}
}
```
### Actual Result
```js
Bridge.assembly("Demo", function ($asm, globals) {
"use strict";
Bridge.define("Demo.C");
Bridge.define("Demo.Program", {
main: function Main () {
var a = new Demo.C();
var b =
var $m = Bridge.setMetadata,
$n = [System,Demo];
$m($n[1].C, function () { return {"att":1048577,"a":2,"m":[{"a":2,"isSynthetic":true,"n":".ctor","t":1,"sn":"ctor"},{"a":2,"n":"op_Implicit","is":true,"t":8,"pi":[{"n":"c","pt":$n[0].String,"ps":0}],"tpc":0,"def":function (c) { return (c); },"rt":$n[1].C,"p":[$n[0].String]}]}; });
$m($n[1].Program, function () { return {"att":1048577,"a":2,"m":[{"a":2,"isSynthetic":true,"n":".ctor","t":1,"sn":"ctor"},{"a":2,"n":"Main","is":true,"t":8,"sn":"Main","rt":$n[0].Void},{"a":2,"n":"method","is":true,"t":8,"sn":"method","rt":$n[0].String}]}; });
});
```
|
1.0
|
[Template] with implicit operators broke compilation - ### Steps To Reproduce
https://deck.net/efa24e49e859f6ebb14dfd44234a74c4
```c#
public class C
{
[Template("({0})")]
[Script("return c")]
static public implicit operator C(string c)
{
return null;
}
}
public class Program
{
public static void Main()
{
var a = new C();
C b = method();
}
static public string method()
{
return "";
}
}
```
### Actual Result
```js
Bridge.assembly("Demo", function ($asm, globals) {
"use strict";
Bridge.define("Demo.C");
Bridge.define("Demo.Program", {
main: function Main () {
var a = new Demo.C();
var b =
var $m = Bridge.setMetadata,
$n = [System,Demo];
$m($n[1].C, function () { return {"att":1048577,"a":2,"m":[{"a":2,"isSynthetic":true,"n":".ctor","t":1,"sn":"ctor"},{"a":2,"n":"op_Implicit","is":true,"t":8,"pi":[{"n":"c","pt":$n[0].String,"ps":0}],"tpc":0,"def":function (c) { return (c); },"rt":$n[1].C,"p":[$n[0].String]}]}; });
$m($n[1].Program, function () { return {"att":1048577,"a":2,"m":[{"a":2,"isSynthetic":true,"n":".ctor","t":1,"sn":"ctor"},{"a":2,"n":"Main","is":true,"t":8,"sn":"Main","rt":$n[0].Void},{"a":2,"n":"method","is":true,"t":8,"sn":"method","rt":$n[0].String}]}; });
});
```
|
defect
|
with implicit operators broke compilation steps to reproduce c public class c static public implicit operator c string c return null public class program public static void main var a new c c b method static public string method return actual result js bridge assembly demo function asm globals use strict bridge define demo c bridge define demo program main function main var a new demo c var b var m bridge setmetadata n m n c function return att a m string ps tpc def function c return c rt n c p string m n program function return att a m void a n method is true t sn method rt n string
| 1
|
54,629
| 13,791,435,857
|
IssuesEvent
|
2020-10-09 12:08:41
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
opened
|
sql service operations hang from cluster split brain test
|
Module: SQL Source: Internal Team: Core Type: Defect
|
http://jenkins.hazelcast.com/view/split/job/split-x/45/console
hz-bench-run-jenkins-split hz/stress/sql-service
/disk1/jenkins/workspace/split-x/4.1-SNAPSHOT/2020_10_09-10_15_12/sql-service
a hanging run with only sql Service operations
```
hanging at
SqlResult sqlResult = sqlService.execute(sqlStatement);
and at result set iteration
for (SqlRow row : sqlResult) {
```
|
1.0
|
sql service operations hang from cluster split brain test -
http://jenkins.hazelcast.com/view/split/job/split-x/45/console
hz-bench-run-jenkins-split hz/stress/sql-service
/disk1/jenkins/workspace/split-x/4.1-SNAPSHOT/2020_10_09-10_15_12/sql-service
a hanging run with only sql Service operations
```
hanging at
SqlResult sqlResult = sqlService.execute(sqlStatement);
and at result set iteration
for (SqlRow row : sqlResult) {
```
|
defect
|
sql service operations hang from cluster split brain test hz bench run jenkins split hz stress sql service jenkins workspace split x snapshot sql service a hanging run with only sql service operations hanging at sqlresult sqlresult sqlservice execute sqlstatement and at result set iteration for sqlrow row sqlresult
| 1
|
41,998
| 10,738,045,687
|
IssuesEvent
|
2019-10-29 14:10:24
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
closed
|
JDK9 + issues
|
Team: Core Type: Defect
|
When running with JDK9, we'll get exceptions like
```
Sep 27, 2019 10:31:49 AM com.hazelcast.internal.metrics.impl.MetricsService
WARNING: [10.212.1.102]:5701 [workers] [4.0-SNAPSHOT] Error when rendering 'memory.totalPhysical'
java.lang.reflect.InvocationTargetException
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at com.hazelcast.internal.metrics.impl.MethodProbe$LongMethodProbe.get(MethodProbe.java:107)
at com.hazelcast.internal.metrics.impl.MetricsRegistryImpl.render(MetricsRegistryImpl.java:322)
at com.hazelcast.internal.metrics.impl.MetricsRegistryImpl.render(MetricsRegistryImpl.java:270)
at com.hazelcast.internal.metrics.impl.MetricsService.collectMetrics(MetricsService.java:155)
at com.hazelcast.internal.metrics.impl.MetricsService.collectMetrics(MetricsService.java:150)
at com.hazelcast.spi.impl.executionservice.impl.DelegateAndSkipOnConcurrentExecutionDecorator$DelegateDecorator.run(DelegateAndSkipOnConcurrentExecutionDecorator.java:77)
at com.hazelcast.internal.util.executor.CachedExecutorServiceDelegate$Worker.run(CachedExecutorServiceDelegate.java:227)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834)
at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64)
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80)
Caused by: java.lang.NoClassDefFoundError: Could not initialize class com.hazelcast.internal.memory.MemoryStatsSupport
at com.hazelcast.internal.memory.DefaultMemoryStats.getTotalPhysical(DefaultMemoryStats.java:34)
... 16 more
```
Instead of getting all of this noise, we should clean it up.
|
1.0
|
JDK9 + issues - When running with JDK9, we'll get exceptions like
```
Sep 27, 2019 10:31:49 AM com.hazelcast.internal.metrics.impl.MetricsService
WARNING: [10.212.1.102]:5701 [workers] [4.0-SNAPSHOT] Error when rendering 'memory.totalPhysical'
java.lang.reflect.InvocationTargetException
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at com.hazelcast.internal.metrics.impl.MethodProbe$LongMethodProbe.get(MethodProbe.java:107)
at com.hazelcast.internal.metrics.impl.MetricsRegistryImpl.render(MetricsRegistryImpl.java:322)
at com.hazelcast.internal.metrics.impl.MetricsRegistryImpl.render(MetricsRegistryImpl.java:270)
at com.hazelcast.internal.metrics.impl.MetricsService.collectMetrics(MetricsService.java:155)
at com.hazelcast.internal.metrics.impl.MetricsService.collectMetrics(MetricsService.java:150)
at com.hazelcast.spi.impl.executionservice.impl.DelegateAndSkipOnConcurrentExecutionDecorator$DelegateDecorator.run(DelegateAndSkipOnConcurrentExecutionDecorator.java:77)
at com.hazelcast.internal.util.executor.CachedExecutorServiceDelegate$Worker.run(CachedExecutorServiceDelegate.java:227)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834)
at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64)
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80)
Caused by: java.lang.NoClassDefFoundError: Could not initialize class com.hazelcast.internal.memory.MemoryStatsSupport
at com.hazelcast.internal.memory.DefaultMemoryStats.getTotalPhysical(DefaultMemoryStats.java:34)
... 16 more
```
Instead of getting all of this noise, we should clean it up.
|
defect
|
issues when running with we ll get exceptions like sep am com hazelcast internal metrics impl metricsservice warning error when rendering memory totalphysical java lang reflect invocationtargetexception at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at com hazelcast internal metrics impl methodprobe longmethodprobe get methodprobe java at com hazelcast internal metrics impl metricsregistryimpl render metricsregistryimpl java at com hazelcast internal metrics impl metricsregistryimpl render metricsregistryimpl java at com hazelcast internal metrics impl metricsservice collectmetrics metricsservice java at com hazelcast internal metrics impl metricsservice collectmetrics metricsservice java at com hazelcast spi impl executionservice impl delegateandskiponconcurrentexecutiondecorator delegatedecorator run delegateandskiponconcurrentexecutiondecorator java at com hazelcast internal util executor cachedexecutorservicedelegate worker run cachedexecutorservicedelegate java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java at com hazelcast internal util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast internal util executor hazelcastmanagedthread run hazelcastmanagedthread java caused by java lang noclassdeffounderror could not initialize class com hazelcast internal memory memorystatssupport at com hazelcast internal memory defaultmemorystats gettotalphysical defaultmemorystats java more instead of getting all of this noise we should clean it up
| 1
|
76,703
| 26,560,309,746
|
IssuesEvent
|
2023-01-20 15:21:02
|
decentraland/unity-renderer
|
https://api.github.com/repos/decentraland/unity-renderer
|
opened
|
[QA] Brief zoom-in before teleportation on fullscreen on MacOS desktop
|
minor new defect stream-core
|
Description:
On MacOS desktop client set to fullscreen a brief zoom-in to the bottom-left corner occurs before every kind of teleportation. Zoom occurs after issuing the teleportation command but before the loading screen is shown. Issue is best observable wit hthe usage of the `/goto` command.
Reproduction rate:
5/5
Steps to reproduce:
1. Launch the desktop client with Desktop Developer Mode on the release/release-20230119 branch with the ENABLE_DECOUPLED_LOADING_SCREEN parameter.
2. Log in either via a guest or a wallet account. Ensure that the client is set to fullscreen in the graphical settings.
3. Use the `/goto random' command to teleport to a random place.
4. Observe the renderer window. Notice the bottom-left section of the screen is briefly zoomed-in before the loading screen appears.
Results:
Bottom-left section of the screen is briefly zoomed-in before the loading screen appears.
Expected results:
No zoom-in occurs before the loading screen appears.
Notes:
- Platforms: Windows and MacOS desktop, wallet and guest login.
- Issue occurs for the [release/release-20230119 branch](https://github.com/decentraland/explorer-desktop/pull/464).
|
1.0
|
[QA] Brief zoom-in before teleportation on fullscreen on MacOS desktop - Description:
On MacOS desktop client set to fullscreen a brief zoom-in to the bottom-left corner occurs before every kind of teleportation. Zoom occurs after issuing the teleportation command but before the loading screen is shown. Issue is best observable wit hthe usage of the `/goto` command.
Reproduction rate:
5/5
Steps to reproduce:
1. Launch the desktop client with Desktop Developer Mode on the release/release-20230119 branch with the ENABLE_DECOUPLED_LOADING_SCREEN parameter.
2. Log in either via a guest or a wallet account. Ensure that the client is set to fullscreen in the graphical settings.
3. Use the `/goto random' command to teleport to a random place.
4. Observe the renderer window. Notice the bottom-left section of the screen is briefly zoomed-in before the loading screen appears.
Results:
Bottom-left section of the screen is briefly zoomed-in before the loading screen appears.
Expected results:
No zoom-in occurs before the loading screen appears.
Notes:
- Platforms: Windows and MacOS desktop, wallet and guest login.
- Issue occurs for the [release/release-20230119 branch](https://github.com/decentraland/explorer-desktop/pull/464).
|
defect
|
brief zoom in before teleportation on fullscreen on macos desktop description on macos desktop client set to fullscreen a brief zoom in to the bottom left corner occurs before every kind of teleportation zoom occurs after issuing the teleportation command but before the loading screen is shown issue is best observable wit hthe usage of the goto command reproduction rate steps to reproduce launch the desktop client with desktop developer mode on the release release branch with the enable decoupled loading screen parameter log in either via a guest or a wallet account ensure that the client is set to fullscreen in the graphical settings use the goto random command to teleport to a random place observe the renderer window notice the bottom left section of the screen is briefly zoomed in before the loading screen appears results bottom left section of the screen is briefly zoomed in before the loading screen appears expected results no zoom in occurs before the loading screen appears notes platforms windows and macos desktop wallet and guest login issue occurs for the
| 1
|
14,544
| 2,822,945,766
|
IssuesEvent
|
2015-05-21 04:28:03
|
gingeleski/menusearch
|
https://api.github.com/repos/gingeleski/menusearch
|
closed
|
SQL exception in MenuDBAccess.retrieve(String query)
|
auto-migrated Priority-High Type-Defect
|
```
A SQL Exception, complaining about incorrect SQL syntax, is being generated in
this method. The problem is likely in the calling method.
```
Original issue reported on code.google.com by `bkmackellar` on 3 May 2015 at 3:52
|
1.0
|
SQL exception in MenuDBAccess.retrieve(String query) - ```
A SQL Exception, complaining about incorrect SQL syntax, is being generated in
this method. The problem is likely in the calling method.
```
Original issue reported on code.google.com by `bkmackellar` on 3 May 2015 at 3:52
|
defect
|
sql exception in menudbaccess retrieve string query a sql exception complaining about incorrect sql syntax is being generated in this method the problem is likely in the calling method original issue reported on code google com by bkmackellar on may at
| 1
|
44,122
| 11,981,891,168
|
IssuesEvent
|
2020-04-07 12:00:06
|
lagom/lagom
|
https://api.github.com/repos/lagom/lagom
|
closed
|
Different behavior in ScaladslServiceRouter.maybeLogException and JavadslServiceRouter.maybeLogException
|
type:defect
|
### Lagom Version (1.2.x / 1.3.x / etc)
1.6.x and below
### API (Scala / Java / Neither / Both)
Both
### Expected Behavior
[ScaladslServiceRouter.maybeLogException](https://github.com/lagom/lagom/blob/fb3fa8c57cd8b5ecb7bbcf8674f9f06e86bbae0a/service/scaladsl/server/src/main/scala/com/lightbend/lagom/internal/scaladsl/server/ScaladslServiceRouter.scala#L96
) and [JavadslServiceRouter.maybeLogException](https://github.com/lagom/lagom/blob/fb3fa8c57cd8b5ecb7bbcf8674f9f06e86bbae0a/service/javadsl/server/src/main/scala/com/lightbend/lagom/internal/javadsl/server/JavadslServerBuilder.scala#L233
) there must be the same behavior
### Actual Behavior
`ScaladslServiceRouter.maybeLogException` logs `BadRequest` exceptions
|
1.0
|
Different behavior in ScaladslServiceRouter.maybeLogException and JavadslServiceRouter.maybeLogException - ### Lagom Version (1.2.x / 1.3.x / etc)
1.6.x and below
### API (Scala / Java / Neither / Both)
Both
### Expected Behavior
[ScaladslServiceRouter.maybeLogException](https://github.com/lagom/lagom/blob/fb3fa8c57cd8b5ecb7bbcf8674f9f06e86bbae0a/service/scaladsl/server/src/main/scala/com/lightbend/lagom/internal/scaladsl/server/ScaladslServiceRouter.scala#L96
) and [JavadslServiceRouter.maybeLogException](https://github.com/lagom/lagom/blob/fb3fa8c57cd8b5ecb7bbcf8674f9f06e86bbae0a/service/javadsl/server/src/main/scala/com/lightbend/lagom/internal/javadsl/server/JavadslServerBuilder.scala#L233
) there must be the same behavior
### Actual Behavior
`ScaladslServiceRouter.maybeLogException` logs `BadRequest` exceptions
|
defect
|
different behavior in scaladslservicerouter maybelogexception and javadslservicerouter maybelogexception lagom version x x etc x and below api scala java neither both both expected behavior and there must be the same behavior actual behavior scaladslservicerouter maybelogexception logs badrequest exceptions
| 1
|
46,236
| 13,055,873,966
|
IssuesEvent
|
2020-07-30 02:59:31
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
[trigger-sim] ShiftFrameObjects from TimeShifter python implementation not found in simprod dataset 11169 (Trac #810)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
Migrated from https://code.icecube.wisc.edu/ticket/810
```json
{
"status": "closed",
"changetime": "2015-01-29T20:08:59",
"description": "When I try to use the time shifter implementation in the trigger_sim __init__.py I get the following error in dataset 11169(http://internal.icecube.wisc.edu/simulation/dataset/11169):\n\n----iceprod.11169.0.err----:\n/lib/I3Tray.py\", line 231, in Execute \nsuper(I3Tray, self).Execute() \nFile \"/var/lib/condor/execute/slot1/dir_23393/tmp/slot3_icesoft/simulation.releases.V04-01-08.r125626.Linux-x86_64.gcc-4.4.6/lib/icecube/trigger_sim/modules/time_shifter.py\", line 56, in DAQ \nShiftFrameObjects(frame,DELTA_T,self.skip_keys) \nRuntimeError: unregistered class \nError: IceTray exited with status (256)'\n\nNot sure what combination of the tarball and the simprod settings are causing this problem...",
"reporter": "melanie.day",
"cc": "nega",
"resolution": "worksforme",
"_ts": "1422562139640742",
"component": "combo simulation",
"summary": "[trigger-sim] ShiftFrameObjects from TimeShifter python implementation not found in simprod dataset 11169",
"priority": "normal",
"keywords": "",
"time": "2014-11-19T22:31:20",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
|
1.0
|
[trigger-sim] ShiftFrameObjects from TimeShifter python implementation not found in simprod dataset 11169 (Trac #810) - Migrated from https://code.icecube.wisc.edu/ticket/810
```json
{
"status": "closed",
"changetime": "2015-01-29T20:08:59",
"description": "When I try to use the time shifter implementation in the trigger_sim __init__.py I get the following error in dataset 11169(http://internal.icecube.wisc.edu/simulation/dataset/11169):\n\n----iceprod.11169.0.err----:\n/lib/I3Tray.py\", line 231, in Execute \nsuper(I3Tray, self).Execute() \nFile \"/var/lib/condor/execute/slot1/dir_23393/tmp/slot3_icesoft/simulation.releases.V04-01-08.r125626.Linux-x86_64.gcc-4.4.6/lib/icecube/trigger_sim/modules/time_shifter.py\", line 56, in DAQ \nShiftFrameObjects(frame,DELTA_T,self.skip_keys) \nRuntimeError: unregistered class \nError: IceTray exited with status (256)'\n\nNot sure what combination of the tarball and the simprod settings are causing this problem...",
"reporter": "melanie.day",
"cc": "nega",
"resolution": "worksforme",
"_ts": "1422562139640742",
"component": "combo simulation",
"summary": "[trigger-sim] ShiftFrameObjects from TimeShifter python implementation not found in simprod dataset 11169",
"priority": "normal",
"keywords": "",
"time": "2014-11-19T22:31:20",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
|
defect
|
shiftframeobjects from timeshifter python implementation not found in simprod dataset trac migrated from json status closed changetime description when i try to use the time shifter implementation in the trigger sim init py i get the following error in dataset line in execute nsuper self execute nfile var lib condor execute dir tmp icesoft simulation releases linux gcc lib icecube trigger sim modules time shifter py line in daq nshiftframeobjects frame delta t self skip keys nruntimeerror unregistered class nerror icetray exited with status n nnot sure what combination of the tarball and the simprod settings are causing this problem reporter melanie day cc nega resolution worksforme ts component combo simulation summary shiftframeobjects from timeshifter python implementation not found in simprod dataset priority normal keywords time milestone owner olivas type defect
| 1
|
31,984
| 6,673,932,383
|
IssuesEvent
|
2017-10-04 16:39:29
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
opened
|
3.6: Do not silently load invalid/inexistent tasks
|
Defect Enhancement
|
This is a (multiple allowed):
* [x] bug
* [x] enhancement
* [ ] feature-discussion (RFC)
```php
public $tasks = [
'Bake.SuperInvalid',
...
];
```
Refs https://github.com/cakephp/bake/pull/362
We should at least trigger an error, better yet throw a clear exception if an invalid task (or typo) is loaded.
|
1.0
|
3.6: Do not silently load invalid/inexistent tasks - This is a (multiple allowed):
* [x] bug
* [x] enhancement
* [ ] feature-discussion (RFC)
```php
public $tasks = [
'Bake.SuperInvalid',
...
];
```
Refs https://github.com/cakephp/bake/pull/362
We should at least trigger an error, better yet throw a clear exception if an invalid task (or typo) is loaded.
|
defect
|
do not silently load invalid inexistent tasks this is a multiple allowed bug enhancement feature discussion rfc php public tasks bake superinvalid refs we should at least trigger an error better yet throw a clear exception if an invalid task or typo is loaded
| 1
|
51,892
| 13,211,334,606
|
IssuesEvent
|
2020-08-15 22:23:21
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
pybdt low test coverage (Trac #1226)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1226">https://code.icecube.wisc.edu/projects/icecube/ticket/1226</a>, reported by jtatarand owned by richman</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:38",
"_ts": "1550067158057333",
"description": "The pybdt test coverage of certain sections of the code is low. Please try to increase the test coverage as much as possible.\n\nFor details please see:\n\nhttp://software.icecube.wisc.edu/coverage/00_LATEST/pybdt/private/pybdt/index.html\n\n",
"reporter": "jtatar",
"cc": "",
"resolution": "fixed",
"time": "2015-08-19T20:59:18",
"component": "combo reconstruction",
"summary": "pybdt low test coverage",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "richman",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
pybdt low test coverage (Trac #1226) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1226">https://code.icecube.wisc.edu/projects/icecube/ticket/1226</a>, reported by jtatarand owned by richman</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:38",
"_ts": "1550067158057333",
"description": "The pybdt test coverage of certain sections of the code is low. Please try to increase the test coverage as much as possible.\n\nFor details please see:\n\nhttp://software.icecube.wisc.edu/coverage/00_LATEST/pybdt/private/pybdt/index.html\n\n",
"reporter": "jtatar",
"cc": "",
"resolution": "fixed",
"time": "2015-08-19T20:59:18",
"component": "combo reconstruction",
"summary": "pybdt low test coverage",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "richman",
"type": "defect"
}
```
</p>
</details>
|
defect
|
pybdt low test coverage trac migrated from json status closed changetime ts description the pybdt test coverage of certain sections of the code is low please try to increase the test coverage as much as possible n nfor details please see n n reporter jtatar cc resolution fixed time component combo reconstruction summary pybdt low test coverage priority normal keywords milestone owner richman type defect
| 1
|
306,784
| 23,172,021,056
|
IssuesEvent
|
2022-07-30 21:39:41
|
bitsy-ai/printnanny-os
|
https://api.github.com/repos/bitsy-ai/printnanny-os
|
opened
|
Docs: How to Choose an edition of PrintNanny OS
|
documentation enhancement
|
**Describe your Request**
Document what's inside of each edition of PrintNanny OS
Link should be: https://docs.printnanny.ai/docs/quick-start/choose-os-edition/
|
1.0
|
Docs: How to Choose an edition of PrintNanny OS - **Describe your Request**
Document what's inside of each edition of PrintNanny OS
Link should be: https://docs.printnanny.ai/docs/quick-start/choose-os-edition/
|
non_defect
|
docs how to choose an edition of printnanny os describe your request document what s inside of each edition of printnanny os link should be
| 0
|
53,131
| 27,980,959,701
|
IssuesEvent
|
2023-03-26 05:59:41
|
jquery/learn.jquery.com
|
https://api.github.com/repos/jquery/learn.jquery.com
|
closed
|
Remove right-hand side specificity recommendation
|
Chapter: performance
|
In http://learn.jquery.com/performance/optimize-selectors/ you say:
"Be specific on the right-hand side of your selector, and less specific on the left."
I've opened a thread on SO: http://stackoverflow.com/questions/34923710/selectors-be-specific-on-right-hand-side
Two people so far have done benchmarks and the version where the right-hand side is more specific is always performing worse. Basically, in modern browsers supporting queryselectorall, as far as I'm aware, the Sizzle engine is not used for processing the selector. So basically, as far as I'm aware, in modern browsers, it's about how many operation there are to evaluate the result.
In `$( ".data .gonzalez" )`, there are 2 operations (1 for looking up CSS class results, 1 for tag results), while in `$( ".data td.gonzalez" )` there are 3 operations (2 for looking up CSS class results, 1 for tag results) which is why I suspect the second version is slower in most browsers.
Correct me if I'm wrong.
|
True
|
Remove right-hand side specificity recommendation - In http://learn.jquery.com/performance/optimize-selectors/ you say:
"Be specific on the right-hand side of your selector, and less specific on the left."
I've opened a thread on SO: http://stackoverflow.com/questions/34923710/selectors-be-specific-on-right-hand-side
Two people so far have done benchmarks and the version where the right-hand side is more specific is always performing worse. Basically, in modern browsers supporting queryselectorall, as far as I'm aware, the Sizzle engine is not used for processing the selector. So basically, as far as I'm aware, in modern browsers, it's about how many operation there are to evaluate the result.
In `$( ".data .gonzalez" )`, there are 2 operations (1 for looking up CSS class results, 1 for tag results), while in `$( ".data td.gonzalez" )` there are 3 operations (2 for looking up CSS class results, 1 for tag results) which is why I suspect the second version is slower in most browsers.
Correct me if I'm wrong.
|
non_defect
|
remove right hand side specificity recommendation in you say be specific on the right hand side of your selector and less specific on the left i ve opened a thread on so two people so far have done benchmarks and the version where the right hand side is more specific is always performing worse basically in modern browsers supporting queryselectorall as far as i m aware the sizzle engine is not used for processing the selector so basically as far as i m aware in modern browsers it s about how many operation there are to evaluate the result in data gonzalez there are operations for looking up css class results for tag results while in data td gonzalez there are operations for looking up css class results for tag results which is why i suspect the second version is slower in most browsers correct me if i m wrong
| 0
|
51,121
| 6,495,382,119
|
IssuesEvent
|
2017-08-22 04:48:52
|
ValerioLyndon/MAL-Public-List-Designs
|
https://api.github.com/repos/ValerioLyndon/MAL-Public-List-Designs
|
closed
|
Set nav bar to not be positioned badly at screen sizes smaller than 1050px
|
problem with design suggestion
|
It looks bad right now if you go below the max screen size.
Everything else is set to the left of the screen while the nav bar is set to center with **calc(50% - ###px)**.

|
1.0
|
Set nav bar to not be positioned badly at screen sizes smaller than 1050px - It looks bad right now if you go below the max screen size.
Everything else is set to the left of the screen while the nav bar is set to center with **calc(50% - ###px)**.

|
non_defect
|
set nav bar to not be positioned badly at screen sizes smaller than it looks bad right now if you go below the max screen size everything else is set to the left of the screen while the nav bar is set to center with calc px
| 0
|
35,467
| 7,749,012,317
|
IssuesEvent
|
2018-05-30 10:02:23
|
PowerDNS/pdns
|
https://api.github.com/repos/PowerDNS/pdns
|
closed
|
rec: actually support NSID
|
defect rec
|
- Program: Recursor
- Issue type: Bug report
### Short description
While our config description claims we do NSID, we do not.
### Environment
- Operating system: any
- Software version: 4.x, master
- Software source: any
### Steps to reproduce
1. Notice `# server-id Returned when queried for 'id.server' TXT or NSID, defaults to hostname` in config
2. start recursor
3. do +nsid query
### Expected behaviour
Response includes NSID.
### Actual behaviour
Response does not include NSID.
### Other information
NSID does work in auth.
<!--
Use the part below to file a feature request and delete the bug report part above.
-->
### Usecase
<!-- Tell what you're trying to achieve, without describing _what_ the requested feature should do -->
### Description
<!-- Describe as extensively as possible what you want the software to do -->
|
1.0
|
rec: actually support NSID - - Program: Recursor
- Issue type: Bug report
### Short description
While our config description claims we do NSID, we do not.
### Environment
- Operating system: any
- Software version: 4.x, master
- Software source: any
### Steps to reproduce
1. Notice `# server-id Returned when queried for 'id.server' TXT or NSID, defaults to hostname` in config
2. start recursor
3. do +nsid query
### Expected behaviour
Response includes NSID.
### Actual behaviour
Response does not include NSID.
### Other information
NSID does work in auth.
<!--
Use the part below to file a feature request and delete the bug report part above.
-->
### Usecase
<!-- Tell what you're trying to achieve, without describing _what_ the requested feature should do -->
### Description
<!-- Describe as extensively as possible what you want the software to do -->
|
defect
|
rec actually support nsid program recursor issue type bug report short description while our config description claims we do nsid we do not environment operating system any software version x master software source any steps to reproduce notice server id returned when queried for id server txt or nsid defaults to hostname in config start recursor do nsid query expected behaviour response includes nsid actual behaviour response does not include nsid other information nsid does work in auth use the part below to file a feature request and delete the bug report part above usecase description
| 1
|
18,062
| 3,022,280,450
|
IssuesEvent
|
2015-07-31 19:22:26
|
catmaid/CATMAID
|
https://api.github.com/repos/catmaid/CATMAID
|
closed
|
Resizing SVG overlay, doesn't refresh it when needed
|
difficulty: low priority: important status: done type: defect
|
When a widget is re-sized or closed and the SVG overlay gains more space, it isn't updated properly. No new node query happens, even if needed.
|
1.0
|
Resizing SVG overlay, doesn't refresh it when needed - When a widget is re-sized or closed and the SVG overlay gains more space, it isn't updated properly. No new node query happens, even if needed.
|
defect
|
resizing svg overlay doesn t refresh it when needed when a widget is re sized or closed and the svg overlay gains more space it isn t updated properly no new node query happens even if needed
| 1
|
192,933
| 15,362,377,621
|
IssuesEvent
|
2021-03-01 19:23:50
|
Topl/Bifrost
|
https://api.github.com/repos/Topl/Bifrost
|
closed
|
Write readme for gjallarhorn back-end folder
|
documentation
|
- overview of connection scenarios
- dedicated actor that communicates with Bifrost
- http server open on this port to listen to commands from front-end
- requirement to have bifrost running if you want gjallarhorn talking to it
|
1.0
|
Write readme for gjallarhorn back-end folder - - overview of connection scenarios
- dedicated actor that communicates with Bifrost
- http server open on this port to listen to commands from front-end
- requirement to have bifrost running if you want gjallarhorn talking to it
|
non_defect
|
write readme for gjallarhorn back end folder overview of connection scenarios dedicated actor that communicates with bifrost http server open on this port to listen to commands from front end requirement to have bifrost running if you want gjallarhorn talking to it
| 0
|
232,952
| 7,687,957,538
|
IssuesEvent
|
2018-05-17 07:56:09
|
EyeSeeTea/pictureapp
|
https://api.github.com/repos/EyeSeeTea/pictureapp
|
closed
|
Save form when app focus changes
|
CNM complexity - med (1-5hr) eReferrals priority - high ready type - maintenance
|
We're currently saving the form we're filling in when the screen switches off, so if the user turns it on again, the form is not lost. However, we're not doing it when the user changes to another app or press the Android home button.
|
1.0
|
Save form when app focus changes - We're currently saving the form we're filling in when the screen switches off, so if the user turns it on again, the form is not lost. However, we're not doing it when the user changes to another app or press the Android home button.
|
non_defect
|
save form when app focus changes we re currently saving the form we re filling in when the screen switches off so if the user turns it on again the form is not lost however we re not doing it when the user changes to another app or press the android home button
| 0
|
4,836
| 2,756,738,505
|
IssuesEvent
|
2015-04-27 10:24:42
|
adrianshort/sutton-council-websites
|
https://api.github.com/repos/adrianshort/sutton-council-websites
|
opened
|
Upload missing payments over £500 CSV files
|
2015-redesign
|
To here: http://www.sutton.gov.uk/downloads/download/453/expenditure_exceeding_500
Missing data is Aug 2010 to May 2014, excluding April 2014.
See also: https://github.com/adrianshort/sutton-council-payments-over-500-pounds/
|
1.0
|
Upload missing payments over £500 CSV files - To here: http://www.sutton.gov.uk/downloads/download/453/expenditure_exceeding_500
Missing data is Aug 2010 to May 2014, excluding April 2014.
See also: https://github.com/adrianshort/sutton-council-payments-over-500-pounds/
|
non_defect
|
upload missing payments over £ csv files to here missing data is aug to may excluding april see also
| 0
|
46,427
| 9,937,658,319
|
IssuesEvent
|
2019-07-02 22:43:26
|
dotnet/coreclr
|
https://api.github.com/repos/dotnet/coreclr
|
closed
|
Assert failure: HasPerInstInfo()
|
arch-x86 area-CodeGen os-windows
|
**Job:**
https://mc.dot.net/#/user/coreclr-outerloop-gcstress-extra/ci~2Fdotnet~2Fcoreclr~2Frefs~2Fheads~2Fmaster/test~2Ffunctional~2Fcli~2F/20190616.1/workItem/JIT/analysis/xunit/JIT_Performance._CodeQuality_BenchmarksGame_regex_redux_regex_redux_5_regex_redux_5_~2F_CodeQuality_BenchmarksGame_regex_redux_regex_redux_5_regex_redux_5_cmd
**Failed tests:**
JIT_Performance._CodeQuality_BenchmarksGame_regex_redux_regex_redux_5_regex_redux_5_._CodeQuality_BenchmarksGame_regex_redux_regex_redux_5_regex_redux_5_cmd
**Log:**
```
Assert failure(PID 4460 [0x0000116c], Thread: 3048 [0x0be8]): HasPerInstInfo()
<no module>! <no symbol> + 0x0 (0x00000000)
File: f:\\workspace\\_work\\1\\s\\src\\vm\\methodtable.h Line: 2955
Image: C:\\dotnetbuild\\work\\d612e967-3739-4ea0-bbcf-da077764d85b\\Payload\\CoreRun.exe
Return code: 1
Raw output file: C:\\dotnetbuild\\work\\d612e967-3739-4ea0-bbcf-da077764d85b\\Work\\e39edb3b-c408-480c-a752-84aca6598fdc\\Exec\\JIT\\Performance\\Reports\\JIT.Performance\\CodeQuality\\BenchmarksGame\\regex-redux\\regex-redux-5\\regex-redux-5.output.txt
Raw output:
BEGIN EXECUTION
\"C:\\dotnetbuild\\work\\d612e967-3739-4ea0-bbcf-da077764d85b\\Payload\\corerun.exe\" regex-redux-5.exe
Expected: 100
Actual: -1073740286
END EXECUTION - FAILED
FAILED
Test Harness Exitcode is : 1
To run the test:
> set CORE_ROOT=C:\\dotnetbuild\\work\\d612e967-3739-4ea0-bbcf-da077764d85b\\Payload
> C:\\dotnetbuild\\work\\d612e967-3739-4ea0-bbcf-da077764d85b\\Work\\e39edb3b-c408-480c-a752-84aca6598fdc\\Exec\\JIT\\Performance\\CodeQuality\\BenchmarksGame\\regex-redux\\regex-redux-5\\regex-redux-5.cmd
Expected: True
Actual: False
```
|
1.0
|
Assert failure: HasPerInstInfo() - **Job:**
https://mc.dot.net/#/user/coreclr-outerloop-gcstress-extra/ci~2Fdotnet~2Fcoreclr~2Frefs~2Fheads~2Fmaster/test~2Ffunctional~2Fcli~2F/20190616.1/workItem/JIT/analysis/xunit/JIT_Performance._CodeQuality_BenchmarksGame_regex_redux_regex_redux_5_regex_redux_5_~2F_CodeQuality_BenchmarksGame_regex_redux_regex_redux_5_regex_redux_5_cmd
**Failed tests:**
JIT_Performance._CodeQuality_BenchmarksGame_regex_redux_regex_redux_5_regex_redux_5_._CodeQuality_BenchmarksGame_regex_redux_regex_redux_5_regex_redux_5_cmd
**Log:**
```
Assert failure(PID 4460 [0x0000116c], Thread: 3048 [0x0be8]): HasPerInstInfo()
<no module>! <no symbol> + 0x0 (0x00000000)
File: f:\\workspace\\_work\\1\\s\\src\\vm\\methodtable.h Line: 2955
Image: C:\\dotnetbuild\\work\\d612e967-3739-4ea0-bbcf-da077764d85b\\Payload\\CoreRun.exe
Return code: 1
Raw output file: C:\\dotnetbuild\\work\\d612e967-3739-4ea0-bbcf-da077764d85b\\Work\\e39edb3b-c408-480c-a752-84aca6598fdc\\Exec\\JIT\\Performance\\Reports\\JIT.Performance\\CodeQuality\\BenchmarksGame\\regex-redux\\regex-redux-5\\regex-redux-5.output.txt
Raw output:
BEGIN EXECUTION
\"C:\\dotnetbuild\\work\\d612e967-3739-4ea0-bbcf-da077764d85b\\Payload\\corerun.exe\" regex-redux-5.exe
Expected: 100
Actual: -1073740286
END EXECUTION - FAILED
FAILED
Test Harness Exitcode is : 1
To run the test:
> set CORE_ROOT=C:\\dotnetbuild\\work\\d612e967-3739-4ea0-bbcf-da077764d85b\\Payload
> C:\\dotnetbuild\\work\\d612e967-3739-4ea0-bbcf-da077764d85b\\Work\\e39edb3b-c408-480c-a752-84aca6598fdc\\Exec\\JIT\\Performance\\CodeQuality\\BenchmarksGame\\regex-redux\\regex-redux-5\\regex-redux-5.cmd
Expected: True
Actual: False
```
|
non_defect
|
assert failure hasperinstinfo job failed tests jit performance codequality benchmarksgame regex redux regex redux regex redux codequality benchmarksgame regex redux regex redux regex redux cmd log assert failure pid thread hasperinstinfo file f workspace work s src vm methodtable h line image c dotnetbuild work bbcf payload corerun exe return code raw output file c dotnetbuild work bbcf work exec jit performance reports jit performance codequality benchmarksgame regex redux regex redux regex redux output txt raw output begin execution c dotnetbuild work bbcf payload corerun exe regex redux exe expected actual end execution failed failed test harness exitcode is to run the test set core root c dotnetbuild work bbcf payload c dotnetbuild work bbcf work exec jit performance codequality benchmarksgame regex redux regex redux regex redux cmd expected true actual false
| 0
|
33,006
| 6,993,943,432
|
IssuesEvent
|
2017-12-15 13:33:47
|
buildo/buildo.io
|
https://api.github.com/repos/buildo/buildo.io
|
closed
|
Adjust scss coding style
|
defect waiting for merge
|
## description
All the `.scss` files have errors (`style-lint` is probably not working in the build):

## how to reproduce
- {optional: describe steps to reproduce defect}
## specs
{optional: describe a possible fix for this defect, if not obvious}
## misc
{optional: other useful info}
|
1.0
|
Adjust scss coding style - ## description
All the `.scss` files have errors (`style-lint` is probably not working in the build):

## how to reproduce
- {optional: describe steps to reproduce defect}
## specs
{optional: describe a possible fix for this defect, if not obvious}
## misc
{optional: other useful info}
|
defect
|
adjust scss coding style description all the scss files have errors style lint is probably not working in the build how to reproduce optional describe steps to reproduce defect specs optional describe a possible fix for this defect if not obvious misc optional other useful info
| 1
|
23,304
| 4,006,282,017
|
IssuesEvent
|
2016-05-12 14:29:28
|
pavanramkumar/pyglmnet
|
https://api.github.com/repos/pavanramkumar/pyglmnet
|
closed
|
shape issue in poissonexp output
|
bug easy FIX tests
|
running README example, `.predict` has output with shape `(10, 10000, 1)` when using `distr=poissonexp`
for the other distributions has shape `(10, 1000)`
|
1.0
|
shape issue in poissonexp output - running README example, `.predict` has output with shape `(10, 10000, 1)` when using `distr=poissonexp`
for the other distributions has shape `(10, 1000)`
|
non_defect
|
shape issue in poissonexp output running readme example predict has output with shape when using distr poissonexp for the other distributions has shape
| 0
|
42,629
| 11,185,654,941
|
IssuesEvent
|
2020-01-01 04:32:07
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
Translate behavior sets `dirty` on contained property if the parent doesn't have the behavior
|
ORM behaviors defect
|
This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.8.4
* Platform and Target: PHP7.2.21
### What you did
To use an example from core unit tests, Articles belongs to Authors. Add Translate behavior to Authors, but not Articles. (Assume that in the actual use case, there's nothing worth translating in Articles.) Load an Article (or Articles; it doesn't seem to matter if you use `find` or `get`), containing Authors.
### What happened
The Article entity has the author property flagged as dirty.
### What you expected to happen
The author property should not be flagged as dirty.
If Articles has the Translate behavior on it, or if neither Articles nor Authors does, or if you use `loadInto` instead of `contain`, the author property is correctly not flagged.
This can be easily replicated in the core tests by commenting out `$table->addBehavior('Translate', ['fields' => ['title', 'body']]);` (and therefore also `$table->setLocale('eng');`) in `Cake\Test\TestCase\ORM\Behavior\TranslateBehaviorTest::testFindSingleLocaleBelongstoLoadInto`.
|
1.0
|
Translate behavior sets `dirty` on contained property if the parent doesn't have the behavior - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.8.4
* Platform and Target: PHP7.2.21
### What you did
To use an example from core unit tests, Articles belongs to Authors. Add Translate behavior to Authors, but not Articles. (Assume that in the actual use case, there's nothing worth translating in Articles.) Load an Article (or Articles; it doesn't seem to matter if you use `find` or `get`), containing Authors.
### What happened
The Article entity has the author property flagged as dirty.
### What you expected to happen
The author property should not be flagged as dirty.
If Articles has the Translate behavior on it, or if neither Articles nor Authors does, or if you use `loadInto` instead of `contain`, the author property is correctly not flagged.
This can be easily replicated in the core tests by commenting out `$table->addBehavior('Translate', ['fields' => ['title', 'body']]);` (and therefore also `$table->setLocale('eng');`) in `Cake\Test\TestCase\ORM\Behavior\TranslateBehaviorTest::testFindSingleLocaleBelongstoLoadInto`.
|
defect
|
translate behavior sets dirty on contained property if the parent doesn t have the behavior this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target what you did to use an example from core unit tests articles belongs to authors add translate behavior to authors but not articles assume that in the actual use case there s nothing worth translating in articles load an article or articles it doesn t seem to matter if you use find or get containing authors what happened the article entity has the author property flagged as dirty what you expected to happen the author property should not be flagged as dirty if articles has the translate behavior on it or if neither articles nor authors does or if you use loadinto instead of contain the author property is correctly not flagged this can be easily replicated in the core tests by commenting out table addbehavior translate and therefore also table setlocale eng in cake test testcase orm behavior translatebehaviortest testfindsinglelocalebelongstoloadinto
| 1
|
17,249
| 2,986,719,366
|
IssuesEvent
|
2015-07-20 07:01:34
|
contao/core
|
https://api.github.com/repos/contao/core
|
closed
|
Minify the markup must not remove all spaces, but collapse instead
|
defect
|
Contao settings allow to "Minify the markup". This removes whitespaces between HTML elements. But that alters the structure of the document. Therefore Contao should collapse whitespaces instead of removing them altogether.
In the examples below decrease the width of the "Result" pane until the menu doesn't fit. Because the HTML changed the results are different between the two examples. You can see the difference instantly because the spaces are visible in the first example. Tested with Chrome and Firefox.
Example with whitespaces:
https://jsfiddle.net/nkqwbfyo/
Example without whitespaces:
https://jsfiddle.net/nkqwbfyo/1/
|
1.0
|
Minify the markup must not remove all spaces, but collapse instead - Contao settings allow to "Minify the markup". This removes whitespaces between HTML elements. But that alters the structure of the document. Therefore Contao should collapse whitespaces instead of removing them altogether.
In the examples below decrease the width of the "Result" pane until the menu doesn't fit. Because the HTML changed the results are different between the two examples. You can see the difference instantly because the spaces are visible in the first example. Tested with Chrome and Firefox.
Example with whitespaces:
https://jsfiddle.net/nkqwbfyo/
Example without whitespaces:
https://jsfiddle.net/nkqwbfyo/1/
|
defect
|
minify the markup must not remove all spaces but collapse instead contao settings allow to minify the markup this removes whitespaces between html elements but that alters the structure of the document therefore contao should collapse whitespaces instead of removing them altogether in the examples below decrease the width of the result pane until the menu doesn t fit because the html changed the results are different between the two examples you can see the difference instantly because the spaces are visible in the first example tested with chrome and firefox example with whitespaces example without whitespaces
| 1
|
23,980
| 3,882,617,502
|
IssuesEvent
|
2016-04-13 10:38:07
|
bridgedotnet/Bridge
|
https://api.github.com/repos/bridgedotnet/Bridge
|
closed
|
Structs as generic arguments are treated like classes
|
defect in progress
|
If you use a struct as a generic argument, it will be passed by reference everywhere and the $clone method will not be inserted anywhere.
### Expected
0
5
### Actual
5
5
### Steps To Reproduce
[Live](http://live.bridge.net/#f01f1e291e860821a6ab1b0913c888c7)
```csharp
public struct Vector2
{
public float x;
public float y;
}
public class App
{
[Ready]
public static void Main()
{
List<Vector2> list = new List<Vector2>();
list.Add(new Vector2() { x = 0.0f, y = 1.0f });
Vector2 vec = list[0];
vec.x = 5.0f;
Window.Alert(list[0].x.ToString()); // Oops.
Window.Alert(vec.x.ToString());
}
}
```
|
1.0
|
Structs as generic arguments are treated like classes - If you use a struct as a generic argument, it will be passed by reference everywhere and the $clone method will not be inserted anywhere.
### Expected
0
5
### Actual
5
5
### Steps To Reproduce
[Live](http://live.bridge.net/#f01f1e291e860821a6ab1b0913c888c7)
```csharp
public struct Vector2
{
public float x;
public float y;
}
public class App
{
[Ready]
public static void Main()
{
List<Vector2> list = new List<Vector2>();
list.Add(new Vector2() { x = 0.0f, y = 1.0f });
Vector2 vec = list[0];
vec.x = 5.0f;
Window.Alert(list[0].x.ToString()); // Oops.
Window.Alert(vec.x.ToString());
}
}
```
|
defect
|
structs as generic arguments are treated like classes if you use a struct as a generic argument it will be passed by reference everywhere and the clone method will not be inserted anywhere expected actual steps to reproduce csharp public struct public float x public float y public class app public static void main list list new list list add new x y vec list vec x window alert list x tostring oops window alert vec x tostring
| 1
|
390,185
| 26,852,006,999
|
IssuesEvent
|
2023-02-03 11:48:42
|
insightsengineering/formatters
|
https://api.github.com/repos/insightsengineering/formatters
|
closed
|
Add example(s) to `make_row_df`
|
documentation sme
|
Please add example(s) to the `make_row_df` function. This is the function that is being referenced in other places as well so it's important to have a high quality docs in there
|
1.0
|
Add example(s) to `make_row_df` - Please add example(s) to the `make_row_df` function. This is the function that is being referenced in other places as well so it's important to have a high quality docs in there
|
non_defect
|
add example s to make row df please add example s to the make row df function this is the function that is being referenced in other places as well so it s important to have a high quality docs in there
| 0
|
51,710
| 13,211,295,370
|
IssuesEvent
|
2020-08-15 22:06:54
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
dataclasses - mask issue in 2011 simulation (Trac #904)
|
Incomplete Migration Migrated from Trac combo core defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/904">https://code.icecube.wisc.edu/projects/icecube/ticket/904</a>, reported by david.schultzand owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-08-14T08:07:52",
"_ts": "1439539672935233",
"description": "In the file /data/sim/IceCube/2011/filtered/level2/CORSIKA-in-ice/10649/00000-00999/Level2_IC86.2011_corsika.010649.000561.i3.bz2, EventID 2142, there is a mask error.\n\n{{{\nIn [3]: for fr in dataio.I3File('/data/sim/IceCube/2011/filtered/level2/CORSIKA-in-ice/10649/00000-00999/Level2_IC86.2011_corsika.010649.000561.i3.bz2'):\n if 'MaskedOfflinePulses' in fr:\n fr['MaskedOfflinePulses'].apply(fr) \n ...: \nFATAL (I3RecoPulseSeriesMapMask): The mask for OM(56,46) has 1 entries, but source pulse vector has 10 entries! (I3MapOMKeyMask.cxx:480 in boost::shared_ptr<const I3Map<OMKey, std::vector<I3RecoPulse, std::allocator<I3RecoPulse> > > > I3RecoPulseSeriesMapMask::Apply(const I3Frame&) const)\n---------------------------------------------------------------------------\nRuntimeError Traceback (most recent call last)\n<ipython-input-3-85e79e1bedde> in <module>()\n 1 for fr in dataio.I3File('/data/sim/IceCube/2011/filtered/level2/CORSIKA-in-ice/10649/00000-00999/Level2_IC86.2011_corsika.010649.000561.i3.bz2'):\n 2 if 'MaskedOfflinePulses' in fr:\n----> 3 fr['MaskedOfflinePulses'].apply(fr)\n 4 \n\nRuntimeError: The mask for OM(56,46) has 1 entries, but source pulse vector has 10 entries! (in boost::shared_ptr<const I3Map<OMKey, std::vector<I3RecoPulse, std::allocator<I3RecoPulse> > > > I3RecoPulseSeriesMapMask::Apply(const I3Frame&) const)\n}}}\n\nThe issue is that the I3SuperDST RecoPulseSeriesMap on OMKey(56,46,0) has more entries than the OfflinePulses mask has entries. This is an issue in creation of the mask itself.\n\nThe software in use is icerec IC2011-L2_V12-08-00_IceSim4compat_V4.\n\nThis changeset might be why this doesn't occur anymore: [96060/IceCube]\n\nSuggestions:\n* Test recreating the pulse mask with and without that changeset.\n* Decide what to do when reading bad masks in the future. Options:\n * remove extra pulses\n * keep extra pulses\n * continue throwing an error, but make it clearer",
"reporter": "david.schultz",
"cc": "jvansanten",
"resolution": "fixed",
"time": "2015-03-24T19:54:43",
"component": "combo core",
"summary": "dataclasses - mask issue in 2011 simulation",
"priority": "blocker",
"keywords": "dataclasses",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
dataclasses - mask issue in 2011 simulation (Trac #904) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/904">https://code.icecube.wisc.edu/projects/icecube/ticket/904</a>, reported by david.schultzand owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-08-14T08:07:52",
"_ts": "1439539672935233",
"description": "In the file /data/sim/IceCube/2011/filtered/level2/CORSIKA-in-ice/10649/00000-00999/Level2_IC86.2011_corsika.010649.000561.i3.bz2, EventID 2142, there is a mask error.\n\n{{{\nIn [3]: for fr in dataio.I3File('/data/sim/IceCube/2011/filtered/level2/CORSIKA-in-ice/10649/00000-00999/Level2_IC86.2011_corsika.010649.000561.i3.bz2'):\n if 'MaskedOfflinePulses' in fr:\n fr['MaskedOfflinePulses'].apply(fr) \n ...: \nFATAL (I3RecoPulseSeriesMapMask): The mask for OM(56,46) has 1 entries, but source pulse vector has 10 entries! (I3MapOMKeyMask.cxx:480 in boost::shared_ptr<const I3Map<OMKey, std::vector<I3RecoPulse, std::allocator<I3RecoPulse> > > > I3RecoPulseSeriesMapMask::Apply(const I3Frame&) const)\n---------------------------------------------------------------------------\nRuntimeError Traceback (most recent call last)\n<ipython-input-3-85e79e1bedde> in <module>()\n 1 for fr in dataio.I3File('/data/sim/IceCube/2011/filtered/level2/CORSIKA-in-ice/10649/00000-00999/Level2_IC86.2011_corsika.010649.000561.i3.bz2'):\n 2 if 'MaskedOfflinePulses' in fr:\n----> 3 fr['MaskedOfflinePulses'].apply(fr)\n 4 \n\nRuntimeError: The mask for OM(56,46) has 1 entries, but source pulse vector has 10 entries! (in boost::shared_ptr<const I3Map<OMKey, std::vector<I3RecoPulse, std::allocator<I3RecoPulse> > > > I3RecoPulseSeriesMapMask::Apply(const I3Frame&) const)\n}}}\n\nThe issue is that the I3SuperDST RecoPulseSeriesMap on OMKey(56,46,0) has more entries than the OfflinePulses mask has entries. This is an issue in creation of the mask itself.\n\nThe software in use is icerec IC2011-L2_V12-08-00_IceSim4compat_V4.\n\nThis changeset might be why this doesn't occur anymore: [96060/IceCube]\n\nSuggestions:\n* Test recreating the pulse mask with and without that changeset.\n* Decide what to do when reading bad masks in the future. Options:\n * remove extra pulses\n * keep extra pulses\n * continue throwing an error, but make it clearer",
"reporter": "david.schultz",
"cc": "jvansanten",
"resolution": "fixed",
"time": "2015-03-24T19:54:43",
"component": "combo core",
"summary": "dataclasses - mask issue in 2011 simulation",
"priority": "blocker",
"keywords": "dataclasses",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
defect
|
dataclasses mask issue in simulation trac migrated from json status closed changetime ts description in the file data sim icecube filtered corsika in ice corsika eventid there is a mask error n n nin for fr in dataio data sim icecube filtered corsika in ice corsika n if maskedofflinepulses in fr n fr apply fr n nfatal the mask for om has entries but source pulse vector has entries cxx in boost shared ptr apply const const n nruntimeerror traceback most recent call last n in n for fr in dataio data sim icecube filtered corsika in ice corsika n if maskedofflinepulses in fr n fr apply fr n n nruntimeerror the mask for om has entries but source pulse vector has entries in boost shared ptr apply const const n n nthe issue is that the recopulseseriesmap on omkey has more entries than the offlinepulses mask has entries this is an issue in creation of the mask itself n nthe software in use is icerec n nthis changeset might be why this doesn t occur anymore n nsuggestions n test recreating the pulse mask with and without that changeset n decide what to do when reading bad masks in the future options n remove extra pulses n keep extra pulses n continue throwing an error but make it clearer reporter david schultz cc jvansanten resolution fixed time component combo core summary dataclasses mask issue in simulation priority blocker keywords dataclasses milestone owner jvansanten type defect
| 1
|
11,077
| 3,456,328,332
|
IssuesEvent
|
2015-12-18 00:35:28
|
space-race/mc-core
|
https://api.github.com/repos/space-race/mc-core
|
opened
|
Health Checks
|
Discussion Extensions Feature / Functionality Reference for Documentation
|
A user should be able to connect external health/monitoring services as "health check" in Mission Control.
- [ ] dashboard: show monitoring alerts
- [ ] health: create a new health check
|
1.0
|
Health Checks - A user should be able to connect external health/monitoring services as "health check" in Mission Control.
- [ ] dashboard: show monitoring alerts
- [ ] health: create a new health check
|
non_defect
|
health checks a user should be able to connect external health monitoring services as health check in mission control dashboard show monitoring alerts health create a new health check
| 0
|
73,154
| 24,475,392,811
|
IssuesEvent
|
2022-10-08 05:05:05
|
zealdocs/zeal
|
https://api.github.com/repos/zealdocs/zeal
|
closed
|
Android Docset Navigation not working
|
type/defect scope/misc/docsets resolution/upstream-problem
|
The Android docs have examples in Kotlin and Java which you can switch between by clicking a button.
In Zeal it's showing the Kotlin examples by default and switching to the Java examples doesn't work unfortunately. When I opened the html downloaded by Zeal in a Browser it shows that some Javascript sources could not be loaded, I was thinking that might be the cause of the problem? But I'm not sure.
You can see this in the MediaPlayer Overview document for example.
|
1.0
|
Android Docset Navigation not working - The Android docs have examples in Kotlin and Java which you can switch between by clicking a button.
In Zeal it's showing the Kotlin examples by default and switching to the Java examples doesn't work unfortunately. When I opened the html downloaded by Zeal in a Browser it shows that some Javascript sources could not be loaded, I was thinking that might be the cause of the problem? But I'm not sure.
You can see this in the MediaPlayer Overview document for example.
|
defect
|
android docset navigation not working the android docs have examples in kotlin and java which you can switch between by clicking a button in zeal it s showing the kotlin examples by default and switching to the java examples doesn t work unfortunately when i opened the html downloaded by zeal in a browser it shows that some javascript sources could not be loaded i was thinking that might be the cause of the problem but i m not sure you can see this in the mediaplayer overview document for example
| 1
|
117,079
| 17,407,950,214
|
IssuesEvent
|
2021-08-03 08:37:09
|
elikkatzgit/TestingPOM
|
https://api.github.com/repos/elikkatzgit/TestingPOM
|
closed
|
CVE-2018-19361 (High) detected in jackson-databind-2.7.2.jar - autoclosed
|
security vulnerability
|
## CVE-2018-19361 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/TestingPOM/commit/349fffeed7cf25f2cf5b8b6a05b5e4367130406e">349fffeed7cf25f2cf5b8b6a05b5e4367130406e</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the openjpa class from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19361>CVE-2018-19361</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19361">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19361</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.8</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.2","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.7.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.8"}],"vulnerabilityIdentifier":"CVE-2018-19361","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the openjpa class from polymorphic deserialization.","vulnerabilityUrl":"https://cve.mitre.org/cgi-bin/cvename.cgi?name\u003dCVE-2018-19361","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2018-19361 (High) detected in jackson-databind-2.7.2.jar - autoclosed - ## CVE-2018-19361 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/TestingPOM/commit/349fffeed7cf25f2cf5b8b6a05b5e4367130406e">349fffeed7cf25f2cf5b8b6a05b5e4367130406e</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the openjpa class from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19361>CVE-2018-19361</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19361">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19361</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.8</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.2","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.7.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.8"}],"vulnerabilityIdentifier":"CVE-2018-19361","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the openjpa class from polymorphic deserialization.","vulnerabilityUrl":"https://cve.mitre.org/cgi-bin/cvename.cgi?name\u003dCVE-2018-19361","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before might allow attackers to have unspecified impact by leveraging failure to block the openjpa class from polymorphic deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before might allow attackers to have unspecified impact by leveraging failure to block the openjpa class from polymorphic deserialization vulnerabilityurl
| 0
|
351,018
| 25,010,313,362
|
IssuesEvent
|
2022-11-03 14:49:32
|
ArctosDB/documentation-wiki
|
https://api.github.com/repos/ArctosDB/documentation-wiki
|
closed
|
How to create labels - HTML
|
Documentation Update
|
Now that we are transitioning away from CFR reports and using HTML tools, this page of the handbook is outdated:
https://handbook.arctosdb.org/how_to/How-to-Create-Labels.html
Does anyone have any resources or templates to get started with the new process that they could share?
Tagging @lkvoong, as suggested on Arctos report printer page
I started using https://wordtohtml.net/ to convert my labels to html, but I'm not sure how to get it to pull directly from Arctos once I have it formatted or how to print multiple labels per page.
Any suggestions on getting started or what resources to use are greatly appreciated. Thanks!
|
1.0
|
How to create labels - HTML - Now that we are transitioning away from CFR reports and using HTML tools, this page of the handbook is outdated:
https://handbook.arctosdb.org/how_to/How-to-Create-Labels.html
Does anyone have any resources or templates to get started with the new process that they could share?
Tagging @lkvoong, as suggested on Arctos report printer page
I started using https://wordtohtml.net/ to convert my labels to html, but I'm not sure how to get it to pull directly from Arctos once I have it formatted or how to print multiple labels per page.
Any suggestions on getting started or what resources to use are greatly appreciated. Thanks!
|
non_defect
|
how to create labels html now that we are transitioning away from cfr reports and using html tools this page of the handbook is outdated does anyone have any resources or templates to get started with the new process that they could share tagging lkvoong as suggested on arctos report printer page i started using to convert my labels to html but i m not sure how to get it to pull directly from arctos once i have it formatted or how to print multiple labels per page any suggestions on getting started or what resources to use are greatly appreciated thanks
| 0
|
136,797
| 30,592,013,417
|
IssuesEvent
|
2023-07-21 17:56:33
|
priyankarpal/ProjectsHut
|
https://api.github.com/repos/priyankarpal/ProjectsHut
|
closed
|
Bug: Copy functionality in Share profile
|
🛠 goal: fix 🏁 status: ready for dev gssoc23 level2 💻 aspect: code
|
### What happened?
Clicking on "Copy" in the Share profile feature does not successfully copy the link to the clipboard.
### Add screenshots
https://github.com/priyankarpal/ProjectsHut/assets/100708686/8da7eca5-24b0-427b-b4ca-ddac0b90a744
### What browsers are you seeing the problem on?
Microsoft Edge
### Record
- [X] I agree to follow this project's Code of Conduct
- [X] I'm a GSSoC'23 contributor
- [X] I want to work on this issue
|
1.0
|
Bug: Copy functionality in Share profile - ### What happened?
Clicking on "Copy" in the Share profile feature does not successfully copy the link to the clipboard.
### Add screenshots
https://github.com/priyankarpal/ProjectsHut/assets/100708686/8da7eca5-24b0-427b-b4ca-ddac0b90a744
### What browsers are you seeing the problem on?
Microsoft Edge
### Record
- [X] I agree to follow this project's Code of Conduct
- [X] I'm a GSSoC'23 contributor
- [X] I want to work on this issue
|
non_defect
|
bug copy functionality in share profile what happened clicking on copy in the share profile feature does not successfully copy the link to the clipboard add screenshots what browsers are you seeing the problem on microsoft edge record i agree to follow this project s code of conduct i m a gssoc contributor i want to work on this issue
| 0
|
199,011
| 15,020,480,840
|
IssuesEvent
|
2021-02-01 14:45:49
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
sql/tests: TestClusterID failed
|
C-test-failure O-robot branch-master
|
[(sql/tests).TestClusterID failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2150997&tab=buildLog) on [master@4f05bc8199a1a8825ad8ddfe5935e04957a62b2d](https://github.com/cockroachdb/cockroach/commits/4f05bc8199a1a8825ad8ddfe5935e04957a62b2d):
```
=== RUN TestClusterID
TestClusterID: test_log_scope.go:85: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestClusterID859317451
TestClusterID: test_log_scope.go:58: use -show-logs to present logs inline
ERROR: signal: segmentation fault (core dumped)
39 runs completed, 1 failures, over 54m14s
context canceled
```
<details><summary>More</summary><p>
Parameters:
- TAGS=
- GOFLAGS=-race -parallel=2
```
make stressrace TESTS=TestClusterID PKG=./pkg/sql/tests TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestClusterID.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
1.0
|
sql/tests: TestClusterID failed - [(sql/tests).TestClusterID failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2150997&tab=buildLog) on [master@4f05bc8199a1a8825ad8ddfe5935e04957a62b2d](https://github.com/cockroachdb/cockroach/commits/4f05bc8199a1a8825ad8ddfe5935e04957a62b2d):
```
=== RUN TestClusterID
TestClusterID: test_log_scope.go:85: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestClusterID859317451
TestClusterID: test_log_scope.go:58: use -show-logs to present logs inline
ERROR: signal: segmentation fault (core dumped)
39 runs completed, 1 failures, over 54m14s
context canceled
```
<details><summary>More</summary><p>
Parameters:
- TAGS=
- GOFLAGS=-race -parallel=2
```
make stressrace TESTS=TestClusterID PKG=./pkg/sql/tests TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestClusterID.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
non_defect
|
sql tests testclusterid failed on run testclusterid testclusterid test log scope go test logs captured to go src github com cockroachdb cockroach artifacts testclusterid test log scope go use show logs to present logs inline error signal segmentation fault core dumped runs completed failures over context canceled more parameters tags goflags race parallel make stressrace tests testclusterid pkg pkg sql tests testtimeout stressflags timeout powered by
| 0
|
62,670
| 17,116,692,136
|
IssuesEvent
|
2021-07-11 13:59:35
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
opened
|
PANIC: blkptr has invalid CHECKSUM 0
|
Status: Triage Needed Type: Defect
|
I have seen a few stale issues with similar title, however my stack traces look somewhat different, so I'm opening a new issue. I'm trying to figure out if I have a hardware problem or if my pool is actually borked, possibly caused by a bug.
### System information
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | 20.04, 21.04
Linux Kernel | 5.8.0-55-generic, 5.11.0-22-generic
Architecture | AMD64
ZFS Version | tracking master, 2.0.2-1ubuntu5, 2.0.5
SPL Version | tracking master, 2.0.2-1ubuntu5, 2.0.5
### Describe the problem you're observing
The system functions purely as a backup receiving sends and deleting old snapshots. After 1-10 days of running, the system panics while receiving with strack trace variations of "blkptr has invalid CHECKSUM 0". Subsequent scrubs comes back clean.
### Describe how to reproduce the problem
I can reproduce by leaving the system receiving backups for a few days. I don't know how to reproduce on another system.
### Include any warning/errors/backtraces from the system logs
- The first two traces are from Ubuntu 20.04 with zfs compiled from master at the time
- The next three are from Ubuntu 21.04 with the provided packages for zfs 2.0.2
- The last one is from Ubuntu 21.04 with packages compiled from the 2.0.5 to test if the changes from #12054 have any effect.
```
Jun 20 01:52:37 ubackup kernel: [482903.384554] PANIC: tank: blkptr at 00000000dbbffcf1 has invalid CHECKSUM 0
Jun 20 01:52:37 ubackup kernel: [482903.384572] Showing stack for process 3425
Jun 20 01:52:37 ubackup kernel: [482903.384580] CPU: 1 PID: 3425 Comm: z_wr_iss Tainted: P OE 5.8.0-55-generic #62~20.04.1-Ubuntu
Jun 20 01:52:37 ubackup kernel: [482903.384582] Hardware name: Gigabyte Technology Co., Ltd. GA-990FXA-UD3/GA-990FXA-UD3, BIOS FB 10/13/2011
Jun 20 01:52:37 ubackup kernel: [482903.384584] Call Trace:
Jun 20 01:52:37 ubackup kernel: [482903.384598] dump_stack+0x74/0x92
Jun 20 01:52:37 ubackup kernel: [482903.384625] spl_dumpstack+0x29/0x2b [spl]
Jun 20 01:52:37 ubackup kernel: [482903.384645] vcmn_err.cold+0x60/0x94 [spl]
Jun 20 01:52:37 ubackup kernel: [482903.384654] ? __blk_mq_sched_dispatch_requests+0x10e/0x170
Jun 20 01:52:37 ubackup kernel: [482903.384662] ? __blk_mq_run_hw_queue+0x5a/0x110
Jun 20 01:52:37 ubackup kernel: [482903.384669] ? ptr_to_id+0xbe/0x220
Jun 20 01:52:37 ubackup kernel: [482903.384916] zfs_panic_recover+0x6f/0x90 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.385130] zfs_blkptr_verify_log+0x94/0x100 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.385342] ? vdev_disk_io_start+0x49f/0x8e0 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.385351] ? _cond_resched+0x19/0x30
Jun 20 01:52:37 ubackup kernel: [482903.385356] ? mutex_lock+0x13/0x40
Jun 20 01:52:37 ubackup kernel: [482903.385568] ? zio_wait_for_children+0x8e/0xd0 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.385780] zfs_blkptr_verify+0x3c9/0x480 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.385993] zio_free+0x27/0x100 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.386191] dsl_free+0x11/0x20 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.386380] dsl_dataset_block_kill+0x4bf/0x4f0 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.386386] ? down_write+0x13/0x50
Jun 20 01:52:37 ubackup kernel: [482903.386565] dbuf_write_done+0x1b3/0x200 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.386740] arc_write_done+0x8f/0x410 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.386953] zio_done+0x407/0x1050 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.387166] zio_execute+0x93/0xf0 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.387185] taskq_thread+0x2fb/0x510 [spl]
Jun 20 01:52:37 ubackup kernel: [482903.387193] ? wake_up_q+0xa0/0xa0
Jun 20 01:52:37 ubackup kernel: [482903.387406] ? zio_taskq_member.isra.0.constprop.0+0x60/0x60 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.387413] kthread+0x114/0x150
Jun 20 01:52:37 ubackup kernel: [482903.387431] ? task_done+0xb0/0xb0 [spl]
Jun 20 01:52:37 ubackup kernel: [482903.387434] ? kthread_park+0x90/0x90
Jun 20 01:52:37 ubackup kernel: [482903.387440] ret_from_fork+0x22/0x30
Jun 25 01:43:58 ubackup kernel: [367708.395650] PANIC: tank: blkptr at 000000004322d8f5 has invalid CHECKSUM 0
Jun 25 01:43:58 ubackup kernel: [367708.395659] Showing stack for process 3381
Jun 25 01:43:58 ubackup kernel: [367708.395663] CPU: 1 PID: 3381 Comm: dp_sync_taskq Tainted: P OE 5.8.0-55-generic #62~20.04.1-Ubuntu
Jun 25 01:43:58 ubackup kernel: [367708.395664] Hardware name: Gigabyte Technology Co., Ltd. GA-990FXA-UD3/GA-990FXA-UD3, BIOS FB 10/13/2011
Jun 25 01:43:58 ubackup kernel: [367708.395664] Call Trace:
Jun 25 01:43:58 ubackup kernel: [367708.395672] dump_stack+0x74/0x92
Jun 25 01:43:58 ubackup kernel: [367708.395684] spl_dumpstack+0x29/0x2b [spl]
Jun 25 01:43:58 ubackup kernel: [367708.395691] vcmn_err.cold+0x60/0x94 [spl]
Jun 25 01:43:58 ubackup kernel: [367708.395698] ? spl_kmem_cache_alloc+0xa9/0x7d0 [spl]
Jun 25 01:43:58 ubackup kernel: [367708.395701] ? ptr_to_id+0xbe/0x220
Jun 25 01:43:58 ubackup kernel: [367708.395826] zfs_panic_recover+0x6f/0x90 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.395902] zfs_blkptr_verify_log+0x94/0x100 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.395963] ? arc_hdr_destroy+0x200/0x200 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.395967] ? try_to_wake_up+0x66/0x540
Jun 25 01:43:58 ubackup kernel: [367708.396028] ? dbuf_issue_final_prefetch+0xd0/0xd0 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396031] ? default_wake_function+0x1a/0x30
Jun 25 01:43:58 ubackup kernel: [367708.396032] ? __wake_up_common+0x7e/0x140
Jun 25 01:43:58 ubackup kernel: [367708.396034] ? __wake_up_common_lock+0x8a/0xc0
Jun 25 01:43:58 ubackup kernel: [367708.396108] zfs_blkptr_verify+0x3c9/0x480 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396182] zio_free+0x27/0x100 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396251] dsl_free+0x11/0x20 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396317] dsl_dataset_block_kill+0x4bf/0x4f0 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396382] free_blocks+0xea/0x1d0 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396456] ? zio_nowait+0xc1/0x150 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396521] dnode_sync_free_range+0x23c/0x270 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396583] ? dbuf_sync_leaf+0x23b/0x500 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396649] ? free_children+0x3d0/0x3d0 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396718] range_tree_walk+0x118/0x1f0 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396784] dnode_sync+0x2e8/0xa90 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396787] ? __switch_to+0x157/0x450
Jun 25 01:43:58 ubackup kernel: [367708.396789] ? _cond_resched+0x19/0x30
Jun 25 01:43:58 ubackup kernel: [367708.396791] ? mutex_lock+0x13/0x40
Jun 25 01:43:58 ubackup kernel: [367708.396855] sync_dnodes_task+0x79/0xb0 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396861] taskq_thread+0x2fb/0x510 [spl]
Jun 25 01:43:58 ubackup kernel: [367708.396864] ? wake_up_q+0xa0/0xa0
Jun 25 01:43:58 ubackup kernel: [367708.396866] kthread+0x114/0x150
Jun 25 01:43:58 ubackup kernel: [367708.396873] ? task_done+0xb0/0xb0 [spl]
Jun 25 01:43:58 ubackup kernel: [367708.396874] ? kthread_park+0x90/0x90
Jun 25 01:43:58 ubackup kernel: [367708.396876] ret_from_fork+0x22/0x30
Jul 1 02:01:26 ubackup kernel: [443020.371564] PANIC: tank: blkptr at 000000002c351e10 has invalid CHECKSUM 0
Jul 1 02:01:26 ubackup kernel: [443020.371590] Showing stack for process 3346
Jul 1 02:01:26 ubackup kernel: [443020.371597] CPU: 1 PID: 3346 Comm: z_wr_iss Tainted: P OE 5.11.0-22-generic #23~20.04.1-Ubuntu
Jul 1 02:01:26 ubackup kernel: [443020.371608] Hardware name: Gigabyte Technology Co., Ltd. GA-990FXA-UD3/GA-990FXA-UD3, BIOS FB 10/13/2011
Jul 1 02:01:26 ubackup kernel: [443020.371613] Call Trace:
Jul 1 02:01:26 ubackup kernel: [443020.371622] dump_stack+0x74/0x92
Jul 1 02:01:26 ubackup kernel: [443020.371640] spl_dumpstack+0x29/0x2b [spl]
Jul 1 02:01:26 ubackup kernel: [443020.371683] vcmn_err.cold+0x60/0x94 [spl]
Jul 1 02:01:26 ubackup kernel: [443020.371722] ? ptr_to_id+0xbe/0x220
Jul 1 02:01:26 ubackup kernel: [443020.371739] zfs_panic_recover+0x6f/0x90 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.372186] zfs_blkptr_verify_log+0x94/0x100 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.372613] ? vdev_disk_io_start+0x4a1/0x8e0 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.373033] ? _cond_resched+0x19/0x30
Jul 1 02:01:26 ubackup kernel: [443020.373045] ? mutex_lock+0x13/0x40
Jul 1 02:01:26 ubackup kernel: [443020.373052] ? zio_wait_for_children+0x8e/0xd0 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.373476] zfs_blkptr_verify+0x3c9/0x480 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.373902] zio_free+0x27/0x100 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.374326] dsl_free+0x11/0x20 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.374719] dsl_dataset_block_kill+0x4bf/0x4f0 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.375097] ? down_write+0x13/0x50
Jul 1 02:01:26 ubackup kernel: [443020.375106] dbuf_write_done+0x1b3/0x200 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.375463] arc_write_done+0x8f/0x410 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.375812] zio_done+0x407/0x1050 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.376237] zio_execute+0x93/0xf0 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.376661] taskq_thread+0x2fb/0x510 [spl]
Jul 1 02:01:26 ubackup kernel: [443020.376698] ? wake_up_q+0xa0/0xa0
Jul 1 02:01:26 ubackup kernel: [443020.376710] ? zio_taskq_member.isra.0.constprop.0+0x60/0x60 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.377136] kthread+0x114/0x150
Jul 1 02:01:26 ubackup kernel: [443020.377146] ? task_done+0xb0/0xb0 [spl]
Jul 1 02:01:26 ubackup kernel: [443020.377181] ? kthread_park+0x90/0x90
Jul 1 02:01:26 ubackup kernel: [443020.377190] ret_from_fork+0x22/0x30
Jul 2 12:41:32 ubackup kernel: [ 7228.720883] PANIC: tank: blkptr at 0000000084e788bd has invalid CHECKSUM 0
Jul 2 12:41:32 ubackup kernel: [ 7228.720901] Showing stack for process 4218
Jul 2 12:41:32 ubackup kernel: [ 7228.720904] CPU: 1 PID: 4218 Comm: z_wr_iss Tainted: P O 5.11.0-22-generic #23-Ubuntu
Jul 2 12:41:32 ubackup kernel: [ 7228.720907] Hardware name: Gigabyte Technology Co., Ltd. GA-990FXA-UD3/GA-990FXA-UD3, BIOS FB 10/13/2011
Jul 2 12:41:32 ubackup kernel: [ 7228.720909] Call Trace:
Jul 2 12:41:32 ubackup kernel: [ 7228.720913] show_stack+0x52/0x58
Jul 2 12:41:32 ubackup kernel: [ 7228.720918] dump_stack+0x70/0x8b
Jul 2 12:41:32 ubackup kernel: [ 7228.720922] spl_dumpstack+0x29/0x2b [spl]
Jul 2 12:41:32 ubackup kernel: [ 7228.720938] vcmn_err.cold+0x60/0x94 [spl]
Jul 2 12:41:32 ubackup kernel: [ 7228.720951] ? ptr_to_id+0xbd/0x270
Jul 2 12:41:32 ubackup kernel: [ 7228.720955] ? pointer+0x19b/0x4d0
Jul 2 12:41:32 ubackup kernel: [ 7228.720958] zfs_panic_recover+0x6d/0x90 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.721155] zfs_blkptr_verify_log+0x94/0x100 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.721297] ? spl_kmem_cache_alloc+0x3b/0x100 [spl]
Jul 2 12:41:32 ubackup kernel: [ 7228.721309] ? _cond_resched+0x1a/0x50
Jul 2 12:41:32 ubackup kernel: [ 7228.721312] ? do_raw_spin_unlock+0x9/0x10 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.721452] ? __raw_spin_unlock+0x9/0x10 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.721592] ? txg_all_lists_empty+0x62/0xb0 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.721731] ? do_softirq_own_stack+0x3d/0x50
Jul 2 12:41:32 ubackup kernel: [ 7228.721735] ? irq_exit_rcu+0x42/0xd0
Jul 2 12:41:32 ubackup kernel: [ 7228.721738] ? common_interrupt+0x88/0x140
Jul 2 12:41:32 ubackup kernel: [ 7228.721740] ? vdev_queue_max_async_writes+0x42/0xe0 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.721881] ? spl_kmem_cache_alloc+0x3b/0x100 [spl]
Jul 2 12:41:32 ubackup kernel: [ 7228.721892] zfs_blkptr_verify+0x359/0x470 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722034] zio_free+0x27/0x100 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722175] dsl_free+0x11/0x20 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722309] dsl_dataset_block_kill+0x45e/0x490 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722439] ? _cond_resched+0x1a/0x50
Jul 2 12:41:32 ubackup kernel: [ 7228.722442] dbuf_write_done+0x19a/0x1c0 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722566] arc_write_done+0x25e/0x420 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722688] ? __raw_spin_unlock+0x9/0x10 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722830] zio_done+0x39d/0xdc0 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722972] zio_execute+0x92/0xe0 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.723113] taskq_thread+0x236/0x420 [spl]
Jul 2 12:41:32 ubackup kernel: [ 7228.723125] ? wake_up_q+0xa0/0xa0
Jul 2 12:41:32 ubackup kernel: [ 7228.723128] ? zio_execute_stack_check.constprop.0+0x10/0x10 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.723271] kthread+0x12f/0x150
Jul 2 12:41:32 ubackup kernel: [ 7228.723274] ? param_set_taskq_kick+0xf0/0xf0 [spl]
Jul 2 12:41:32 ubackup kernel: [ 7228.723285] ? __kthread_bind_mask+0x70/0x70
Jul 2 12:41:32 ubackup kernel: [ 7228.723288] ret_from_fork+0x22/0x30
Jul 8 01:32:19 ubackup kernel: [463029.048674] PANIC: tank: blkptr at 000000008e4479a1 has invalid CHECKSUM 0
Jul 8 01:32:19 ubackup kernel: [463029.048684] Showing stack for process 3084
Jul 8 01:32:19 ubackup kernel: [463029.048686] CPU: 0 PID: 3084 Comm: z_wr_iss Tainted: P O 5.11.0-22-generic #23-Ubuntu
Jul 8 01:32:19 ubackup kernel: [463029.048689] Hardware name: Gigabyte Technology Co., Ltd. GA-990FXA-UD3/GA-990FXA-UD3, BIOS FB 10/13/2011
Jul 8 01:32:19 ubackup kernel: [463029.048691] Call Trace:
Jul 8 01:32:19 ubackup kernel: [463029.048695] show_stack+0x52/0x58
Jul 8 01:32:19 ubackup kernel: [463029.048700] dump_stack+0x70/0x8b
Jul 8 01:32:19 ubackup kernel: [463029.048704] spl_dumpstack+0x29/0x2b [spl]
Jul 8 01:32:19 ubackup kernel: [463029.048720] vcmn_err.cold+0x60/0x94 [spl]
Jul 8 01:32:19 ubackup kernel: [463029.048733] ? ptr_to_id+0xbd/0x270
Jul 8 01:32:19 ubackup kernel: [463029.048737] ? pointer+0x19b/0x4d0
Jul 8 01:32:19 ubackup kernel: [463029.048740] zfs_panic_recover+0x6d/0x90 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.048932] zfs_blkptr_verify_log+0x94/0x100 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049075] ? spl_kmem_cache_alloc+0x3b/0x100 [spl]
Jul 8 01:32:19 ubackup kernel: [463029.049087] ? _cond_resched+0x1a/0x50
Jul 8 01:32:19 ubackup kernel: [463029.049090] ? do_raw_spin_unlock+0x9/0x10 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049229] ? __raw_spin_unlock+0x9/0x10 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049369] ? txg_all_lists_empty+0x62/0xb0 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049509] ? spa_has_pending_synctask+0x46/0x60 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049648] ? vdev_queue_max_async_writes+0x42/0xe0 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049788] ? vdev_queue_class_to_issue+0xf7/0x120 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049928] zfs_blkptr_verify+0x359/0x470 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050070] ? _cond_resched+0x1a/0x50
Jul 8 01:32:19 ubackup kernel: [463029.050072] zio_free+0x27/0x100 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050213] dsl_free+0x11/0x20 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050347] dsl_dataset_block_kill+0x45e/0x490 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050477] ? _cond_resched+0x1a/0x50
Jul 8 01:32:19 ubackup kernel: [463029.050480] dbuf_write_done+0x19a/0x1c0 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050604] arc_write_done+0x25e/0x420 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050726] ? __raw_spin_unlock+0x9/0x10 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050868] zio_done+0x39d/0xdc0 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.051010] zio_execute+0x92/0xe0 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.051151] taskq_thread+0x236/0x420 [spl]
Jul 8 01:32:19 ubackup kernel: [463029.051163] ? wake_up_q+0xa0/0xa0
Jul 8 01:32:19 ubackup kernel: [463029.051167] ? zio_execute_stack_check.constprop.0+0x10/0x10 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.051309] kthread+0x12f/0x150
Jul 8 01:32:19 ubackup kernel: [463029.051312] ? param_set_taskq_kick+0xf0/0xf0 [spl]
Jul 8 01:32:19 ubackup kernel: [463029.051323] ? __kthread_bind_mask+0x70/0x70
Jul 8 01:32:19 ubackup kernel: [463029.051326] ret_from_fork+0x22/0x30
Jul 10 01:56:12 ubackup kernel: [47704.696345] PANIC: tank: blkptr at 00000000fa306804 has invalid CHECKSUM 0
Jul 10 01:56:12 ubackup kernel: [47704.696355] Showing stack for process 3313
Jul 10 01:56:12 ubackup kernel: [47704.696358] CPU: 0 PID: 3313 Comm: z_wr_int Tainted: P OE 5.11.0-22-generic #23-Ubuntu
Jul 10 01:56:12 ubackup kernel: [47704.696361] Hardware name: Gigabyte Technology Co., Ltd. GA-990FXA-UD3/GA-990FXA-UD3, BIOS FB 10/13/2011
Jul 10 01:56:12 ubackup kernel: [47704.696363] Call Trace:
Jul 10 01:56:12 ubackup kernel: [47704.696367] show_stack+0x52/0x58
Jul 10 01:56:12 ubackup kernel: [47704.696372] dump_stack+0x70/0x8b
Jul 10 01:56:12 ubackup kernel: [47704.696376] spl_dumpstack+0x29/0x2b [spl]
Jul 10 01:56:12 ubackup kernel: [47704.696390] vcmn_err.cold+0x60/0x94 [spl]
Jul 10 01:56:12 ubackup kernel: [47704.696400] ? zfs_btree_insert_into_leaf+0x24a/0x2c0 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.696545] ? ptr_to_id+0xbd/0x270
Jul 10 01:56:12 ubackup kernel: [47704.696549] ? pointer+0x19b/0x4d0
Jul 10 01:56:12 ubackup kernel: [47704.696552] zfs_panic_recover+0x6d/0x90 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.696666] zfs_blkptr_verify_log+0x94/0x100 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.696782] ? range_tree_add+0x11/0x20 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.696893] ? metaslab_free_concrete+0x10b/0x260 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697004] ? metaslab_free_impl+0xaf/0xe0 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697114] ? _cond_resched+0x1a/0x50
Jul 10 01:56:12 ubackup kernel: [47704.697119] zfs_blkptr_verify+0x359/0x470 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697234] zio_free+0x27/0x100 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697349] dsl_free+0x11/0x20 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697458] dsl_dataset_block_kill+0x4b6/0x4f0 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697564] dbuf_write_done+0x1ad/0x1f0 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697664] arc_write_done+0x8f/0x420 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697763] zio_done+0x405/0x11b0 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697879] zio_execute+0x8b/0x130 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697994] taskq_thread+0x2b7/0x500 [spl]
Jul 10 01:56:12 ubackup kernel: [47704.698003] ? wake_up_q+0xa0/0xa0
Jul 10 01:56:12 ubackup kernel: [47704.698007] ? zio_gang_tree_free+0x70/0x70 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.698123] kthread+0x12f/0x150
Jul 10 01:56:12 ubackup kernel: [47704.698126] ? taskq_thread_spawn+0x60/0x60 [spl]
Jul 10 01:56:12 ubackup kernel: [47704.698135] ? __kthread_bind_mask+0x70/0x70
Jul 10 01:56:12 ubackup kernel: [47704.698138] ret_from_fork+0x22/0x30
```
|
1.0
|
PANIC: blkptr has invalid CHECKSUM 0 - I have seen a few stale issues with similar title, however my stack traces look somewhat different, so I'm opening a new issue. I'm trying to figure out if I have a hardware problem or if my pool is actually borked, possibly caused by a bug.
### System information
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | 20.04, 21.04
Linux Kernel | 5.8.0-55-generic, 5.11.0-22-generic
Architecture | AMD64
ZFS Version | tracking master, 2.0.2-1ubuntu5, 2.0.5
SPL Version | tracking master, 2.0.2-1ubuntu5, 2.0.5
### Describe the problem you're observing
The system functions purely as a backup receiving sends and deleting old snapshots. After 1-10 days of running, the system panics while receiving with strack trace variations of "blkptr has invalid CHECKSUM 0". Subsequent scrubs comes back clean.
### Describe how to reproduce the problem
I can reproduce by leaving the system receiving backups for a few days. I don't know how to reproduce on another system.
### Include any warning/errors/backtraces from the system logs
- The first two traces are from Ubuntu 20.04 with zfs compiled from master at the time
- The next three are from Ubuntu 21.04 with the provided packages for zfs 2.0.2
- The last one is from Ubuntu 21.04 with packages compiled from the 2.0.5 to test if the changes from #12054 have any effect.
```
Jun 20 01:52:37 ubackup kernel: [482903.384554] PANIC: tank: blkptr at 00000000dbbffcf1 has invalid CHECKSUM 0
Jun 20 01:52:37 ubackup kernel: [482903.384572] Showing stack for process 3425
Jun 20 01:52:37 ubackup kernel: [482903.384580] CPU: 1 PID: 3425 Comm: z_wr_iss Tainted: P OE 5.8.0-55-generic #62~20.04.1-Ubuntu
Jun 20 01:52:37 ubackup kernel: [482903.384582] Hardware name: Gigabyte Technology Co., Ltd. GA-990FXA-UD3/GA-990FXA-UD3, BIOS FB 10/13/2011
Jun 20 01:52:37 ubackup kernel: [482903.384584] Call Trace:
Jun 20 01:52:37 ubackup kernel: [482903.384598] dump_stack+0x74/0x92
Jun 20 01:52:37 ubackup kernel: [482903.384625] spl_dumpstack+0x29/0x2b [spl]
Jun 20 01:52:37 ubackup kernel: [482903.384645] vcmn_err.cold+0x60/0x94 [spl]
Jun 20 01:52:37 ubackup kernel: [482903.384654] ? __blk_mq_sched_dispatch_requests+0x10e/0x170
Jun 20 01:52:37 ubackup kernel: [482903.384662] ? __blk_mq_run_hw_queue+0x5a/0x110
Jun 20 01:52:37 ubackup kernel: [482903.384669] ? ptr_to_id+0xbe/0x220
Jun 20 01:52:37 ubackup kernel: [482903.384916] zfs_panic_recover+0x6f/0x90 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.385130] zfs_blkptr_verify_log+0x94/0x100 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.385342] ? vdev_disk_io_start+0x49f/0x8e0 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.385351] ? _cond_resched+0x19/0x30
Jun 20 01:52:37 ubackup kernel: [482903.385356] ? mutex_lock+0x13/0x40
Jun 20 01:52:37 ubackup kernel: [482903.385568] ? zio_wait_for_children+0x8e/0xd0 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.385780] zfs_blkptr_verify+0x3c9/0x480 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.385993] zio_free+0x27/0x100 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.386191] dsl_free+0x11/0x20 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.386380] dsl_dataset_block_kill+0x4bf/0x4f0 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.386386] ? down_write+0x13/0x50
Jun 20 01:52:37 ubackup kernel: [482903.386565] dbuf_write_done+0x1b3/0x200 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.386740] arc_write_done+0x8f/0x410 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.386953] zio_done+0x407/0x1050 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.387166] zio_execute+0x93/0xf0 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.387185] taskq_thread+0x2fb/0x510 [spl]
Jun 20 01:52:37 ubackup kernel: [482903.387193] ? wake_up_q+0xa0/0xa0
Jun 20 01:52:37 ubackup kernel: [482903.387406] ? zio_taskq_member.isra.0.constprop.0+0x60/0x60 [zfs]
Jun 20 01:52:37 ubackup kernel: [482903.387413] kthread+0x114/0x150
Jun 20 01:52:37 ubackup kernel: [482903.387431] ? task_done+0xb0/0xb0 [spl]
Jun 20 01:52:37 ubackup kernel: [482903.387434] ? kthread_park+0x90/0x90
Jun 20 01:52:37 ubackup kernel: [482903.387440] ret_from_fork+0x22/0x30
Jun 25 01:43:58 ubackup kernel: [367708.395650] PANIC: tank: blkptr at 000000004322d8f5 has invalid CHECKSUM 0
Jun 25 01:43:58 ubackup kernel: [367708.395659] Showing stack for process 3381
Jun 25 01:43:58 ubackup kernel: [367708.395663] CPU: 1 PID: 3381 Comm: dp_sync_taskq Tainted: P OE 5.8.0-55-generic #62~20.04.1-Ubuntu
Jun 25 01:43:58 ubackup kernel: [367708.395664] Hardware name: Gigabyte Technology Co., Ltd. GA-990FXA-UD3/GA-990FXA-UD3, BIOS FB 10/13/2011
Jun 25 01:43:58 ubackup kernel: [367708.395664] Call Trace:
Jun 25 01:43:58 ubackup kernel: [367708.395672] dump_stack+0x74/0x92
Jun 25 01:43:58 ubackup kernel: [367708.395684] spl_dumpstack+0x29/0x2b [spl]
Jun 25 01:43:58 ubackup kernel: [367708.395691] vcmn_err.cold+0x60/0x94 [spl]
Jun 25 01:43:58 ubackup kernel: [367708.395698] ? spl_kmem_cache_alloc+0xa9/0x7d0 [spl]
Jun 25 01:43:58 ubackup kernel: [367708.395701] ? ptr_to_id+0xbe/0x220
Jun 25 01:43:58 ubackup kernel: [367708.395826] zfs_panic_recover+0x6f/0x90 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.395902] zfs_blkptr_verify_log+0x94/0x100 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.395963] ? arc_hdr_destroy+0x200/0x200 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.395967] ? try_to_wake_up+0x66/0x540
Jun 25 01:43:58 ubackup kernel: [367708.396028] ? dbuf_issue_final_prefetch+0xd0/0xd0 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396031] ? default_wake_function+0x1a/0x30
Jun 25 01:43:58 ubackup kernel: [367708.396032] ? __wake_up_common+0x7e/0x140
Jun 25 01:43:58 ubackup kernel: [367708.396034] ? __wake_up_common_lock+0x8a/0xc0
Jun 25 01:43:58 ubackup kernel: [367708.396108] zfs_blkptr_verify+0x3c9/0x480 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396182] zio_free+0x27/0x100 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396251] dsl_free+0x11/0x20 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396317] dsl_dataset_block_kill+0x4bf/0x4f0 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396382] free_blocks+0xea/0x1d0 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396456] ? zio_nowait+0xc1/0x150 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396521] dnode_sync_free_range+0x23c/0x270 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396583] ? dbuf_sync_leaf+0x23b/0x500 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396649] ? free_children+0x3d0/0x3d0 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396718] range_tree_walk+0x118/0x1f0 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396784] dnode_sync+0x2e8/0xa90 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396787] ? __switch_to+0x157/0x450
Jun 25 01:43:58 ubackup kernel: [367708.396789] ? _cond_resched+0x19/0x30
Jun 25 01:43:58 ubackup kernel: [367708.396791] ? mutex_lock+0x13/0x40
Jun 25 01:43:58 ubackup kernel: [367708.396855] sync_dnodes_task+0x79/0xb0 [zfs]
Jun 25 01:43:58 ubackup kernel: [367708.396861] taskq_thread+0x2fb/0x510 [spl]
Jun 25 01:43:58 ubackup kernel: [367708.396864] ? wake_up_q+0xa0/0xa0
Jun 25 01:43:58 ubackup kernel: [367708.396866] kthread+0x114/0x150
Jun 25 01:43:58 ubackup kernel: [367708.396873] ? task_done+0xb0/0xb0 [spl]
Jun 25 01:43:58 ubackup kernel: [367708.396874] ? kthread_park+0x90/0x90
Jun 25 01:43:58 ubackup kernel: [367708.396876] ret_from_fork+0x22/0x30
Jul 1 02:01:26 ubackup kernel: [443020.371564] PANIC: tank: blkptr at 000000002c351e10 has invalid CHECKSUM 0
Jul 1 02:01:26 ubackup kernel: [443020.371590] Showing stack for process 3346
Jul 1 02:01:26 ubackup kernel: [443020.371597] CPU: 1 PID: 3346 Comm: z_wr_iss Tainted: P OE 5.11.0-22-generic #23~20.04.1-Ubuntu
Jul 1 02:01:26 ubackup kernel: [443020.371608] Hardware name: Gigabyte Technology Co., Ltd. GA-990FXA-UD3/GA-990FXA-UD3, BIOS FB 10/13/2011
Jul 1 02:01:26 ubackup kernel: [443020.371613] Call Trace:
Jul 1 02:01:26 ubackup kernel: [443020.371622] dump_stack+0x74/0x92
Jul 1 02:01:26 ubackup kernel: [443020.371640] spl_dumpstack+0x29/0x2b [spl]
Jul 1 02:01:26 ubackup kernel: [443020.371683] vcmn_err.cold+0x60/0x94 [spl]
Jul 1 02:01:26 ubackup kernel: [443020.371722] ? ptr_to_id+0xbe/0x220
Jul 1 02:01:26 ubackup kernel: [443020.371739] zfs_panic_recover+0x6f/0x90 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.372186] zfs_blkptr_verify_log+0x94/0x100 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.372613] ? vdev_disk_io_start+0x4a1/0x8e0 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.373033] ? _cond_resched+0x19/0x30
Jul 1 02:01:26 ubackup kernel: [443020.373045] ? mutex_lock+0x13/0x40
Jul 1 02:01:26 ubackup kernel: [443020.373052] ? zio_wait_for_children+0x8e/0xd0 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.373476] zfs_blkptr_verify+0x3c9/0x480 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.373902] zio_free+0x27/0x100 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.374326] dsl_free+0x11/0x20 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.374719] dsl_dataset_block_kill+0x4bf/0x4f0 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.375097] ? down_write+0x13/0x50
Jul 1 02:01:26 ubackup kernel: [443020.375106] dbuf_write_done+0x1b3/0x200 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.375463] arc_write_done+0x8f/0x410 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.375812] zio_done+0x407/0x1050 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.376237] zio_execute+0x93/0xf0 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.376661] taskq_thread+0x2fb/0x510 [spl]
Jul 1 02:01:26 ubackup kernel: [443020.376698] ? wake_up_q+0xa0/0xa0
Jul 1 02:01:26 ubackup kernel: [443020.376710] ? zio_taskq_member.isra.0.constprop.0+0x60/0x60 [zfs]
Jul 1 02:01:26 ubackup kernel: [443020.377136] kthread+0x114/0x150
Jul 1 02:01:26 ubackup kernel: [443020.377146] ? task_done+0xb0/0xb0 [spl]
Jul 1 02:01:26 ubackup kernel: [443020.377181] ? kthread_park+0x90/0x90
Jul 1 02:01:26 ubackup kernel: [443020.377190] ret_from_fork+0x22/0x30
Jul 2 12:41:32 ubackup kernel: [ 7228.720883] PANIC: tank: blkptr at 0000000084e788bd has invalid CHECKSUM 0
Jul 2 12:41:32 ubackup kernel: [ 7228.720901] Showing stack for process 4218
Jul 2 12:41:32 ubackup kernel: [ 7228.720904] CPU: 1 PID: 4218 Comm: z_wr_iss Tainted: P O 5.11.0-22-generic #23-Ubuntu
Jul 2 12:41:32 ubackup kernel: [ 7228.720907] Hardware name: Gigabyte Technology Co., Ltd. GA-990FXA-UD3/GA-990FXA-UD3, BIOS FB 10/13/2011
Jul 2 12:41:32 ubackup kernel: [ 7228.720909] Call Trace:
Jul 2 12:41:32 ubackup kernel: [ 7228.720913] show_stack+0x52/0x58
Jul 2 12:41:32 ubackup kernel: [ 7228.720918] dump_stack+0x70/0x8b
Jul 2 12:41:32 ubackup kernel: [ 7228.720922] spl_dumpstack+0x29/0x2b [spl]
Jul 2 12:41:32 ubackup kernel: [ 7228.720938] vcmn_err.cold+0x60/0x94 [spl]
Jul 2 12:41:32 ubackup kernel: [ 7228.720951] ? ptr_to_id+0xbd/0x270
Jul 2 12:41:32 ubackup kernel: [ 7228.720955] ? pointer+0x19b/0x4d0
Jul 2 12:41:32 ubackup kernel: [ 7228.720958] zfs_panic_recover+0x6d/0x90 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.721155] zfs_blkptr_verify_log+0x94/0x100 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.721297] ? spl_kmem_cache_alloc+0x3b/0x100 [spl]
Jul 2 12:41:32 ubackup kernel: [ 7228.721309] ? _cond_resched+0x1a/0x50
Jul 2 12:41:32 ubackup kernel: [ 7228.721312] ? do_raw_spin_unlock+0x9/0x10 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.721452] ? __raw_spin_unlock+0x9/0x10 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.721592] ? txg_all_lists_empty+0x62/0xb0 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.721731] ? do_softirq_own_stack+0x3d/0x50
Jul 2 12:41:32 ubackup kernel: [ 7228.721735] ? irq_exit_rcu+0x42/0xd0
Jul 2 12:41:32 ubackup kernel: [ 7228.721738] ? common_interrupt+0x88/0x140
Jul 2 12:41:32 ubackup kernel: [ 7228.721740] ? vdev_queue_max_async_writes+0x42/0xe0 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.721881] ? spl_kmem_cache_alloc+0x3b/0x100 [spl]
Jul 2 12:41:32 ubackup kernel: [ 7228.721892] zfs_blkptr_verify+0x359/0x470 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722034] zio_free+0x27/0x100 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722175] dsl_free+0x11/0x20 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722309] dsl_dataset_block_kill+0x45e/0x490 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722439] ? _cond_resched+0x1a/0x50
Jul 2 12:41:32 ubackup kernel: [ 7228.722442] dbuf_write_done+0x19a/0x1c0 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722566] arc_write_done+0x25e/0x420 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722688] ? __raw_spin_unlock+0x9/0x10 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722830] zio_done+0x39d/0xdc0 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.722972] zio_execute+0x92/0xe0 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.723113] taskq_thread+0x236/0x420 [spl]
Jul 2 12:41:32 ubackup kernel: [ 7228.723125] ? wake_up_q+0xa0/0xa0
Jul 2 12:41:32 ubackup kernel: [ 7228.723128] ? zio_execute_stack_check.constprop.0+0x10/0x10 [zfs]
Jul 2 12:41:32 ubackup kernel: [ 7228.723271] kthread+0x12f/0x150
Jul 2 12:41:32 ubackup kernel: [ 7228.723274] ? param_set_taskq_kick+0xf0/0xf0 [spl]
Jul 2 12:41:32 ubackup kernel: [ 7228.723285] ? __kthread_bind_mask+0x70/0x70
Jul 2 12:41:32 ubackup kernel: [ 7228.723288] ret_from_fork+0x22/0x30
Jul 8 01:32:19 ubackup kernel: [463029.048674] PANIC: tank: blkptr at 000000008e4479a1 has invalid CHECKSUM 0
Jul 8 01:32:19 ubackup kernel: [463029.048684] Showing stack for process 3084
Jul 8 01:32:19 ubackup kernel: [463029.048686] CPU: 0 PID: 3084 Comm: z_wr_iss Tainted: P O 5.11.0-22-generic #23-Ubuntu
Jul 8 01:32:19 ubackup kernel: [463029.048689] Hardware name: Gigabyte Technology Co., Ltd. GA-990FXA-UD3/GA-990FXA-UD3, BIOS FB 10/13/2011
Jul 8 01:32:19 ubackup kernel: [463029.048691] Call Trace:
Jul 8 01:32:19 ubackup kernel: [463029.048695] show_stack+0x52/0x58
Jul 8 01:32:19 ubackup kernel: [463029.048700] dump_stack+0x70/0x8b
Jul 8 01:32:19 ubackup kernel: [463029.048704] spl_dumpstack+0x29/0x2b [spl]
Jul 8 01:32:19 ubackup kernel: [463029.048720] vcmn_err.cold+0x60/0x94 [spl]
Jul 8 01:32:19 ubackup kernel: [463029.048733] ? ptr_to_id+0xbd/0x270
Jul 8 01:32:19 ubackup kernel: [463029.048737] ? pointer+0x19b/0x4d0
Jul 8 01:32:19 ubackup kernel: [463029.048740] zfs_panic_recover+0x6d/0x90 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.048932] zfs_blkptr_verify_log+0x94/0x100 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049075] ? spl_kmem_cache_alloc+0x3b/0x100 [spl]
Jul 8 01:32:19 ubackup kernel: [463029.049087] ? _cond_resched+0x1a/0x50
Jul 8 01:32:19 ubackup kernel: [463029.049090] ? do_raw_spin_unlock+0x9/0x10 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049229] ? __raw_spin_unlock+0x9/0x10 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049369] ? txg_all_lists_empty+0x62/0xb0 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049509] ? spa_has_pending_synctask+0x46/0x60 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049648] ? vdev_queue_max_async_writes+0x42/0xe0 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049788] ? vdev_queue_class_to_issue+0xf7/0x120 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.049928] zfs_blkptr_verify+0x359/0x470 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050070] ? _cond_resched+0x1a/0x50
Jul 8 01:32:19 ubackup kernel: [463029.050072] zio_free+0x27/0x100 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050213] dsl_free+0x11/0x20 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050347] dsl_dataset_block_kill+0x45e/0x490 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050477] ? _cond_resched+0x1a/0x50
Jul 8 01:32:19 ubackup kernel: [463029.050480] dbuf_write_done+0x19a/0x1c0 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050604] arc_write_done+0x25e/0x420 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050726] ? __raw_spin_unlock+0x9/0x10 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.050868] zio_done+0x39d/0xdc0 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.051010] zio_execute+0x92/0xe0 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.051151] taskq_thread+0x236/0x420 [spl]
Jul 8 01:32:19 ubackup kernel: [463029.051163] ? wake_up_q+0xa0/0xa0
Jul 8 01:32:19 ubackup kernel: [463029.051167] ? zio_execute_stack_check.constprop.0+0x10/0x10 [zfs]
Jul 8 01:32:19 ubackup kernel: [463029.051309] kthread+0x12f/0x150
Jul 8 01:32:19 ubackup kernel: [463029.051312] ? param_set_taskq_kick+0xf0/0xf0 [spl]
Jul 8 01:32:19 ubackup kernel: [463029.051323] ? __kthread_bind_mask+0x70/0x70
Jul 8 01:32:19 ubackup kernel: [463029.051326] ret_from_fork+0x22/0x30
Jul 10 01:56:12 ubackup kernel: [47704.696345] PANIC: tank: blkptr at 00000000fa306804 has invalid CHECKSUM 0
Jul 10 01:56:12 ubackup kernel: [47704.696355] Showing stack for process 3313
Jul 10 01:56:12 ubackup kernel: [47704.696358] CPU: 0 PID: 3313 Comm: z_wr_int Tainted: P OE 5.11.0-22-generic #23-Ubuntu
Jul 10 01:56:12 ubackup kernel: [47704.696361] Hardware name: Gigabyte Technology Co., Ltd. GA-990FXA-UD3/GA-990FXA-UD3, BIOS FB 10/13/2011
Jul 10 01:56:12 ubackup kernel: [47704.696363] Call Trace:
Jul 10 01:56:12 ubackup kernel: [47704.696367] show_stack+0x52/0x58
Jul 10 01:56:12 ubackup kernel: [47704.696372] dump_stack+0x70/0x8b
Jul 10 01:56:12 ubackup kernel: [47704.696376] spl_dumpstack+0x29/0x2b [spl]
Jul 10 01:56:12 ubackup kernel: [47704.696390] vcmn_err.cold+0x60/0x94 [spl]
Jul 10 01:56:12 ubackup kernel: [47704.696400] ? zfs_btree_insert_into_leaf+0x24a/0x2c0 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.696545] ? ptr_to_id+0xbd/0x270
Jul 10 01:56:12 ubackup kernel: [47704.696549] ? pointer+0x19b/0x4d0
Jul 10 01:56:12 ubackup kernel: [47704.696552] zfs_panic_recover+0x6d/0x90 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.696666] zfs_blkptr_verify_log+0x94/0x100 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.696782] ? range_tree_add+0x11/0x20 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.696893] ? metaslab_free_concrete+0x10b/0x260 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697004] ? metaslab_free_impl+0xaf/0xe0 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697114] ? _cond_resched+0x1a/0x50
Jul 10 01:56:12 ubackup kernel: [47704.697119] zfs_blkptr_verify+0x359/0x470 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697234] zio_free+0x27/0x100 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697349] dsl_free+0x11/0x20 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697458] dsl_dataset_block_kill+0x4b6/0x4f0 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697564] dbuf_write_done+0x1ad/0x1f0 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697664] arc_write_done+0x8f/0x420 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697763] zio_done+0x405/0x11b0 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697879] zio_execute+0x8b/0x130 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.697994] taskq_thread+0x2b7/0x500 [spl]
Jul 10 01:56:12 ubackup kernel: [47704.698003] ? wake_up_q+0xa0/0xa0
Jul 10 01:56:12 ubackup kernel: [47704.698007] ? zio_gang_tree_free+0x70/0x70 [zfs]
Jul 10 01:56:12 ubackup kernel: [47704.698123] kthread+0x12f/0x150
Jul 10 01:56:12 ubackup kernel: [47704.698126] ? taskq_thread_spawn+0x60/0x60 [spl]
Jul 10 01:56:12 ubackup kernel: [47704.698135] ? __kthread_bind_mask+0x70/0x70
Jul 10 01:56:12 ubackup kernel: [47704.698138] ret_from_fork+0x22/0x30
```
|
defect
|
panic blkptr has invalid checksum i have seen a few stale issues with similar title however my stack traces look somewhat different so i m opening a new issue i m trying to figure out if i have a hardware problem or if my pool is actually borked possibly caused by a bug system information type version name distribution name ubuntu distribution version linux kernel generic generic architecture zfs version tracking master spl version tracking master describe the problem you re observing the system functions purely as a backup receiving sends and deleting old snapshots after days of running the system panics while receiving with strack trace variations of blkptr has invalid checksum subsequent scrubs comes back clean describe how to reproduce the problem i can reproduce by leaving the system receiving backups for a few days i don t know how to reproduce on another system include any warning errors backtraces from the system logs the first two traces are from ubuntu with zfs compiled from master at the time the next three are from ubuntu with the provided packages for zfs the last one is from ubuntu with packages compiled from the to test if the changes from have any effect jun ubackup kernel panic tank blkptr at has invalid checksum jun ubackup kernel showing stack for process jun ubackup kernel cpu pid comm z wr iss tainted p oe generic ubuntu jun ubackup kernel hardware name gigabyte technology co ltd ga ga bios fb jun ubackup kernel call trace jun ubackup kernel dump stack jun ubackup kernel spl dumpstack jun ubackup kernel vcmn err cold jun ubackup kernel blk mq sched dispatch requests jun ubackup kernel blk mq run hw queue jun ubackup kernel ptr to id jun ubackup kernel zfs panic recover jun ubackup kernel zfs blkptr verify log jun ubackup kernel vdev disk io start jun ubackup kernel cond resched jun ubackup kernel mutex lock jun ubackup kernel zio wait for children jun ubackup kernel zfs blkptr verify jun ubackup kernel zio free jun ubackup kernel dsl free jun ubackup kernel dsl dataset block kill jun ubackup kernel down write jun ubackup kernel dbuf write done jun ubackup kernel arc write done jun ubackup kernel zio done jun ubackup kernel zio execute jun ubackup kernel taskq thread jun ubackup kernel wake up q jun ubackup kernel zio taskq member isra constprop jun ubackup kernel kthread jun ubackup kernel task done jun ubackup kernel kthread park jun ubackup kernel ret from fork jun ubackup kernel panic tank blkptr at has invalid checksum jun ubackup kernel showing stack for process jun ubackup kernel cpu pid comm dp sync taskq tainted p oe generic ubuntu jun ubackup kernel hardware name gigabyte technology co ltd ga ga bios fb jun ubackup kernel call trace jun ubackup kernel dump stack jun ubackup kernel spl dumpstack jun ubackup kernel vcmn err cold jun ubackup kernel spl kmem cache alloc jun ubackup kernel ptr to id jun ubackup kernel zfs panic recover jun ubackup kernel zfs blkptr verify log jun ubackup kernel arc hdr destroy jun ubackup kernel try to wake up jun ubackup kernel dbuf issue final prefetch jun ubackup kernel default wake function jun ubackup kernel wake up common jun ubackup kernel wake up common lock jun ubackup kernel zfs blkptr verify jun ubackup kernel zio free jun ubackup kernel dsl free jun ubackup kernel dsl dataset block kill jun ubackup kernel free blocks jun ubackup kernel zio nowait jun ubackup kernel dnode sync free range jun ubackup kernel dbuf sync leaf jun ubackup kernel free children jun ubackup kernel range tree walk jun ubackup kernel dnode sync jun ubackup kernel switch to jun ubackup kernel cond resched jun ubackup kernel mutex lock jun ubackup kernel sync dnodes task jun ubackup kernel taskq thread jun ubackup kernel wake up q jun ubackup kernel kthread jun ubackup kernel task done jun ubackup kernel kthread park jun ubackup kernel ret from fork jul ubackup kernel panic tank blkptr at has invalid checksum jul ubackup kernel showing stack for process jul ubackup kernel cpu pid comm z wr iss tainted p oe generic ubuntu jul ubackup kernel hardware name gigabyte technology co ltd ga ga bios fb jul ubackup kernel call trace jul ubackup kernel dump stack jul ubackup kernel spl dumpstack jul ubackup kernel vcmn err cold jul ubackup kernel ptr to id jul ubackup kernel zfs panic recover jul ubackup kernel zfs blkptr verify log jul ubackup kernel vdev disk io start jul ubackup kernel cond resched jul ubackup kernel mutex lock jul ubackup kernel zio wait for children jul ubackup kernel zfs blkptr verify jul ubackup kernel zio free jul ubackup kernel dsl free jul ubackup kernel dsl dataset block kill jul ubackup kernel down write jul ubackup kernel dbuf write done jul ubackup kernel arc write done jul ubackup kernel zio done jul ubackup kernel zio execute jul ubackup kernel taskq thread jul ubackup kernel wake up q jul ubackup kernel zio taskq member isra constprop jul ubackup kernel kthread jul ubackup kernel task done jul ubackup kernel kthread park jul ubackup kernel ret from fork jul ubackup kernel panic tank blkptr at has invalid checksum jul ubackup kernel showing stack for process jul ubackup kernel cpu pid comm z wr iss tainted p o generic ubuntu jul ubackup kernel hardware name gigabyte technology co ltd ga ga bios fb jul ubackup kernel call trace jul ubackup kernel show stack jul ubackup kernel dump stack jul ubackup kernel spl dumpstack jul ubackup kernel vcmn err cold jul ubackup kernel ptr to id jul ubackup kernel pointer jul ubackup kernel zfs panic recover jul ubackup kernel zfs blkptr verify log jul ubackup kernel spl kmem cache alloc jul ubackup kernel cond resched jul ubackup kernel do raw spin unlock jul ubackup kernel raw spin unlock jul ubackup kernel txg all lists empty jul ubackup kernel do softirq own stack jul ubackup kernel irq exit rcu jul ubackup kernel common interrupt jul ubackup kernel vdev queue max async writes jul ubackup kernel spl kmem cache alloc jul ubackup kernel zfs blkptr verify jul ubackup kernel zio free jul ubackup kernel dsl free jul ubackup kernel dsl dataset block kill jul ubackup kernel cond resched jul ubackup kernel dbuf write done jul ubackup kernel arc write done jul ubackup kernel raw spin unlock jul ubackup kernel zio done jul ubackup kernel zio execute jul ubackup kernel taskq thread jul ubackup kernel wake up q jul ubackup kernel zio execute stack check constprop jul ubackup kernel kthread jul ubackup kernel param set taskq kick jul ubackup kernel kthread bind mask jul ubackup kernel ret from fork jul ubackup kernel panic tank blkptr at has invalid checksum jul ubackup kernel showing stack for process jul ubackup kernel cpu pid comm z wr iss tainted p o generic ubuntu jul ubackup kernel hardware name gigabyte technology co ltd ga ga bios fb jul ubackup kernel call trace jul ubackup kernel show stack jul ubackup kernel dump stack jul ubackup kernel spl dumpstack jul ubackup kernel vcmn err cold jul ubackup kernel ptr to id jul ubackup kernel pointer jul ubackup kernel zfs panic recover jul ubackup kernel zfs blkptr verify log jul ubackup kernel spl kmem cache alloc jul ubackup kernel cond resched jul ubackup kernel do raw spin unlock jul ubackup kernel raw spin unlock jul ubackup kernel txg all lists empty jul ubackup kernel spa has pending synctask jul ubackup kernel vdev queue max async writes jul ubackup kernel vdev queue class to issue jul ubackup kernel zfs blkptr verify jul ubackup kernel cond resched jul ubackup kernel zio free jul ubackup kernel dsl free jul ubackup kernel dsl dataset block kill jul ubackup kernel cond resched jul ubackup kernel dbuf write done jul ubackup kernel arc write done jul ubackup kernel raw spin unlock jul ubackup kernel zio done jul ubackup kernel zio execute jul ubackup kernel taskq thread jul ubackup kernel wake up q jul ubackup kernel zio execute stack check constprop jul ubackup kernel kthread jul ubackup kernel param set taskq kick jul ubackup kernel kthread bind mask jul ubackup kernel ret from fork jul ubackup kernel panic tank blkptr at has invalid checksum jul ubackup kernel showing stack for process jul ubackup kernel cpu pid comm z wr int tainted p oe generic ubuntu jul ubackup kernel hardware name gigabyte technology co ltd ga ga bios fb jul ubackup kernel call trace jul ubackup kernel show stack jul ubackup kernel dump stack jul ubackup kernel spl dumpstack jul ubackup kernel vcmn err cold jul ubackup kernel zfs btree insert into leaf jul ubackup kernel ptr to id jul ubackup kernel pointer jul ubackup kernel zfs panic recover jul ubackup kernel zfs blkptr verify log jul ubackup kernel range tree add jul ubackup kernel metaslab free concrete jul ubackup kernel metaslab free impl jul ubackup kernel cond resched jul ubackup kernel zfs blkptr verify jul ubackup kernel zio free jul ubackup kernel dsl free jul ubackup kernel dsl dataset block kill jul ubackup kernel dbuf write done jul ubackup kernel arc write done jul ubackup kernel zio done jul ubackup kernel zio execute jul ubackup kernel taskq thread jul ubackup kernel wake up q jul ubackup kernel zio gang tree free jul ubackup kernel kthread jul ubackup kernel taskq thread spawn jul ubackup kernel kthread bind mask jul ubackup kernel ret from fork
| 1
|
6,347
| 2,610,241,099
|
IssuesEvent
|
2015-02-26 19:16:49
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
台州割包茎手术去哪家医院好
|
auto-migrated Priority-Medium Type-Defect
|
```
台州割包茎手术去哪家医院好【台州五洲生殖医院】24小时健
康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:
台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104�
��108、118、198及椒江一金清公交车直达枫南小区,乘坐107、105
、109、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 31 May 2014 at 12:06
|
1.0
|
台州割包茎手术去哪家医院好 - ```
台州割包茎手术去哪家医院好【台州五洲生殖医院】24小时健
康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:
台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104�
��108、118、198及椒江一金清公交车直达枫南小区,乘坐107、105
、109、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 31 May 2014 at 12:06
|
defect
|
台州割包茎手术去哪家医院好 台州割包茎手术去哪家医院好【台州五洲生殖医院】 康咨询热线 微信号tzwzszyy 医院地址 (枫南大转盘旁)乘车线路 � �� 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 1
|
49,862
| 13,187,282,477
|
IssuesEvent
|
2020-08-13 02:55:18
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
[PROPOSAL] Trunk doesn't render in browser - ReadmeRenderer bug (Trac #2193)
|
Incomplete Migration Migrated from Trac defect infrastructure
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2193">https://code.icecube.wisc.edu/ticket/2193</a>, reported by olivas and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-12T20:17:51",
"description": "\nhttps://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/PROPOSAL/trunk\n\nThat produces an error: \n\nGenshi UnicodeDecodeError error while rendering template 'None', line -1, char -1",
"reporter": "olivas",
"cc": "nega",
"resolution": "fixed",
"_ts": "1550002671314091",
"component": "infrastructure",
"summary": "[PROPOSAL] Trunk doesn't render in browser - ReadmeRenderer bug",
"priority": "normal",
"keywords": "trac plugin python",
"time": "2018-10-01T21:13:31",
"milestone": "Vernal Equinox 2019",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[PROPOSAL] Trunk doesn't render in browser - ReadmeRenderer bug (Trac #2193) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2193">https://code.icecube.wisc.edu/ticket/2193</a>, reported by olivas and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-12T20:17:51",
"description": "\nhttps://code.icecube.wisc.edu/projects/icecube/browser/IceCube/projects/PROPOSAL/trunk\n\nThat produces an error: \n\nGenshi UnicodeDecodeError error while rendering template 'None', line -1, char -1",
"reporter": "olivas",
"cc": "nega",
"resolution": "fixed",
"_ts": "1550002671314091",
"component": "infrastructure",
"summary": "[PROPOSAL] Trunk doesn't render in browser - ReadmeRenderer bug",
"priority": "normal",
"keywords": "trac plugin python",
"time": "2018-10-01T21:13:31",
"milestone": "Vernal Equinox 2019",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
defect
|
trunk doesn t render in browser readmerenderer bug trac migrated from json status closed changetime description n produces an error n ngenshi unicodedecodeerror error while rendering template none line char reporter olivas cc nega resolution fixed ts component infrastructure summary trunk doesn t render in browser readmerenderer bug priority normal keywords trac plugin python time milestone vernal equinox owner nega type defect
| 1
|
32,128
| 6,718,211,352
|
IssuesEvent
|
2017-10-15 09:45:35
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
p:panelGrid, header lacks padding when columns > 0
|
defect
|
## 1) Environment
- PrimeFaces version: 6.1
- Does it work on the newest released PrimeFaces version? Version? NO
- Does it work on the newest sources in GitHub? (Build by source -> https://github.com/primefaces/primefaces/wiki/Building-From-Source) NO
- Application server + version: Tomcat 7
- Affected browsers: All
## 2) Expected behavior
The header of the `p:panelGrid` should have proper padding, even if the `columns` attribute is provided.
## 3) Actual behavior
The header doesn't have the proper padding, becuase a css class is missing. Rendered output:

## 4) Steps to reproduce
Just open the provided xhtml.
## 5) Sample XHTML
```xml
<!DOCTYPE html>
<html xmlns="http://www.w3.org/1999/xhtml"
xmlns:h="http://xmlns.jcp.org/jsf/html"
xmlns:f="http://xmlns.jcp.org/jsf/core"
xmlns:p="http://primefaces.org/ui">
<h:head>
<title>Facelet Title</title>
</h:head>
<h:body>
<p:panelGrid columns="1">
<f:facet name="header">
<h:outputText value="With columns"/>
</f:facet>
<h:outputText value="Content"/>
</p:panelGrid>
<p:spacer height="10"/>
<p:panelGrid>
<f:facet name="header">
<p:row>
<p:column>
<h:outputText value="Without columns"/>
</p:column>
</p:row>
</f:facet>
<p:row>
<p:column>
<h:outputText value="Content"/>
</p:column>
</p:row>
</p:panelGrid>
</h:body>
</html>
```
## 6) Sample bean
Not needed
|
1.0
|
p:panelGrid, header lacks padding when columns > 0 - ## 1) Environment
- PrimeFaces version: 6.1
- Does it work on the newest released PrimeFaces version? Version? NO
- Does it work on the newest sources in GitHub? (Build by source -> https://github.com/primefaces/primefaces/wiki/Building-From-Source) NO
- Application server + version: Tomcat 7
- Affected browsers: All
## 2) Expected behavior
The header of the `p:panelGrid` should have proper padding, even if the `columns` attribute is provided.
## 3) Actual behavior
The header doesn't have the proper padding, becuase a css class is missing. Rendered output:

## 4) Steps to reproduce
Just open the provided xhtml.
## 5) Sample XHTML
```xml
<!DOCTYPE html>
<html xmlns="http://www.w3.org/1999/xhtml"
xmlns:h="http://xmlns.jcp.org/jsf/html"
xmlns:f="http://xmlns.jcp.org/jsf/core"
xmlns:p="http://primefaces.org/ui">
<h:head>
<title>Facelet Title</title>
</h:head>
<h:body>
<p:panelGrid columns="1">
<f:facet name="header">
<h:outputText value="With columns"/>
</f:facet>
<h:outputText value="Content"/>
</p:panelGrid>
<p:spacer height="10"/>
<p:panelGrid>
<f:facet name="header">
<p:row>
<p:column>
<h:outputText value="Without columns"/>
</p:column>
</p:row>
</f:facet>
<p:row>
<p:column>
<h:outputText value="Content"/>
</p:column>
</p:row>
</p:panelGrid>
</h:body>
</html>
```
## 6) Sample bean
Not needed
|
defect
|
p panelgrid header lacks padding when columns environment primefaces version does it work on the newest released primefaces version version no does it work on the newest sources in github build by source no application server version tomcat affected browsers all expected behavior the header of the p panelgrid should have proper padding even if the columns attribute is provided actual behavior the header doesn t have the proper padding becuase a css class is missing rendered output steps to reproduce just open the provided xhtml sample xhtml xml html xmlns xmlns h xmlns f xmlns p facelet title sample bean not needed
| 1
|
25,435
| 6,662,031,774
|
IssuesEvent
|
2017-10-02 11:21:38
|
openvstorage/framework-alba-plugin
|
https://api.github.com/repos/openvstorage/framework-alba-plugin
|
closed
|
Make it possible to assign externally managed Arakoon clusters to dedicated Backend
|
priority_normal state_codecomplete type_enhancement
|
_From @kvanhijf on August 3, 2017 12:33_
Our NSM checkup currently claims any Arakoon cluster which has the 'in_use' flag set to False when a Backend needs more NSM clusters.
This behavior is not optimal and should be improved
Possible solutions:
1) We could add a flag in the Arakoon cluster's metadata called: 'backend_name'. This flag could then be used to check for which backend the cluster was created. Our NSM checkup should then check on the name flag and act accordingly.
This could prevent that certain unused arakoon clusters, which have been deployed on 5 nodes, are registered to an ABM cluster only deployed on 3 nodes.
2) Do nothing at all when NSM checkup notices all NSM clusters being overloaded, just log a warning for monitoring to react upon.
Downsides for 1st option:
* What to do when no 'name' flag is provided? Always claim or log warning?
* What to do when unknown backend name is provided?
Downsides for 2nd option:
* We need for every NSM cluster a service in our model. This would have to become a scheduled task to keep these services in sync.
_Copied from original issue: openvstorage/framework#1711_
|
1.0
|
Make it possible to assign externally managed Arakoon clusters to dedicated Backend - _From @kvanhijf on August 3, 2017 12:33_
Our NSM checkup currently claims any Arakoon cluster which has the 'in_use' flag set to False when a Backend needs more NSM clusters.
This behavior is not optimal and should be improved
Possible solutions:
1) We could add a flag in the Arakoon cluster's metadata called: 'backend_name'. This flag could then be used to check for which backend the cluster was created. Our NSM checkup should then check on the name flag and act accordingly.
This could prevent that certain unused arakoon clusters, which have been deployed on 5 nodes, are registered to an ABM cluster only deployed on 3 nodes.
2) Do nothing at all when NSM checkup notices all NSM clusters being overloaded, just log a warning for monitoring to react upon.
Downsides for 1st option:
* What to do when no 'name' flag is provided? Always claim or log warning?
* What to do when unknown backend name is provided?
Downsides for 2nd option:
* We need for every NSM cluster a service in our model. This would have to become a scheduled task to keep these services in sync.
_Copied from original issue: openvstorage/framework#1711_
|
non_defect
|
make it possible to assign externally managed arakoon clusters to dedicated backend from kvanhijf on august our nsm checkup currently claims any arakoon cluster which has the in use flag set to false when a backend needs more nsm clusters this behavior is not optimal and should be improved possible solutions we could add a flag in the arakoon cluster s metadata called backend name this flag could then be used to check for which backend the cluster was created our nsm checkup should then check on the name flag and act accordingly this could prevent that certain unused arakoon clusters which have been deployed on nodes are registered to an abm cluster only deployed on nodes do nothing at all when nsm checkup notices all nsm clusters being overloaded just log a warning for monitoring to react upon downsides for option what to do when no name flag is provided always claim or log warning what to do when unknown backend name is provided downsides for option we need for every nsm cluster a service in our model this would have to become a scheduled task to keep these services in sync copied from original issue openvstorage framework
| 0
|
92,745
| 18,945,663,727
|
IssuesEvent
|
2021-11-18 09:54:06
|
UNIST-Almight/ps-study-2021-fall
|
https://api.github.com/repos/UNIST-Almight/ps-study-2021-fall
|
closed
|
[초급반 6회차] B번 문제 리뷰 (gracelee)
|
code review beginner week 6
|
[B번](http://boj.kr/421cbe8511104e7589a9960e52625955)
for문으로 투포인터의 첫번째 포인터(st)를 정해주고, while문을 통해 조건을 만족할 때까지 두번째 포인터(en)을 늘려서 정해줍니다.
궁금한 점은 투포인터로 O(n^2)을 O(n)으로 처리할 수 있게 해준다고 했는데, for문으로 st를 정하고 그 안에서 while문이 돌아가면 이게 왜 O(n)이 되는 건지 궁금합니다. 그래서 시간 초과가 났던 코드가 왜 시간 초과가 났는지, 제 코드는 O(n^2)으로 돌아가고 있는건지가 궁금합니다.!
|
1.0
|
[초급반 6회차] B번 문제 리뷰 (gracelee) - [B번](http://boj.kr/421cbe8511104e7589a9960e52625955)
for문으로 투포인터의 첫번째 포인터(st)를 정해주고, while문을 통해 조건을 만족할 때까지 두번째 포인터(en)을 늘려서 정해줍니다.
궁금한 점은 투포인터로 O(n^2)을 O(n)으로 처리할 수 있게 해준다고 했는데, for문으로 st를 정하고 그 안에서 while문이 돌아가면 이게 왜 O(n)이 되는 건지 궁금합니다. 그래서 시간 초과가 났던 코드가 왜 시간 초과가 났는지, 제 코드는 O(n^2)으로 돌아가고 있는건지가 궁금합니다.!
|
non_defect
|
b번 문제 리뷰 gracelee for문으로 투포인터의 첫번째 포인터 st 를 정해주고 while문을 통해 조건을 만족할 때까지 두번째 포인터 en 을 늘려서 정해줍니다 궁금한 점은 투포인터로 o n 을 o n 으로 처리할 수 있게 해준다고 했는데 for문으로 st를 정하고 그 안에서 while문이 돌아가면 이게 왜 o n 이 되는 건지 궁금합니다 그래서 시간 초과가 났던 코드가 왜 시간 초과가 났는지 제 코드는 o n 으로 돌아가고 있는건지가 궁금합니다
| 0
|
200,339
| 22,763,567,675
|
IssuesEvent
|
2022-07-08 00:23:59
|
jeffgran/granguerra
|
https://api.github.com/repos/jeffgran/granguerra
|
opened
|
sdoc-0.4.1.gem: 1 vulnerabilities (highest severity is: 7.0)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sdoc-0.4.1.gem</b></p></summary>
<p></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rdoc-4.2.0.gem</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/jeffgran/granguerra/commit/69b1396cbc478e4798268536112ce2c1054adf84">69b1396cbc478e4798268536112ce2c1054adf84</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-31799](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-31799) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.0 | rdoc-4.2.0.gem | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-31799</summary>
### Vulnerable Library - <b>rdoc-4.2.0.gem</b></p>
<p>RDoc produces HTML and command-line documentation for Ruby projects. RDoc
includes the +rdoc+ and +ri+ tools for generating and displaying documentation
from the command-line.</p>
<p>Library home page: <a href="https://rubygems.org/gems/rdoc-4.2.0.gem">https://rubygems.org/gems/rdoc-4.2.0.gem</a></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rdoc-4.2.0.gem</p>
<p>
Dependency Hierarchy:
- sdoc-0.4.1.gem (Root Library)
- :x: **rdoc-4.2.0.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jeffgran/granguerra/commit/69b1396cbc478e4798268536112ce2c1054adf84">69b1396cbc478e4798268536112ce2c1054adf84</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In RDoc 3.11 through 6.x before 6.3.1, as distributed with Ruby through 3.0.1, it is possible to execute arbitrary code via | and tags in a filename.
<p>Publish Date: 2021-07-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-31799>CVE-2021-31799</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.0</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.ruby-lang.org/en/news/2021/05/02/os-command-injection-in-rdoc/">https://www.ruby-lang.org/en/news/2021/05/02/os-command-injection-in-rdoc/</a></p>
<p>Release Date: 2021-07-30</p>
<p>Fix Resolution: rdoc - 6.3.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
sdoc-0.4.1.gem: 1 vulnerabilities (highest severity is: 7.0) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sdoc-0.4.1.gem</b></p></summary>
<p></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rdoc-4.2.0.gem</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/jeffgran/granguerra/commit/69b1396cbc478e4798268536112ce2c1054adf84">69b1396cbc478e4798268536112ce2c1054adf84</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-31799](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-31799) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.0 | rdoc-4.2.0.gem | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-31799</summary>
### Vulnerable Library - <b>rdoc-4.2.0.gem</b></p>
<p>RDoc produces HTML and command-line documentation for Ruby projects. RDoc
includes the +rdoc+ and +ri+ tools for generating and displaying documentation
from the command-line.</p>
<p>Library home page: <a href="https://rubygems.org/gems/rdoc-4.2.0.gem">https://rubygems.org/gems/rdoc-4.2.0.gem</a></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rdoc-4.2.0.gem</p>
<p>
Dependency Hierarchy:
- sdoc-0.4.1.gem (Root Library)
- :x: **rdoc-4.2.0.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jeffgran/granguerra/commit/69b1396cbc478e4798268536112ce2c1054adf84">69b1396cbc478e4798268536112ce2c1054adf84</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In RDoc 3.11 through 6.x before 6.3.1, as distributed with Ruby through 3.0.1, it is possible to execute arbitrary code via | and tags in a filename.
<p>Publish Date: 2021-07-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-31799>CVE-2021-31799</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.0</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.ruby-lang.org/en/news/2021/05/02/os-command-injection-in-rdoc/">https://www.ruby-lang.org/en/news/2021/05/02/os-command-injection-in-rdoc/</a></p>
<p>Release Date: 2021-07-30</p>
<p>Fix Resolution: rdoc - 6.3.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_defect
|
sdoc gem vulnerabilities highest severity is vulnerable library sdoc gem path to dependency file gemfile lock path to vulnerable library home wss scanner gem ruby cache rdoc gem found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high rdoc gem transitive n a details cve vulnerable library rdoc gem rdoc produces html and command line documentation for ruby projects rdoc includes the rdoc and ri tools for generating and displaying documentation from the command line library home page a href path to dependency file gemfile lock path to vulnerable library home wss scanner gem ruby cache rdoc gem dependency hierarchy sdoc gem root library x rdoc gem vulnerable library found in head commit a href found in base branch master vulnerability details in rdoc through x before as distributed with ruby through it is possible to execute arbitrary code via and tags in a filename publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rdoc step up your open source security game with mend
| 0
|
199,961
| 22,739,357,907
|
IssuesEvent
|
2022-07-07 01:05:06
|
amccool/AngularASPNETCore2WebApiAuth
|
https://api.github.com/repos/amccool/AngularASPNETCore2WebApiAuth
|
opened
|
WS-2020-0450 (Medium) detected in handlebars-4.0.11.tgz
|
security vulnerability
|
## WS-2020-0450 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.11.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz</a></p>
<p>Path to dependency file: /src/package.json</p>
<p>Path to vulnerable library: /src/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- karma-coverage-istanbul-reporter-1.3.3.tgz (Root Library)
- istanbul-api-1.2.1.tgz
- istanbul-reports-1.1.3.tgz
- :x: **handlebars-4.0.11.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/amccool/AngularASPNETCore2WebApiAuth/commit/aaa0eeb0237c3fe61b491b5e658b404be0a3a83f">aaa0eeb0237c3fe61b491b5e658b404be0a3a83f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Handlebars before 4.6.0 vulnerable to Prototype Pollution. Prototype access to the template engine allows for potential code execution, which may lead to Denial Of Service (DoS).
<p>Publish Date: 2020-01-09
<p>URL: <a href=https://github.com/handlebars-lang/handlebars.js/commit/33a3b46bc205f768f8edbc67241c68591fe3472c>WS-2020-0450</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-01-09</p>
<p>Fix Resolution (handlebars): 4.1.2-0</p>
<p>Direct dependency fix Resolution (karma-coverage-istanbul-reporter): 1.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2020-0450 (Medium) detected in handlebars-4.0.11.tgz - ## WS-2020-0450 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.11.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz</a></p>
<p>Path to dependency file: /src/package.json</p>
<p>Path to vulnerable library: /src/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- karma-coverage-istanbul-reporter-1.3.3.tgz (Root Library)
- istanbul-api-1.2.1.tgz
- istanbul-reports-1.1.3.tgz
- :x: **handlebars-4.0.11.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/amccool/AngularASPNETCore2WebApiAuth/commit/aaa0eeb0237c3fe61b491b5e658b404be0a3a83f">aaa0eeb0237c3fe61b491b5e658b404be0a3a83f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Handlebars before 4.6.0 vulnerable to Prototype Pollution. Prototype access to the template engine allows for potential code execution, which may lead to Denial Of Service (DoS).
<p>Publish Date: 2020-01-09
<p>URL: <a href=https://github.com/handlebars-lang/handlebars.js/commit/33a3b46bc205f768f8edbc67241c68591fe3472c>WS-2020-0450</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-01-09</p>
<p>Fix Resolution (handlebars): 4.1.2-0</p>
<p>Direct dependency fix Resolution (karma-coverage-istanbul-reporter): 1.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
ws medium detected in handlebars tgz ws medium severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file src package json path to vulnerable library src node modules handlebars package json dependency hierarchy karma coverage istanbul reporter tgz root library istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href vulnerability details handlebars before vulnerable to prototype pollution prototype access to the template engine allows for potential code execution which may lead to denial of service dos publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution handlebars direct dependency fix resolution karma coverage istanbul reporter step up your open source security game with mend
| 0
|
213,688
| 24,016,306,726
|
IssuesEvent
|
2022-09-15 01:16:51
|
brogers588/netflix_conductor_fork
|
https://api.github.com/repos/brogers588/netflix_conductor_fork
|
opened
|
WS-2020-0042 (High) detected in acorn-5.7.4.tgz
|
security vulnerability
|
## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.4.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz</a></p>
<p>
Dependency Hierarchy:
- react-0.14.9.tgz (Root Library)
- envify-3.4.1.tgz
- jstransform-11.0.3.tgz
- commoner-0.10.8.tgz
- detective-4.7.1.tgz
- :x: **acorn-5.7.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/brogers588/netflix_conductor_fork/commit/975ea99358eaa6f34b7c8c0c0dce2a0a92a39da5">975ea99358eaa6f34b7c8c0c0dce2a0a92a39da5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-01
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-01</p>
<p>Fix Resolution (acorn): 6.4.1</p>
<p>Direct dependency fix Resolution (react): 15.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2020-0042 (High) detected in acorn-5.7.4.tgz - ## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>acorn-5.7.4.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.4.tgz</a></p>
<p>
Dependency Hierarchy:
- react-0.14.9.tgz (Root Library)
- envify-3.4.1.tgz
- jstransform-11.0.3.tgz
- commoner-0.10.8.tgz
- detective-4.7.1.tgz
- :x: **acorn-5.7.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/brogers588/netflix_conductor_fork/commit/975ea99358eaa6f34b7c8c0c0dce2a0a92a39da5">975ea99358eaa6f34b7c8c0c0dce2a0a92a39da5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-01
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-01</p>
<p>Fix Resolution (acorn): 6.4.1</p>
<p>Direct dependency fix Resolution (react): 15.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
ws high detected in acorn tgz ws high severity vulnerability vulnerable library acorn tgz ecmascript parser library home page a href dependency hierarchy react tgz root library envify tgz jstransform tgz commoner tgz detective tgz x acorn tgz vulnerable library found in head commit a href found in base branch master vulnerability details acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution acorn direct dependency fix resolution react step up your open source security game with mend
| 0
|
71,650
| 23,741,982,067
|
IssuesEvent
|
2022-08-31 13:11:21
|
SeleniumHQ/selenium
|
https://api.github.com/repos/SeleniumHQ/selenium
|
opened
|
[🐛 Bug]: Intercept Network Fetch responses using JavaScript
|
I-defect needs-triaging
|
### What happened?
Hi, I've seen that it is possible to intercept or retrieve network responses using Selenium and Chrome DevTools in Java. However, nowhere can I find an example of how to do this in JavaScript, is it not possible?
Example: https://stackoverflow.com/questions/68668853/how-to-get-the-response-body-from-network-tab-using-selenium-4-devtools
The above is a java example.
I have the goal of retrieving the response strings from requests such as those that can be seen over the Chrome browser Network tab.
I do not intend to manipulate the response, I only wish to be able to extract these strings to a log along with their relevant request URL string.
I have made numerous efforts using cdpConnection and selenium-webdriver (also using sendAndGetDevToolsCommand)...
However, I have been only unsuccessful in this effort and cannot find any documentation or examples on how to accomplish this over JavaScript. Any help would be greatly appreciated.
### How can we reproduce the issue?
```shell
<Asking for documentation and example, no reproduction necessary>
```
### Relevant log output
```shell
<Not applicable>
```
### Operating System
Windows
### Selenium version
Node v14.16.1 javascript
### What are the browser(s) and version(s) where you see this issue?
Chrome 104
### What are the browser driver(s) and version(s) where you see this issue?
ChromeDriver 104
### Are you using Selenium Grid?
_No response_
|
1.0
|
[🐛 Bug]: Intercept Network Fetch responses using JavaScript - ### What happened?
Hi, I've seen that it is possible to intercept or retrieve network responses using Selenium and Chrome DevTools in Java. However, nowhere can I find an example of how to do this in JavaScript, is it not possible?
Example: https://stackoverflow.com/questions/68668853/how-to-get-the-response-body-from-network-tab-using-selenium-4-devtools
The above is a java example.
I have the goal of retrieving the response strings from requests such as those that can be seen over the Chrome browser Network tab.
I do not intend to manipulate the response, I only wish to be able to extract these strings to a log along with their relevant request URL string.
I have made numerous efforts using cdpConnection and selenium-webdriver (also using sendAndGetDevToolsCommand)...
However, I have been only unsuccessful in this effort and cannot find any documentation or examples on how to accomplish this over JavaScript. Any help would be greatly appreciated.
### How can we reproduce the issue?
```shell
<Asking for documentation and example, no reproduction necessary>
```
### Relevant log output
```shell
<Not applicable>
```
### Operating System
Windows
### Selenium version
Node v14.16.1 javascript
### What are the browser(s) and version(s) where you see this issue?
Chrome 104
### What are the browser driver(s) and version(s) where you see this issue?
ChromeDriver 104
### Are you using Selenium Grid?
_No response_
|
defect
|
intercept network fetch responses using javascript what happened hi i ve seen that it is possible to intercept or retrieve network responses using selenium and chrome devtools in java however nowhere can i find an example of how to do this in javascript is it not possible example the above is a java example i have the goal of retrieving the response strings from requests such as those that can be seen over the chrome browser network tab i do not intend to manipulate the response i only wish to be able to extract these strings to a log along with their relevant request url string i have made numerous efforts using cdpconnection and selenium webdriver also using sendandgetdevtoolscommand however i have been only unsuccessful in this effort and cannot find any documentation or examples on how to accomplish this over javascript any help would be greatly appreciated how can we reproduce the issue shell relevant log output shell operating system windows selenium version node javascript what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue chromedriver are you using selenium grid no response
| 1
|
52,651
| 13,224,881,516
|
IssuesEvent
|
2020-08-17 20:02:19
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
ithon and tarballing broken (Trac #106)
|
Migrated from Trac cmake defect
|
ithon has a new destination path, lib/python/ithon.so, and this doesn't tarball correctly
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/106">https://code.icecube.wisc.edu/projects/icecube/ticket/106</a>, reported by troyand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-09T22:33:45",
"_ts": "1194647625000000",
"description": "ithon has a new destination path, lib/python/ithon.so, and this doesn't tarball correctly\n",
"reporter": "troy",
"cc": "",
"resolution": "duplicate",
"time": "2007-08-30T19:03:32",
"component": "cmake",
"summary": "ithon and tarballing broken",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
ithon and tarballing broken (Trac #106) - ithon has a new destination path, lib/python/ithon.so, and this doesn't tarball correctly
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/106">https://code.icecube.wisc.edu/projects/icecube/ticket/106</a>, reported by troyand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-09T22:33:45",
"_ts": "1194647625000000",
"description": "ithon has a new destination path, lib/python/ithon.so, and this doesn't tarball correctly\n",
"reporter": "troy",
"cc": "",
"resolution": "duplicate",
"time": "2007-08-30T19:03:32",
"component": "cmake",
"summary": "ithon and tarballing broken",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
|
defect
|
ithon and tarballing broken trac ithon has a new destination path lib python ithon so and this doesn t tarball correctly migrated from json status closed changetime ts description ithon has a new destination path lib python ithon so and this doesn t tarball correctly n reporter troy cc resolution duplicate time component cmake summary ithon and tarballing broken priority normal keywords milestone owner troy type defect
| 1
|
37,829
| 8,528,313,806
|
IssuesEvent
|
2018-11-02 23:05:18
|
alex-hhh/ActivityLog2
|
https://api.github.com/repos/alex-hhh/ActivityLog2
|
closed
|
Histogram inspector does not remember its settings
|
defect
|
This happens occasionally, and I don't have the steps to reproduce, but, occasionally, the histogram options don't seem to be in sync with what is drawn in the plot. For example, "color by zone" is applied to the histogram when the respective checkbox is not selected, or the outliers are not trimmed when there is a percentage in the "trim outliers" input field. Need to review the code to determine what is going on.
|
1.0
|
Histogram inspector does not remember its settings - This happens occasionally, and I don't have the steps to reproduce, but, occasionally, the histogram options don't seem to be in sync with what is drawn in the plot. For example, "color by zone" is applied to the histogram when the respective checkbox is not selected, or the outliers are not trimmed when there is a percentage in the "trim outliers" input field. Need to review the code to determine what is going on.
|
defect
|
histogram inspector does not remember its settings this happens occasionally and i don t have the steps to reproduce but occasionally the histogram options don t seem to be in sync with what is drawn in the plot for example color by zone is applied to the histogram when the respective checkbox is not selected or the outliers are not trimmed when there is a percentage in the trim outliers input field need to review the code to determine what is going on
| 1
|
40,193
| 9,887,963,669
|
IssuesEvent
|
2019-06-25 10:22:12
|
vector-im/riot-web
|
https://api.github.com/repos/vector-im/riot-web
|
opened
|
Forward message modal looks strange
|
bug defect needs-design p2 ui/ux
|
I am not sure what the intended appearance is, but at the moment, when you choose forward message, we show some text and the top of the screen and an oddly placed close button:
<img width="791" alt="2019-06-25 at 11 15" src="https://user-images.githubusercontent.com/279572/60090805-1ef8d780-973b-11e9-93e1-73f1eaac0260.png">
We currently appear to have no CSS for `mx_ForwardMessage`. We could do some archaeology to find what was their historically, but it might be easier to just design what we want to have now.
|
1.0
|
Forward message modal looks strange - I am not sure what the intended appearance is, but at the moment, when you choose forward message, we show some text and the top of the screen and an oddly placed close button:
<img width="791" alt="2019-06-25 at 11 15" src="https://user-images.githubusercontent.com/279572/60090805-1ef8d780-973b-11e9-93e1-73f1eaac0260.png">
We currently appear to have no CSS for `mx_ForwardMessage`. We could do some archaeology to find what was their historically, but it might be easier to just design what we want to have now.
|
defect
|
forward message modal looks strange i am not sure what the intended appearance is but at the moment when you choose forward message we show some text and the top of the screen and an oddly placed close button img width alt at src we currently appear to have no css for mx forwardmessage we could do some archaeology to find what was their historically but it might be easier to just design what we want to have now
| 1
|
416,470
| 12,147,065,218
|
IssuesEvent
|
2020-04-24 12:22:18
|
googleapis/repo-automation-bots
|
https://api.github.com/repos/googleapis/repo-automation-bots
|
opened
|
Synthesis failed for repo-automation-bots
|
autosynth failure priority: p1 type: bug
|
Hello! Autosynth couldn't regenerate repo-automation-bots. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to a new branch 'autosynth'
Cloning into '/tmpfs/tmp/tmpc5n3m331/synthtool'...
Note: checking out '3a070bedf9de0c6fa587f6427ac9e66cb8854435'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 3a070be chore: update npm scripts and synth.py (#497)
Note: checking out '716f741f2d307b48cbe8a5bc3bc883571212344a'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 716f741 fix(python): adjust regex for fix_pb2_headers (#500)
Switched to a new branch 'autosynth-1'
2020-04-24 05:19:31 [INFO] Running synthtool
2020-04-24 05:19:31 [INFO] ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-24 05:19:31,274 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/synth.py.
On branch autosynth-1
nothing to commit, working tree clean
2020-04-24 05:19:31,335 synthtool > Using precloned repo /tmpfs/tmp/tmpc5n3m331/synthtool
.eslintignore
.eslintrc.json
.github/ISSUE_TEMPLATE/bug_report.md
.github/ISSUE_TEMPLATE/feature_request.md
.github/ISSUE_TEMPLATE/support_request.md
.github/PULL_REQUEST_TEMPLATE.md
.github/publish.yml
.github/release-please.yml
.github/workflows/ci.yaml
.kokoro/common.cfg
.kokoro/continuous/node10/common.cfg
.kokoro/continuous/node10/docs.cfg
.kokoro/continuous/node10/lint.cfg
.kokoro/continuous/node10/samples-test.cfg
.kokoro/continuous/node10/system-test.cfg
.kokoro/continuous/node10/test.cfg
.kokoro/continuous/node12/common.cfg
.kokoro/continuous/node12/test.cfg
.kokoro/continuous/node8/common.cfg
.kokoro/continuous/node8/test.cfg
.kokoro/docs.sh
.kokoro/lint.sh
.kokoro/presubmit/node10/common.cfg
.kokoro/presubmit/node10/docs.cfg
.kokoro/presubmit/node10/lint.cfg
.kokoro/presubmit/node10/samples-test.cfg
.kokoro/presubmit/node10/system-test.cfg
.kokoro/presubmit/node10/test.cfg
.kokoro/presubmit/node12/common.cfg
.kokoro/presubmit/node12/test.cfg
.kokoro/presubmit/node8/common.cfg
.kokoro/presubmit/node8/test.cfg
.kokoro/presubmit/windows/common.cfg
.kokoro/presubmit/windows/test.cfg
.kokoro/publish.sh
.kokoro/release/docs.cfg
.kokoro/release/docs.sh
.kokoro/release/publish.cfg
.kokoro/samples-test.sh
.kokoro/system-test.sh
.kokoro/test.bat
.kokoro/test.sh
.kokoro/trampoline.sh
.mocharc.js
.nycrc
.prettierignore
.prettierrc.js
CODE_OF_CONDUCT.md
CONTRIBUTING.md
LICENSE
README.md
renovate.json
Skipping: samples/README.md
2020-04-24 05:19:31,520 synthtool > Installing dependencies...
npm WARN npm npm does not support Node.js v12.16.2
npm WARN npm You should probably upgrade to a newer version of node as we
npm WARN npm can't make any promises that npm will work with this version.
npm WARN npm Supported releases of Node.js are the latest release of 6, 8, 9, 10, 11.
npm WARN npm You can find the latest version at https://nodejs.org/
added 106 packages from 86 contributors and audited 162 packages in 2.992s
found 0 vulnerabilities
2020-04-24 05:19:34,849 synthtool > Running prelint...
npm WARN npm npm does not support Node.js v12.16.2
npm WARN npm You should probably upgrade to a newer version of node as we
npm WARN npm can't make any promises that npm will work with this version.
npm WARN npm Supported releases of Node.js are the latest release of 6, 8, 9, 10, 11.
npm WARN npm You can find the latest version at https://nodejs.org/
npm ERR! missing script: prelint
npm ERR! A complete log of this run can be found in:
npm ERR! /home/kbuilder/.npm/_logs/2020-04-24T12_19_35_011Z-debug.log
2020-04-24 05:19:35,021 synthtool > Running fix...
npm WARN npm npm does not support Node.js v12.16.2
npm WARN npm You should probably upgrade to a newer version of node as we
npm WARN npm can't make any promises that npm will work with this version.
npm WARN npm Supported releases of Node.js are the latest release of 6, 8, 9, 10, 11.
npm WARN npm You can find the latest version at https://nodejs.org/
> repo-automation-bots@ fix /tmpfs/src/git/autosynth/working_repo
> gts fix
sh: 1: gts: not found
npm ERR! file sh
npm ERR! code ELIFECYCLE
npm ERR! errno ENOENT
npm ERR! syscall spawn
npm ERR! repo-automation-bots@ fix: `gts fix`
npm ERR! spawn ENOENT
npm ERR!
npm ERR! Failed at the repo-automation-bots@ fix script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! /home/kbuilder/.npm/_logs/2020-04-24T12_19_35_196Z-debug.log
2020-04-24 05:19:35,204 synthtool > Failed executing npm run fix:
None
2020-04-24 05:19:35,205 synthtool > Wrote metadata to synth.metadata.
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/tmpfs/src/git/autosynth/working_repo/synth.py", line 26, in <module>
node.fix()
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/languages/node.py", line 114, in fix
shell.run(["npm", "run", "fix"], hide_output=hide_output)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['npm', 'run', 'fix']' returned non-zero exit status 1.
2020-04-24 05:19:35 [ERROR] Synthesis failed
HEAD is now at 3a070be chore: update npm scripts and synth.py (#497)
Switched to branch 'autosynth'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 584, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 465, in main
return _inner_main(temp_dir)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 574, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 358, in synthesize_loop
synthesize_inner_loop(toolbox, synthesizer)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 368, in synthesize_inner_loop
synthesizer, len(toolbox.versions) - 1
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 259, in synthesize_version_in_new_branch
synthesizer.synthesize(self.environ)
File "/tmpfs/src/git/autosynth/autosynth/synthesizer.py", line 115, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/07cef893-0724-4dfe-840a-4ede71699211).
|
1.0
|
Synthesis failed for repo-automation-bots - Hello! Autosynth couldn't regenerate repo-automation-bots. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to a new branch 'autosynth'
Cloning into '/tmpfs/tmp/tmpc5n3m331/synthtool'...
Note: checking out '3a070bedf9de0c6fa587f6427ac9e66cb8854435'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 3a070be chore: update npm scripts and synth.py (#497)
Note: checking out '716f741f2d307b48cbe8a5bc3bc883571212344a'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 716f741 fix(python): adjust regex for fix_pb2_headers (#500)
Switched to a new branch 'autosynth-1'
2020-04-24 05:19:31 [INFO] Running synthtool
2020-04-24 05:19:31 [INFO] ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-24 05:19:31,274 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/synth.py.
On branch autosynth-1
nothing to commit, working tree clean
2020-04-24 05:19:31,335 synthtool > Using precloned repo /tmpfs/tmp/tmpc5n3m331/synthtool
.eslintignore
.eslintrc.json
.github/ISSUE_TEMPLATE/bug_report.md
.github/ISSUE_TEMPLATE/feature_request.md
.github/ISSUE_TEMPLATE/support_request.md
.github/PULL_REQUEST_TEMPLATE.md
.github/publish.yml
.github/release-please.yml
.github/workflows/ci.yaml
.kokoro/common.cfg
.kokoro/continuous/node10/common.cfg
.kokoro/continuous/node10/docs.cfg
.kokoro/continuous/node10/lint.cfg
.kokoro/continuous/node10/samples-test.cfg
.kokoro/continuous/node10/system-test.cfg
.kokoro/continuous/node10/test.cfg
.kokoro/continuous/node12/common.cfg
.kokoro/continuous/node12/test.cfg
.kokoro/continuous/node8/common.cfg
.kokoro/continuous/node8/test.cfg
.kokoro/docs.sh
.kokoro/lint.sh
.kokoro/presubmit/node10/common.cfg
.kokoro/presubmit/node10/docs.cfg
.kokoro/presubmit/node10/lint.cfg
.kokoro/presubmit/node10/samples-test.cfg
.kokoro/presubmit/node10/system-test.cfg
.kokoro/presubmit/node10/test.cfg
.kokoro/presubmit/node12/common.cfg
.kokoro/presubmit/node12/test.cfg
.kokoro/presubmit/node8/common.cfg
.kokoro/presubmit/node8/test.cfg
.kokoro/presubmit/windows/common.cfg
.kokoro/presubmit/windows/test.cfg
.kokoro/publish.sh
.kokoro/release/docs.cfg
.kokoro/release/docs.sh
.kokoro/release/publish.cfg
.kokoro/samples-test.sh
.kokoro/system-test.sh
.kokoro/test.bat
.kokoro/test.sh
.kokoro/trampoline.sh
.mocharc.js
.nycrc
.prettierignore
.prettierrc.js
CODE_OF_CONDUCT.md
CONTRIBUTING.md
LICENSE
README.md
renovate.json
Skipping: samples/README.md
2020-04-24 05:19:31,520 synthtool > Installing dependencies...
npm WARN npm npm does not support Node.js v12.16.2
npm WARN npm You should probably upgrade to a newer version of node as we
npm WARN npm can't make any promises that npm will work with this version.
npm WARN npm Supported releases of Node.js are the latest release of 6, 8, 9, 10, 11.
npm WARN npm You can find the latest version at https://nodejs.org/
added 106 packages from 86 contributors and audited 162 packages in 2.992s
found 0 vulnerabilities
2020-04-24 05:19:34,849 synthtool > Running prelint...
npm WARN npm npm does not support Node.js v12.16.2
npm WARN npm You should probably upgrade to a newer version of node as we
npm WARN npm can't make any promises that npm will work with this version.
npm WARN npm Supported releases of Node.js are the latest release of 6, 8, 9, 10, 11.
npm WARN npm You can find the latest version at https://nodejs.org/
npm ERR! missing script: prelint
npm ERR! A complete log of this run can be found in:
npm ERR! /home/kbuilder/.npm/_logs/2020-04-24T12_19_35_011Z-debug.log
2020-04-24 05:19:35,021 synthtool > Running fix...
npm WARN npm npm does not support Node.js v12.16.2
npm WARN npm You should probably upgrade to a newer version of node as we
npm WARN npm can't make any promises that npm will work with this version.
npm WARN npm Supported releases of Node.js are the latest release of 6, 8, 9, 10, 11.
npm WARN npm You can find the latest version at https://nodejs.org/
> repo-automation-bots@ fix /tmpfs/src/git/autosynth/working_repo
> gts fix
sh: 1: gts: not found
npm ERR! file sh
npm ERR! code ELIFECYCLE
npm ERR! errno ENOENT
npm ERR! syscall spawn
npm ERR! repo-automation-bots@ fix: `gts fix`
npm ERR! spawn ENOENT
npm ERR!
npm ERR! Failed at the repo-automation-bots@ fix script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! /home/kbuilder/.npm/_logs/2020-04-24T12_19_35_196Z-debug.log
2020-04-24 05:19:35,204 synthtool > Failed executing npm run fix:
None
2020-04-24 05:19:35,205 synthtool > Wrote metadata to synth.metadata.
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/tmpfs/src/git/autosynth/working_repo/synth.py", line 26, in <module>
node.fix()
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/languages/node.py", line 114, in fix
shell.run(["npm", "run", "fix"], hide_output=hide_output)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['npm', 'run', 'fix']' returned non-zero exit status 1.
2020-04-24 05:19:35 [ERROR] Synthesis failed
HEAD is now at 3a070be chore: update npm scripts and synth.py (#497)
Switched to branch 'autosynth'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 584, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 465, in main
return _inner_main(temp_dir)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 574, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 358, in synthesize_loop
synthesize_inner_loop(toolbox, synthesizer)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 368, in synthesize_inner_loop
synthesizer, len(toolbox.versions) - 1
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 259, in synthesize_version_in_new_branch
synthesizer.synthesize(self.environ)
File "/tmpfs/src/git/autosynth/autosynth/synthesizer.py", line 115, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/07cef893-0724-4dfe-840a-4ede71699211).
|
non_defect
|
synthesis failed for repo automation bots hello autosynth couldn t regenerate repo automation bots broken heart here s the output from running synth py cloning into working repo switched to a new branch autosynth cloning into tmpfs tmp synthtool note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at chore update npm scripts and synth py note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at fix python adjust regex for fix headers switched to a new branch autosynth running synthtool synthtool executing tmpfs src git autosynth working repo synth py on branch autosynth nothing to commit working tree clean synthtool using precloned repo tmpfs tmp synthtool eslintignore eslintrc json github issue template bug report md github issue template feature request md github issue template support request md github pull request template md github publish yml github release please yml github workflows ci yaml kokoro common cfg kokoro continuous common cfg kokoro continuous docs cfg kokoro continuous lint cfg kokoro continuous samples test cfg kokoro continuous system test cfg kokoro continuous test cfg kokoro continuous common cfg kokoro continuous test cfg kokoro continuous common cfg kokoro continuous test cfg kokoro docs sh kokoro lint sh kokoro presubmit common cfg kokoro presubmit docs cfg kokoro presubmit lint cfg kokoro presubmit samples test cfg kokoro presubmit system test cfg kokoro presubmit test cfg kokoro presubmit common cfg kokoro presubmit test cfg kokoro presubmit common cfg kokoro presubmit test cfg kokoro presubmit windows common cfg kokoro presubmit windows test cfg kokoro publish sh kokoro release docs cfg kokoro release docs sh kokoro release publish cfg kokoro samples test sh kokoro system test sh kokoro test bat kokoro test sh kokoro trampoline sh mocharc js nycrc prettierignore prettierrc js code of conduct md contributing md license readme md renovate json skipping samples readme md synthtool installing dependencies npm warn npm npm does not support node js npm warn npm you should probably upgrade to a newer version of node as we npm warn npm can t make any promises that npm will work with this version npm warn npm supported releases of node js are the latest release of npm warn npm you can find the latest version at added packages from contributors and audited packages in found vulnerabilities synthtool running prelint npm warn npm npm does not support node js npm warn npm you should probably upgrade to a newer version of node as we npm warn npm can t make any promises that npm will work with this version npm warn npm supported releases of node js are the latest release of npm warn npm you can find the latest version at npm err missing script prelint npm err a complete log of this run can be found in npm err home kbuilder npm logs debug log synthtool running fix npm warn npm npm does not support node js npm warn npm you should probably upgrade to a newer version of node as we npm warn npm can t make any promises that npm will work with this version npm warn npm supported releases of node js are the latest release of npm warn npm you can find the latest version at repo automation bots fix tmpfs src git autosynth working repo gts fix sh gts not found npm err file sh npm err code elifecycle npm err errno enoent npm err syscall spawn npm err repo automation bots fix gts fix npm err spawn enoent npm err npm err failed at the repo automation bots fix script npm err this is probably not a problem with npm there is likely additional logging output above npm err a complete log of this run can be found in npm err home kbuilder npm logs debug log synthtool failed executing npm run fix none synthtool wrote metadata to synth metadata traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth env lib site packages synthtool main py line in main file tmpfs src git autosynth env lib site packages click core py line in call return self main args kwargs file tmpfs src git autosynth env lib site packages click core py line in main rv self invoke ctx file tmpfs src git autosynth env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src git autosynth env lib site packages click core py line in invoke return callback args kwargs file tmpfs src git autosynth env lib site packages synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file tmpfs src git autosynth working repo synth py line in node fix file tmpfs src git autosynth env lib site packages synthtool languages node py line in fix shell run hide output hide output file tmpfs src git autosynth env lib site packages synthtool shell py line in run raise exc file tmpfs src git autosynth env lib site packages synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status synthesis failed head is now at chore update npm scripts and synth py switched to branch autosynth traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main return inner main temp dir file tmpfs src git autosynth autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src git autosynth autosynth synth py line in synthesize loop synthesize inner loop toolbox synthesizer file tmpfs src git autosynth autosynth synth py line in synthesize inner loop synthesizer len toolbox versions file tmpfs src git autosynth autosynth synth py line in synthesize version in new branch synthesizer synthesize self environ file tmpfs src git autosynth autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
| 0
|
15,808
| 2,869,074,438
|
IssuesEvent
|
2015-06-05 23:07:29
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
Running JS version of Chessboard example does not resize board
|
Area-Pkg Pkg-Polymer PolymerMilestone-Next Priority-Medium Triaged Type-Defect
|
*This issue was originally filed by aforsel...@gmail.com*
_____
**What steps will reproduce the problem?**
1. Clone https://github.com/andersforsell/chessboard.dart.git
2. Run example/app.html in Dartium and verify that the chessboard is resized when resizing browser window
3. Run example/app.html as JS in Chrome and now the chessboard is NOT resized when resizing browser window
**What is the expected output? What do you see instead?**
The JS version should resize chessboard as well.
**What version of the product are you using? On what operating system?**
I am in OSX 10.9.3 and using Dart 1.6.0-dev.1.2
**Please provide any additional information below.**
|
1.0
|
Running JS version of Chessboard example does not resize board - *This issue was originally filed by aforsel...@gmail.com*
_____
**What steps will reproduce the problem?**
1. Clone https://github.com/andersforsell/chessboard.dart.git
2. Run example/app.html in Dartium and verify that the chessboard is resized when resizing browser window
3. Run example/app.html as JS in Chrome and now the chessboard is NOT resized when resizing browser window
**What is the expected output? What do you see instead?**
The JS version should resize chessboard as well.
**What version of the product are you using? On what operating system?**
I am in OSX 10.9.3 and using Dart 1.6.0-dev.1.2
**Please provide any additional information below.**
|
defect
|
running js version of chessboard example does not resize board this issue was originally filed by aforsel gmail com what steps will reproduce the problem clone run example app html in dartium and verify that the chessboard is resized when resizing browser window run example app html as js in chrome and now the chessboard is not resized when resizing browser window what is the expected output what do you see instead the js version should resize chessboard as well what version of the product are you using on what operating system i am in osx and using dart dev please provide any additional information below
| 1
|
19,655
| 3,228,805,606
|
IssuesEvent
|
2015-10-12 07:12:26
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
closed
|
[TEST-FAILURE] MigrationAwareServiceTest timeout
|
Team: Core Type: Defect
|
```
06:09:29 Running com.hazelcast.spi.MigrationAwareServiceTest
07:53:43 Build timed out (after 150 minutes). Marking the build as aborted.
```
https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-IbmJDK1.6/583/console
|
1.0
|
[TEST-FAILURE] MigrationAwareServiceTest timeout - ```
06:09:29 Running com.hazelcast.spi.MigrationAwareServiceTest
07:53:43 Build timed out (after 150 minutes). Marking the build as aborted.
```
https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-IbmJDK1.6/583/console
|
defect
|
migrationawareservicetest timeout running com hazelcast spi migrationawareservicetest build timed out after minutes marking the build as aborted
| 1
|
224,341
| 17,689,717,135
|
IssuesEvent
|
2021-08-24 08:27:40
|
quarkusio/quarkus
|
https://api.github.com/repos/quarkusio/quarkus
|
closed
|
No (mock) injection when using @Nested JUnit tests
|
kind/bug area/testing
|
### Describe the bug
Mock injection is not performed before execution of unit tests inside nested test classes.
### Expected behavior
Attributes of the top level class annotated with `@InjectMock` are injected before the tests of `@Nested` test classes are executed.
### Actual behavior
Attributes annotated with `@InjectMock` are `null`.
### How to Reproduce?
A test like this fails because the `service` is not injected:
```java
@QuarkusTest
class TestClass {
@InjectMock
private MyService service;
@Nested
class NestedTest {
@Test
void myTest() {
assertNotNull(service);
}
}
}
```
And it passes if the nested class wrapper is removed.
### Output of `uname -a` or `ver`
2.1.0.Final
### Output of `java -version`
11
### GraalVM version (if different from Java)
_No response_
### Quarkus version or git rev
_No response_
### Build tool (ie. output of `mvnw --version` or `gradlew --version`)
Maven
### Additional information
_No response_
|
1.0
|
No (mock) injection when using @Nested JUnit tests - ### Describe the bug
Mock injection is not performed before execution of unit tests inside nested test classes.
### Expected behavior
Attributes of the top level class annotated with `@InjectMock` are injected before the tests of `@Nested` test classes are executed.
### Actual behavior
Attributes annotated with `@InjectMock` are `null`.
### How to Reproduce?
A test like this fails because the `service` is not injected:
```java
@QuarkusTest
class TestClass {
@InjectMock
private MyService service;
@Nested
class NestedTest {
@Test
void myTest() {
assertNotNull(service);
}
}
}
```
And it passes if the nested class wrapper is removed.
### Output of `uname -a` or `ver`
2.1.0.Final
### Output of `java -version`
11
### GraalVM version (if different from Java)
_No response_
### Quarkus version or git rev
_No response_
### Build tool (ie. output of `mvnw --version` or `gradlew --version`)
Maven
### Additional information
_No response_
|
non_defect
|
no mock injection when using nested junit tests describe the bug mock injection is not performed before execution of unit tests inside nested test classes expected behavior attributes of the top level class annotated with injectmock are injected before the tests of nested test classes are executed actual behavior attributes annotated with injectmock are null how to reproduce a test like this fails because the service is not injected java quarkustest class testclass injectmock private myservice service nested class nestedtest test void mytest assertnotnull service and it passes if the nested class wrapper is removed output of uname a or ver final output of java version graalvm version if different from java no response quarkus version or git rev no response build tool ie output of mvnw version or gradlew version maven additional information no response
| 0
|
8,412
| 2,611,496,541
|
IssuesEvent
|
2015-02-27 05:35:55
|
chrsmith/hedgewars
|
https://api.github.com/repos/chrsmith/hedgewars
|
closed
|
DLC list is not hidden, just moved away from the screen
|
auto-migrated Priority-Minor Type-Defect
|
```
What steps will reproduce the problem?
1. Resize hedgewars to a small window
2. Enter DLC from network room
3. Leave DLC
4. Maximize hedgewars
5. Here's the result: http://img.vos.uz/65ww.png
What is the expected output? What do you see instead?
DLC is hidden entirely
What version of the product are you using? On what operating system?
0.9.18-7740-4ba77e6178cd / win32
```
Original issue reported on code.google.com by `v...@vos.uz` on 26 Oct 2012 at 6:53
|
1.0
|
DLC list is not hidden, just moved away from the screen - ```
What steps will reproduce the problem?
1. Resize hedgewars to a small window
2. Enter DLC from network room
3. Leave DLC
4. Maximize hedgewars
5. Here's the result: http://img.vos.uz/65ww.png
What is the expected output? What do you see instead?
DLC is hidden entirely
What version of the product are you using? On what operating system?
0.9.18-7740-4ba77e6178cd / win32
```
Original issue reported on code.google.com by `v...@vos.uz` on 26 Oct 2012 at 6:53
|
defect
|
dlc list is not hidden just moved away from the screen what steps will reproduce the problem resize hedgewars to a small window enter dlc from network room leave dlc maximize hedgewars here s the result what is the expected output what do you see instead dlc is hidden entirely what version of the product are you using on what operating system original issue reported on code google com by v vos uz on oct at
| 1
|
258,419
| 27,563,970,767
|
IssuesEvent
|
2023-03-08 01:19:42
|
panasalap/linux-4.1.15
|
https://api.github.com/repos/panasalap/linux-4.1.15
|
opened
|
CVE-2023-1095 (Medium) detected in linuxlinux-4.1.17
|
security vulnerability
|
## CVE-2023-1095 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.17</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_tables_api.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_tables_api.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In nf_tables_updtable, if nf_tables_table_enable returns an error, nft_trans_destroy is called to free the transaction object. nft_trans_destroy() calls list_del(), but the transaction was never placed on a list -- the list head is all zeroes, this results in a NULL pointer dereference.
<p>Publish Date: 2023-02-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1095>CVE-2023-1095</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1095">https://www.linuxkernelcves.com/cves/CVE-2023-1095</a></p>
<p>Release Date: 2023-02-28</p>
<p>Fix Resolution: v4.9.326,v4.14.291,v4.19.256,v5.4.211,v5.10.137,v5.15.61,v5.18.18</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-1095 (Medium) detected in linuxlinux-4.1.17 - ## CVE-2023-1095 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.17</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_tables_api.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_tables_api.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In nf_tables_updtable, if nf_tables_table_enable returns an error, nft_trans_destroy is called to free the transaction object. nft_trans_destroy() calls list_del(), but the transaction was never placed on a list -- the list head is all zeroes, this results in a NULL pointer dereference.
<p>Publish Date: 2023-02-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1095>CVE-2023-1095</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1095">https://www.linuxkernelcves.com/cves/CVE-2023-1095</a></p>
<p>Release Date: 2023-02-28</p>
<p>Fix Resolution: v4.9.326,v4.14.291,v4.19.256,v5.4.211,v5.10.137,v5.15.61,v5.18.18</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in base branch master vulnerable source files net netfilter nf tables api c net netfilter nf tables api c vulnerability details in nf tables updtable if nf tables table enable returns an error nft trans destroy is called to free the transaction object nft trans destroy calls list del but the transaction was never placed on a list the list head is all zeroes this results in a null pointer dereference publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
65,617
| 19,595,897,749
|
IssuesEvent
|
2022-01-05 17:48:17
|
SeleniumHQ/selenium
|
https://api.github.com/repos/SeleniumHQ/selenium
|
opened
|
[🐛 Bug]: Grid-Url in TOML node-config file is not passed through
|
I-defect needs-triaging
|
### What happened?
In Selenium GRID 3.x I configured the hub and node with the Json config files.
Now I want to configure Selenium GRID 4.x with TOML configuration file for the nodes and need to setup parameters (grid-url, max connections) as specified.
Either the documentation seems to be misleading or the parameter is not correctly passed through:
Section "Node", parameter "--grid-url"
https://www.selenium.dev/documentation/grid/configuration/cli_options/
Each node must connect to the grid hub instance, which is located on a different machine within my network.
When I start the node instance with the config-file it does not connect to the given URL and it seems to try to connect to my localhost (see the first part in the log).
When I start the node with CLI and use parameter "hub" or use the parameter "hub" instead of "grid-url" the node connects correctly to the grid hub instance.
### How can we reproduce the issue?
```shell
(The TOML file is validated)
This was assumed to work fine, but does not:
[node]
grid-url = "http://mygridserver.com"
max-sessions = 4
I coincidentally found out, that using the works:
[node]
#the parameter actually should be "grid-url" but it works with the CLI parameter "hub"
hub = "http://mygridserver.com"
max-sessions = 4
I start the node: java -jar selenium-server-4.1.0.jar node --config node-config.toml
```
### Relevant log output
```shell
When using "grid-url" as parameter in TOML config file:
C:\Testen\SeleniumGrid>java -jar selenium-server-4.1.0.jar node --config node-config.toml
18:11:13.869 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding
18:11:13.872 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing
18:11:13.987 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://0.0.0.0:4442 and tcp://0.0.0.0:4443
18:11:14.017 INFO [UnboundZmqEventBus.<init>] - Sockets created
18:11:15.029 INFO [UnboundZmqEventBus.<init>] - Event bus ready
18:11:15.314 INFO [NodeServer.createHandlers] - Reporting self as: http://172.20.80.1:5555
18:11:15.541 INFO [NodeOptions.getSessionFactories] - Detected 20 available processors
18:11:15.560 INFO [NodeOptions.discoverDrivers] - Discovered 0 driver(s)
18:11:15.589 INFO [NodeOptions.report] - Adding Firefox for {"browserVersion": "95","browserName": "firefox","moz:firefoxOptions": {"binary": "C:\u002fTesten\u002fFirefox\u002f"},"platformName": "WIN10"} 8 times
18:11:15.590 INFO [NodeOptions.report] - Adding Chrome for {"goog:chromeOptions": {"binary": "C:\u002fProgram Files\u002fGoogle\u002fChrome\u002fApplication\u002f"},"browserVersion": "96","browserName": "chrome","se:webDriverExecutable": "C:\\Testen\\Driver\\chromedriver.exe","platformName": "WIN10"} 8 times
18:11:15.602 INFO [Node.<init>] - Binding additional locator mechanisms: id, relative, name
18:11:15.973 INFO [NodeServer$1.start] - Starting registration process for node id 67a37301-14c8-449c-b2d7-0206c3084152
18:11:15.974 INFO [NodeServer.execute] - Started Selenium node 4.1.0 (revision 87802e897b): http://172.20.80.1:5555
When using "hub" as parameter in TOML config file:
C:\Testen\SeleniumGrid>java -jar selenium-server-4.1.0.jar node --config node-config.toml
18:12:21.653 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding
18:12:21.656 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing
18:12:21.765 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://mygridserver.com:4442 and tcp://mygridserver.com:4443
18:12:21.800 INFO [UnboundZmqEventBus.<init>] - Sockets created
18:12:22.806 INFO [UnboundZmqEventBus.<init>] - Event bus ready
18:12:23.077 INFO [NodeServer.createHandlers] - Reporting self as: http://172.20.80.1:5555
18:12:23.302 INFO [NodeOptions.getSessionFactories] - Detected 20 available processors
18:12:23.319 INFO [NodeOptions.discoverDrivers] - Discovered 0 driver(s)
18:12:23.346 INFO [NodeOptions.report] - Adding Firefox for {"browserVersion": "95","browserName": "firefox","moz:firefoxOptions": {"binary": "C:\u002fTesten\u002fFirefox\u002f"},"platformName": "WIN10"} 8 times
18:12:23.347 INFO [NodeOptions.report] - Adding Chrome for {"goog:chromeOptions": {"binary": "C:\u002fProgram Files\u002fGoogle\u002fChrome\u002fApplication\u002f"},"browserVersion": "96","browserName": "chrome","se:webDriverExecutable": "C:\\Testen\\Driver\\chromedriver.exe","platformName": "WIN10"} 8 times
18:12:23.359 INFO [Node.<init>] - Binding additional locator mechanisms: name, relative, id
18:12:23.734 INFO [NodeServer$1.start] - Starting registration process for node id 3179bdd2-a1c9-448d-895a-ab281abe9e9d
18:12:23.735 INFO [NodeServer.execute] - Started Selenium node 4.1.0 (revision 87802e897b): http://172.20.80.1:5555
```
### Operating System
Windows 10
### Selenium version
.Net - Selenium 4.1.0
### What are the browser(s) and version(s) where you see this issue?
Chrome 96, Firefox 95
### What are the browser driver(s) and version(s) where you see this issue?
Chromdriver 96.0.4664.45, Geckodriver 0.30.0
### Are you using Selenium Grid?
4.1.0
|
1.0
|
[🐛 Bug]: Grid-Url in TOML node-config file is not passed through - ### What happened?
In Selenium GRID 3.x I configured the hub and node with the Json config files.
Now I want to configure Selenium GRID 4.x with TOML configuration file for the nodes and need to setup parameters (grid-url, max connections) as specified.
Either the documentation seems to be misleading or the parameter is not correctly passed through:
Section "Node", parameter "--grid-url"
https://www.selenium.dev/documentation/grid/configuration/cli_options/
Each node must connect to the grid hub instance, which is located on a different machine within my network.
When I start the node instance with the config-file it does not connect to the given URL and it seems to try to connect to my localhost (see the first part in the log).
When I start the node with CLI and use parameter "hub" or use the parameter "hub" instead of "grid-url" the node connects correctly to the grid hub instance.
### How can we reproduce the issue?
```shell
(The TOML file is validated)
This was assumed to work fine, but does not:
[node]
grid-url = "http://mygridserver.com"
max-sessions = 4
I coincidentally found out, that using the works:
[node]
#the parameter actually should be "grid-url" but it works with the CLI parameter "hub"
hub = "http://mygridserver.com"
max-sessions = 4
I start the node: java -jar selenium-server-4.1.0.jar node --config node-config.toml
```
### Relevant log output
```shell
When using "grid-url" as parameter in TOML config file:
C:\Testen\SeleniumGrid>java -jar selenium-server-4.1.0.jar node --config node-config.toml
18:11:13.869 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding
18:11:13.872 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing
18:11:13.987 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://0.0.0.0:4442 and tcp://0.0.0.0:4443
18:11:14.017 INFO [UnboundZmqEventBus.<init>] - Sockets created
18:11:15.029 INFO [UnboundZmqEventBus.<init>] - Event bus ready
18:11:15.314 INFO [NodeServer.createHandlers] - Reporting self as: http://172.20.80.1:5555
18:11:15.541 INFO [NodeOptions.getSessionFactories] - Detected 20 available processors
18:11:15.560 INFO [NodeOptions.discoverDrivers] - Discovered 0 driver(s)
18:11:15.589 INFO [NodeOptions.report] - Adding Firefox for {"browserVersion": "95","browserName": "firefox","moz:firefoxOptions": {"binary": "C:\u002fTesten\u002fFirefox\u002f"},"platformName": "WIN10"} 8 times
18:11:15.590 INFO [NodeOptions.report] - Adding Chrome for {"goog:chromeOptions": {"binary": "C:\u002fProgram Files\u002fGoogle\u002fChrome\u002fApplication\u002f"},"browserVersion": "96","browserName": "chrome","se:webDriverExecutable": "C:\\Testen\\Driver\\chromedriver.exe","platformName": "WIN10"} 8 times
18:11:15.602 INFO [Node.<init>] - Binding additional locator mechanisms: id, relative, name
18:11:15.973 INFO [NodeServer$1.start] - Starting registration process for node id 67a37301-14c8-449c-b2d7-0206c3084152
18:11:15.974 INFO [NodeServer.execute] - Started Selenium node 4.1.0 (revision 87802e897b): http://172.20.80.1:5555
When using "hub" as parameter in TOML config file:
C:\Testen\SeleniumGrid>java -jar selenium-server-4.1.0.jar node --config node-config.toml
18:12:21.653 INFO [LoggingOptions.configureLogEncoding] - Using the system default encoding
18:12:21.656 INFO [OpenTelemetryTracer.createTracer] - Using OpenTelemetry for tracing
18:12:21.765 INFO [UnboundZmqEventBus.<init>] - Connecting to tcp://mygridserver.com:4442 and tcp://mygridserver.com:4443
18:12:21.800 INFO [UnboundZmqEventBus.<init>] - Sockets created
18:12:22.806 INFO [UnboundZmqEventBus.<init>] - Event bus ready
18:12:23.077 INFO [NodeServer.createHandlers] - Reporting self as: http://172.20.80.1:5555
18:12:23.302 INFO [NodeOptions.getSessionFactories] - Detected 20 available processors
18:12:23.319 INFO [NodeOptions.discoverDrivers] - Discovered 0 driver(s)
18:12:23.346 INFO [NodeOptions.report] - Adding Firefox for {"browserVersion": "95","browserName": "firefox","moz:firefoxOptions": {"binary": "C:\u002fTesten\u002fFirefox\u002f"},"platformName": "WIN10"} 8 times
18:12:23.347 INFO [NodeOptions.report] - Adding Chrome for {"goog:chromeOptions": {"binary": "C:\u002fProgram Files\u002fGoogle\u002fChrome\u002fApplication\u002f"},"browserVersion": "96","browserName": "chrome","se:webDriverExecutable": "C:\\Testen\\Driver\\chromedriver.exe","platformName": "WIN10"} 8 times
18:12:23.359 INFO [Node.<init>] - Binding additional locator mechanisms: name, relative, id
18:12:23.734 INFO [NodeServer$1.start] - Starting registration process for node id 3179bdd2-a1c9-448d-895a-ab281abe9e9d
18:12:23.735 INFO [NodeServer.execute] - Started Selenium node 4.1.0 (revision 87802e897b): http://172.20.80.1:5555
```
### Operating System
Windows 10
### Selenium version
.Net - Selenium 4.1.0
### What are the browser(s) and version(s) where you see this issue?
Chrome 96, Firefox 95
### What are the browser driver(s) and version(s) where you see this issue?
Chromdriver 96.0.4664.45, Geckodriver 0.30.0
### Are you using Selenium Grid?
4.1.0
|
defect
|
grid url in toml node config file is not passed through what happened in selenium grid x i configured the hub and node with the json config files now i want to configure selenium grid x with toml configuration file for the nodes and need to setup parameters grid url max connections as specified either the documentation seems to be misleading or the parameter is not correctly passed through section node parameter grid url each node must connect to the grid hub instance which is located on a different machine within my network when i start the node instance with the config file it does not connect to the given url and it seems to try to connect to my localhost see the first part in the log when i start the node with cli and use parameter hub or use the parameter hub instead of grid url the node connects correctly to the grid hub instance how can we reproduce the issue shell the toml file is validated this was assumed to work fine but does not grid url max sessions i coincidentally found out that using the works the parameter actually should be grid url but it works with the cli parameter hub hub max sessions i start the node java jar selenium server jar node config node config toml relevant log output shell when using grid url as parameter in toml config file c testen seleniumgrid java jar selenium server jar node config node config toml info using the system default encoding info using opentelemetry for tracing info connecting to tcp and tcp info sockets created info event bus ready info reporting self as info detected available processors info discovered driver s info adding firefox for browserversion browsername firefox moz firefoxoptions binary c platformname times info adding chrome for goog chromeoptions binary c files browserversion browsername chrome se webdriverexecutable c testen driver chromedriver exe platformname times info binding additional locator mechanisms id relative name info starting registration process for node id info started selenium node revision when using hub as parameter in toml config file c testen seleniumgrid java jar selenium server jar node config node config toml info using the system default encoding info using opentelemetry for tracing info connecting to tcp mygridserver com and tcp mygridserver com info sockets created info event bus ready info reporting self as info detected available processors info discovered driver s info adding firefox for browserversion browsername firefox moz firefoxoptions binary c platformname times info adding chrome for goog chromeoptions binary c files browserversion browsername chrome se webdriverexecutable c testen driver chromedriver exe platformname times info binding additional locator mechanisms name relative id info starting registration process for node id info started selenium node revision operating system windows selenium version net selenium what are the browser s and version s where you see this issue chrome firefox what are the browser driver s and version s where you see this issue chromdriver geckodriver are you using selenium grid
| 1
|
28,425
| 4,104,697,003
|
IssuesEvent
|
2016-06-05 15:07:13
|
JuliaLang/julia
|
https://api.github.com/repos/JuliaLang/julia
|
opened
|
2-arg A_mul_B! should be consistent, or not exist
|
design linear algebra
|
Example methods are here: https://github.com/JuliaLang/julia/blob/643bdbec6aeca8053e33a242df86e2de2327f815/base/linalg/triangular.jl#L402-L403 where these clearly follow the BLAS `trmm` convention of mutating whichever argument is full. So `A_mul_B!(A, B)` might mutate A, or it might mutate B, depending on their types. This is the case for several other combinations of input types as well, and I think it's bad for generic programming. Ref discussion at https://github.com/JuliaLang/julia/pull/16615#discussion_r65652965
If we can't apply a consistent rule here, "always mutates the second argument" (or the first?), then we shouldn't define this method. Better to favor the more explicit `A_mul_B!(C, A, B)` method where the output is a separate argument. Some combinations of types will allow one of `A_mul_B!(A, A, B)` or `A_mul_B!(B, A, B)` to work, but writing code that does that will be inherently non-generic. The individual methods of `A_mul_B!` should be responsible for checking whether the inputs are `===` (and possibly more complicated forms of alias checks if we have more view types in common use) and only allowing the specific combinations that are expected to work.
This issue is still present for all the in-place multiplication methods even if #6837 renames all 7 (or 9 if you count possible combinations we don't define right now, ref #5332) variants to methods of `mul!`. I think this is an explicitly post-0.5 cleanup to make though since we don't have time to deal with it right now.
|
1.0
|
2-arg A_mul_B! should be consistent, or not exist - Example methods are here: https://github.com/JuliaLang/julia/blob/643bdbec6aeca8053e33a242df86e2de2327f815/base/linalg/triangular.jl#L402-L403 where these clearly follow the BLAS `trmm` convention of mutating whichever argument is full. So `A_mul_B!(A, B)` might mutate A, or it might mutate B, depending on their types. This is the case for several other combinations of input types as well, and I think it's bad for generic programming. Ref discussion at https://github.com/JuliaLang/julia/pull/16615#discussion_r65652965
If we can't apply a consistent rule here, "always mutates the second argument" (or the first?), then we shouldn't define this method. Better to favor the more explicit `A_mul_B!(C, A, B)` method where the output is a separate argument. Some combinations of types will allow one of `A_mul_B!(A, A, B)` or `A_mul_B!(B, A, B)` to work, but writing code that does that will be inherently non-generic. The individual methods of `A_mul_B!` should be responsible for checking whether the inputs are `===` (and possibly more complicated forms of alias checks if we have more view types in common use) and only allowing the specific combinations that are expected to work.
This issue is still present for all the in-place multiplication methods even if #6837 renames all 7 (or 9 if you count possible combinations we don't define right now, ref #5332) variants to methods of `mul!`. I think this is an explicitly post-0.5 cleanup to make though since we don't have time to deal with it right now.
|
non_defect
|
arg a mul b should be consistent or not exist example methods are here where these clearly follow the blas trmm convention of mutating whichever argument is full so a mul b a b might mutate a or it might mutate b depending on their types this is the case for several other combinations of input types as well and i think it s bad for generic programming ref discussion at if we can t apply a consistent rule here always mutates the second argument or the first then we shouldn t define this method better to favor the more explicit a mul b c a b method where the output is a separate argument some combinations of types will allow one of a mul b a a b or a mul b b a b to work but writing code that does that will be inherently non generic the individual methods of a mul b should be responsible for checking whether the inputs are and possibly more complicated forms of alias checks if we have more view types in common use and only allowing the specific combinations that are expected to work this issue is still present for all the in place multiplication methods even if renames all or if you count possible combinations we don t define right now ref variants to methods of mul i think this is an explicitly post cleanup to make though since we don t have time to deal with it right now
| 0
|
752,268
| 26,278,362,620
|
IssuesEvent
|
2023-01-07 02:41:26
|
yugabyte/yugabyte-db
|
https://api.github.com/repos/yugabyte/yugabyte-db
|
closed
|
[DocDB] [xCluster][Atomicity and Ordering] Data inconsistency after replication pause
|
kind/bug area/docdb priority/high xCluster
|
Jira Link: [DB-4940](https://yugabyte.atlassian.net/browse/DB-4940)
### Description
**Steps**:
1. Create 2 universes(Source and Target) with Turn ON the GFlag **enable_replicate_transaction_status_table** on master of both universes and disable automatic tablet splitting
2. Create database and 3 tables at both sides
3. Setup replication from source to target
4. Change the xCluster role at Target
`yb-admin -master_addresses <master-address> change_xcluster_role STANDBY`
5. Updates(within transaction) are happening in one table
6. Inserts(within transactions) are happening in other 2 tables
7. After sometime pause the replication
**Observations**,
1. After pausing the replication observed Data inconsistency,
> Update transaction partially replicated at target.
> Insert transaction partially replicated at target.
8. After resuming the replication(after 2 min)
> Partially un-replicated data(which was within transactions) is replicated for all 3 tables
> And data loss (next ran transactions’ data is not replicated)
9. Observed below logs in source’s tserver log file
`I0106 07:55:38.122043 24527 tablet_rpc.cc:458] Operation failed. Try again (yb/tserver/tablet_service.cc:1070): Failed UpdateTransaction: tablet_id: "aae7e35e001d4fc6ab1def3bc3d06f0b" state { transaction_id: "\301\341\326\010+\354D\022\237\210\\\3439\236\331\017" status: APPLYING tablets: "de5564bbee144b7fa635f06cf5fb2b2d" commit_hybrid_time: 6852572107115016192 sealed: false aborted { } external_hybrid_time: 6852572107115016192 } propagated_hybrid_time: 6852574159347347456 is_external: true, retrier: { task_id: -1 state: kIdle deadline: 6806.796s } to tablet aae7e35e001d4fc6ab1def3bc3d06f0b on tablet server { uuid: 725e7ad0b47649be9ce0c6491e03098b private: [host: "172.161.59.194" port: 9100] cloud_info: placement_cloud: "aws" placement_region: "us-east-2" placement_zone: "us-east-2c" after 1 attempt(s): XCluster safe time not found for namespace 00004005000030008000000000000000`
[DB-4939]: https://yugabyte.atlassian.net/browse/DB-4939?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
[DB-4940]: https://yugabyte.atlassian.net/browse/DB-4940?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
|
1.0
|
[DocDB] [xCluster][Atomicity and Ordering] Data inconsistency after replication pause - Jira Link: [DB-4940](https://yugabyte.atlassian.net/browse/DB-4940)
### Description
**Steps**:
1. Create 2 universes(Source and Target) with Turn ON the GFlag **enable_replicate_transaction_status_table** on master of both universes and disable automatic tablet splitting
2. Create database and 3 tables at both sides
3. Setup replication from source to target
4. Change the xCluster role at Target
`yb-admin -master_addresses <master-address> change_xcluster_role STANDBY`
5. Updates(within transaction) are happening in one table
6. Inserts(within transactions) are happening in other 2 tables
7. After sometime pause the replication
**Observations**,
1. After pausing the replication observed Data inconsistency,
> Update transaction partially replicated at target.
> Insert transaction partially replicated at target.
8. After resuming the replication(after 2 min)
> Partially un-replicated data(which was within transactions) is replicated for all 3 tables
> And data loss (next ran transactions’ data is not replicated)
9. Observed below logs in source’s tserver log file
`I0106 07:55:38.122043 24527 tablet_rpc.cc:458] Operation failed. Try again (yb/tserver/tablet_service.cc:1070): Failed UpdateTransaction: tablet_id: "aae7e35e001d4fc6ab1def3bc3d06f0b" state { transaction_id: "\301\341\326\010+\354D\022\237\210\\\3439\236\331\017" status: APPLYING tablets: "de5564bbee144b7fa635f06cf5fb2b2d" commit_hybrid_time: 6852572107115016192 sealed: false aborted { } external_hybrid_time: 6852572107115016192 } propagated_hybrid_time: 6852574159347347456 is_external: true, retrier: { task_id: -1 state: kIdle deadline: 6806.796s } to tablet aae7e35e001d4fc6ab1def3bc3d06f0b on tablet server { uuid: 725e7ad0b47649be9ce0c6491e03098b private: [host: "172.161.59.194" port: 9100] cloud_info: placement_cloud: "aws" placement_region: "us-east-2" placement_zone: "us-east-2c" after 1 attempt(s): XCluster safe time not found for namespace 00004005000030008000000000000000`
[DB-4939]: https://yugabyte.atlassian.net/browse/DB-4939?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
[DB-4940]: https://yugabyte.atlassian.net/browse/DB-4940?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
|
non_defect
|
data inconsistency after replication pause jira link description steps create universes source and target with turn on the gflag enable replicate transaction status table on master of both universes and disable automatic tablet splitting create database and tables at both sides setup replication from source to target change the xcluster role at target yb admin master addresses change xcluster role standby updates within transaction are happening in one table inserts within transactions are happening in other tables after sometime pause the replication observations after pausing the replication observed data inconsistency update transaction partially replicated at target insert transaction partially replicated at target after resuming the replication after min partially un replicated data which was within transactions is replicated for all tables and data loss next ran transactions’ data is not replicated observed below logs in source’s tserver log file tablet rpc cc operation failed try again yb tserver tablet service cc failed updatetransaction tablet id state transaction id status applying tablets commit hybrid time sealed false aborted external hybrid time propagated hybrid time is external true retrier task id state kidle deadline to tablet on tablet server uuid private cloud info placement cloud aws placement region us east placement zone us east after attempt s xcluster safe time not found for namespace
| 0
|
9,925
| 2,616,011,008
|
IssuesEvent
|
2015-03-02 00:54:14
|
jasonhall/bwapi
|
https://api.github.com/repos/jasonhall/bwapi
|
closed
|
Details on a geyser are incorrect after a refinery on it has been destroyed
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Play a game and have the enemy steal your gas
2. After destroying it, its details either don't change or are incorrect
Unsure whether it affects all races / refineries but it was experienced with
the bot playing Protoss and a
Terran enemy building a refinery in the bots base.
What is the expected output? What do you see instead?
BWAPI::Broodwar->canBuildHere(NULL, geyser->getTilePosition(),
BWAPI::Broodwar->self()-
>getRace().getRefinery()) returns false on the empty geyser afterwards
What version of the product are you using? On what operating system?
r2255 on Windows 7 x64
```
Original issue reported on code.google.com by `quaver.smith` on 30 May 2010 at 2:07
|
1.0
|
Details on a geyser are incorrect after a refinery on it has been destroyed - ```
What steps will reproduce the problem?
1. Play a game and have the enemy steal your gas
2. After destroying it, its details either don't change or are incorrect
Unsure whether it affects all races / refineries but it was experienced with
the bot playing Protoss and a
Terran enemy building a refinery in the bots base.
What is the expected output? What do you see instead?
BWAPI::Broodwar->canBuildHere(NULL, geyser->getTilePosition(),
BWAPI::Broodwar->self()-
>getRace().getRefinery()) returns false on the empty geyser afterwards
What version of the product are you using? On what operating system?
r2255 on Windows 7 x64
```
Original issue reported on code.google.com by `quaver.smith` on 30 May 2010 at 2:07
|
defect
|
details on a geyser are incorrect after a refinery on it has been destroyed what steps will reproduce the problem play a game and have the enemy steal your gas after destroying it its details either don t change or are incorrect unsure whether it affects all races refineries but it was experienced with the bot playing protoss and a terran enemy building a refinery in the bots base what is the expected output what do you see instead bwapi broodwar canbuildhere null geyser gettileposition bwapi broodwar self getrace getrefinery returns false on the empty geyser afterwards what version of the product are you using on what operating system on windows original issue reported on code google com by quaver smith on may at
| 1
|
33,672
| 7,195,406,479
|
IssuesEvent
|
2018-02-04 16:50:46
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
Cake 2 - getAffectedRows can return 0 when query is ran via Model->query()
|
Defect models
|
This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 2.10.6
### What you did
I have a shell that generates some SQL (a MySQL INSERT IGNORE INTO query) that runs via:
```PHP
$rows = 0;
$sql = $this->obtainSQL($activityType, $lastRun); #returns valid SQL
$result = $this->Model->query($sql);
if($result) {
$rows = $this->Model->getAffectedRows();
}
```
(Actual code, Model is a different model, but `query` and `getAffectedRows` are not overridden anywhere)
### What happened
`$rows` is always 0
### What you expected to happen
`$rows` to be non-zero when the INSERT IGNORE query inserts records.
So what I suspect is happening is that when you call getAffectedRows, you end up calling [DboSource::hasResult](https://github.com/cakephp/cakephp/blob/2.x/lib/Cake/Model/Datasource/DboSource.php#L946-L949)
The issue is, that [DboSource::_execute can sometimes move the result into `$this->_results`](https://github.com/cakephp/cakephp/blob/2.x/lib/Cake/Model/Datasource/DboSource.php#L473), rather than `$this->_result` (note the extra `s`).
This means that when you call $this->Model->getAffectedRows();, it thinks that DboSource::hasResult is false, and just returns 0 (Ref [DboSource::lastAffected](https://github.com/cakephp/cakephp/blob/2.x/lib/Cake/Model/Datasource/DboSource.php#L519-L524))
There's only one `$this->_results` in all of the DboSource.php file, and there's no tests at all on `getAffectedRows`, `hasResult` and the tests for `lastAffected` don't try a query run via `Model->query()`
|
1.0
|
Cake 2 - getAffectedRows can return 0 when query is ran via Model->query() - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 2.10.6
### What you did
I have a shell that generates some SQL (a MySQL INSERT IGNORE INTO query) that runs via:
```PHP
$rows = 0;
$sql = $this->obtainSQL($activityType, $lastRun); #returns valid SQL
$result = $this->Model->query($sql);
if($result) {
$rows = $this->Model->getAffectedRows();
}
```
(Actual code, Model is a different model, but `query` and `getAffectedRows` are not overridden anywhere)
### What happened
`$rows` is always 0
### What you expected to happen
`$rows` to be non-zero when the INSERT IGNORE query inserts records.
So what I suspect is happening is that when you call getAffectedRows, you end up calling [DboSource::hasResult](https://github.com/cakephp/cakephp/blob/2.x/lib/Cake/Model/Datasource/DboSource.php#L946-L949)
The issue is, that [DboSource::_execute can sometimes move the result into `$this->_results`](https://github.com/cakephp/cakephp/blob/2.x/lib/Cake/Model/Datasource/DboSource.php#L473), rather than `$this->_result` (note the extra `s`).
This means that when you call $this->Model->getAffectedRows();, it thinks that DboSource::hasResult is false, and just returns 0 (Ref [DboSource::lastAffected](https://github.com/cakephp/cakephp/blob/2.x/lib/Cake/Model/Datasource/DboSource.php#L519-L524))
There's only one `$this->_results` in all of the DboSource.php file, and there's no tests at all on `getAffectedRows`, `hasResult` and the tests for `lastAffected` don't try a query run via `Model->query()`
|
defect
|
cake getaffectedrows can return when query is ran via model query this is a multiple allowed bug enhancement feature discussion rfc cakephp version what you did i have a shell that generates some sql a mysql insert ignore into query that runs via php rows sql this obtainsql activitytype lastrun returns valid sql result this model query sql if result rows this model getaffectedrows actual code model is a different model but query and getaffectedrows are not overridden anywhere what happened rows is always what you expected to happen rows to be non zero when the insert ignore query inserts records so what i suspect is happening is that when you call getaffectedrows you end up calling the issue is that rather than this result note the extra s this means that when you call this model getaffectedrows it thinks that dbosource hasresult is false and just returns ref there s only one this results in all of the dbosource php file and there s no tests at all on getaffectedrows hasresult and the tests for lastaffected don t try a query run via model query
| 1
|
38,995
| 9,121,777,620
|
IssuesEvent
|
2019-02-23 01:21:31
|
techo/voluntariado-eventual
|
https://api.github.com/repos/techo/voluntariado-eventual
|
closed
|
[Error] Algunos usuarios inscriptos en Argentina aparecen con país México.
|
Defecto
|
**Describí el error**
Algunos usuarios inscriptos en Argentina aparecen con país México. Esto pasa cuando los usuarios llegan desde la integración con Pilote.
**Para reproducirlo**
Pasos para reproducir el comportamiento:
1. Crear un usuario nuevo en Pilote con país "Argentina"
2. Esperar unos minutos
3. Buscar el usuario en la plataforma nueva
4. Ver que el país se asignó como "México"
**Comportamiento esperando**
El usuario debería aparecer con país "Argentina"
**Contexto adicional**
Reportado por Cami Minor.
|
1.0
|
[Error] Algunos usuarios inscriptos en Argentina aparecen con país México. - **Describí el error**
Algunos usuarios inscriptos en Argentina aparecen con país México. Esto pasa cuando los usuarios llegan desde la integración con Pilote.
**Para reproducirlo**
Pasos para reproducir el comportamiento:
1. Crear un usuario nuevo en Pilote con país "Argentina"
2. Esperar unos minutos
3. Buscar el usuario en la plataforma nueva
4. Ver que el país se asignó como "México"
**Comportamiento esperando**
El usuario debería aparecer con país "Argentina"
**Contexto adicional**
Reportado por Cami Minor.
|
defect
|
algunos usuarios inscriptos en argentina aparecen con país méxico describí el error algunos usuarios inscriptos en argentina aparecen con país méxico esto pasa cuando los usuarios llegan desde la integración con pilote para reproducirlo pasos para reproducir el comportamiento crear un usuario nuevo en pilote con país argentina esperar unos minutos buscar el usuario en la plataforma nueva ver que el país se asignó como méxico comportamiento esperando el usuario debería aparecer con país argentina contexto adicional reportado por cami minor
| 1
|
39,530
| 9,530,760,459
|
IssuesEvent
|
2019-04-29 14:34:02
|
primefaces/primereact
|
https://api.github.com/repos/primefaces/primereact
|
closed
|
Calendar used in DataTable is unable to switch months
|
defect
|
**I'm submitting a ...
[ *] bug report
[ ] feature request
[ ] support request
You can find a working exaple of the problem here
https://codesandbox.io/s/6wv3qkyl4n
**Current behavior**
I can not change months when Calendar is used in a DataTable (but can when used normal)
**Expected behavior**
I should be able to change the month when Calendar is used in a DataTable
**Minimal reproduction of the problem with instructions**
Workin example here:
https://codesandbox.io/s/6wv3qkyl4n
**Please tell us about your environment:**
MAC OS
* **React version:**
react": "^16.8.6"
* **PrimeReact version:**
"primereact": "^3.1.2"
* **Browser:**
I tested Firefox Chrome
* **Language:** [all | TypeScript X.X | ES6/7 | ES5]
|
1.0
|
Calendar used in DataTable is unable to switch months - **I'm submitting a ...
[ *] bug report
[ ] feature request
[ ] support request
You can find a working exaple of the problem here
https://codesandbox.io/s/6wv3qkyl4n
**Current behavior**
I can not change months when Calendar is used in a DataTable (but can when used normal)
**Expected behavior**
I should be able to change the month when Calendar is used in a DataTable
**Minimal reproduction of the problem with instructions**
Workin example here:
https://codesandbox.io/s/6wv3qkyl4n
**Please tell us about your environment:**
MAC OS
* **React version:**
react": "^16.8.6"
* **PrimeReact version:**
"primereact": "^3.1.2"
* **Browser:**
I tested Firefox Chrome
* **Language:** [all | TypeScript X.X | ES6/7 | ES5]
|
defect
|
calendar used in datatable is unable to switch months i m submitting a bug report feature request support request you can find a working exaple of the problem here current behavior i can not change months when calendar is used in a datatable but can when used normal expected behavior i should be able to change the month when calendar is used in a datatable minimal reproduction of the problem with instructions workin example here please tell us about your environment mac os react version react primereact version primereact browser i tested firefox chrome language
| 1
|
172,237
| 13,283,442,439
|
IssuesEvent
|
2020-08-24 03:14:05
|
OpenPaaS-Suite/esn-frontend-inbox
|
https://api.github.com/repos/OpenPaaS-Suite/esn-frontend-inbox
|
reopened
|
Unable to move to email
|
QA:Testing bug
|
Reproduce steps:
- Select to go to email's detail in the INBOX
- Click on "MOVE TO" icon
- Verify issue
Actual Results: Unable to move to email and it displays the error in the Console tab.

Expected Results: Please help to investigate this and fix to able to move to email
Tested on Go to https://dev.open-paas.org/ and using Chrome/Firefox on Ubuntu
|
1.0
|
Unable to move to email - Reproduce steps:
- Select to go to email's detail in the INBOX
- Click on "MOVE TO" icon
- Verify issue
Actual Results: Unable to move to email and it displays the error in the Console tab.

Expected Results: Please help to investigate this and fix to able to move to email
Tested on Go to https://dev.open-paas.org/ and using Chrome/Firefox on Ubuntu
|
non_defect
|
unable to move to email reproduce steps select to go to email s detail in the inbox click on move to icon verify issue actual results unable to move to email and it displays the error in the console tab expected results please help to investigate this and fix to able to move to email tested on go to and using chrome firefox on ubuntu
| 0
|
61,893
| 17,023,802,359
|
IssuesEvent
|
2021-07-03 03:56:14
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
mod_tile compilation error (problem with image_32 buf on line 445 of gen_tile.cpp)
|
Component: mod_tile Priority: blocker Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 9.58am, Thursday, 7th June 2012]**
Hello,
I've encoutered some problems compiling mod_tile, with the command "make".[[BR]]
Indeed, when the compiler is focusing on gen_tile.cpp, the console gives me those errors :[[BR]]
```
g++ -DHAVE_CONFIG_H -I. -I/usr/include/freetype2 -pthread -I/usr/include -D_REENTRANT -I/usr/local/include -g -O2 -MT gen_tile.o -MD -MP -MF .deps/gen_tile.Tpo -c -o gen_tile.o gen_tile.cpp
In file included from gen_tile.cpp:5:
/usr/local/include/mapnik/filter_factory.hpp:1:2: warning: #warning "filter_factory.hpp" is now called "expression.hpp"
gen_tile.cpp: In function protoCmd render(mapnik::Map&, char*, mapnik::projection&, int, int, int, unsigned int, metaTile&):
gen_tile.cpp:445: error: variable mapnik::image_32 buf has initializer but incomplete type
gen_tile.cpp:453: error: image_view was not declared in this scope
gen_tile.cpp:453: error: expected primary-expression before > token
gen_tile.cpp:453: error: vw was not declared in this scope
make[2]: *** [gen_tile.o] Error 1
make[2]: Leaving directory `/home/direst/osm/mod_tile'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/home/direst/osm/mod_tile'
make: *** [all] Error 2
```
Apparently, I can't instantiante the value buf and can't understand why.[[BR]]
My OS is Debian 6 Squeeze and I've installed Mapnik properly, according to this following site :[[BR]]
[https://github.com/mapnik/mapnik/wiki/Mapnik2
][[BR]]
The libboost version I have is 1.49 and I've also the latest ICU version.[[BR]]
In order to install mod_tile, I've performed those operations :[[BR]]
```
cd ~/osm/
svn co http://svn.openstreetmap.org/applications/utils/mod_tile
cd mod_tile
./autogen.sh
./configure
make
sudo make install
sudo make install-mod_tile
```
[[BR]]
But the "make" and "sudo make install" commands are not the working. "sudo make install-mod_tile" is working actually.[[BR]]
If I can't compile, I will not be able to perform this commande : "
./renderd -f"[[BR]]
So if anyone has a solution...[[BR]]
Thanks
|
1.0
|
mod_tile compilation error (problem with image_32 buf on line 445 of gen_tile.cpp) - **[Submitted to the original trac issue database at 9.58am, Thursday, 7th June 2012]**
Hello,
I've encoutered some problems compiling mod_tile, with the command "make".[[BR]]
Indeed, when the compiler is focusing on gen_tile.cpp, the console gives me those errors :[[BR]]
```
g++ -DHAVE_CONFIG_H -I. -I/usr/include/freetype2 -pthread -I/usr/include -D_REENTRANT -I/usr/local/include -g -O2 -MT gen_tile.o -MD -MP -MF .deps/gen_tile.Tpo -c -o gen_tile.o gen_tile.cpp
In file included from gen_tile.cpp:5:
/usr/local/include/mapnik/filter_factory.hpp:1:2: warning: #warning "filter_factory.hpp" is now called "expression.hpp"
gen_tile.cpp: In function protoCmd render(mapnik::Map&, char*, mapnik::projection&, int, int, int, unsigned int, metaTile&):
gen_tile.cpp:445: error: variable mapnik::image_32 buf has initializer but incomplete type
gen_tile.cpp:453: error: image_view was not declared in this scope
gen_tile.cpp:453: error: expected primary-expression before > token
gen_tile.cpp:453: error: vw was not declared in this scope
make[2]: *** [gen_tile.o] Error 1
make[2]: Leaving directory `/home/direst/osm/mod_tile'
make[1]: *** [all-recursive] Error 1
make[1]: Leaving directory `/home/direst/osm/mod_tile'
make: *** [all] Error 2
```
Apparently, I can't instantiante the value buf and can't understand why.[[BR]]
My OS is Debian 6 Squeeze and I've installed Mapnik properly, according to this following site :[[BR]]
[https://github.com/mapnik/mapnik/wiki/Mapnik2
][[BR]]
The libboost version I have is 1.49 and I've also the latest ICU version.[[BR]]
In order to install mod_tile, I've performed those operations :[[BR]]
```
cd ~/osm/
svn co http://svn.openstreetmap.org/applications/utils/mod_tile
cd mod_tile
./autogen.sh
./configure
make
sudo make install
sudo make install-mod_tile
```
[[BR]]
But the "make" and "sudo make install" commands are not the working. "sudo make install-mod_tile" is working actually.[[BR]]
If I can't compile, I will not be able to perform this commande : "
./renderd -f"[[BR]]
So if anyone has a solution...[[BR]]
Thanks
|
defect
|
mod tile compilation error problem with image buf on line of gen tile cpp hello i ve encoutered some problems compiling mod tile with the command make indeed when the compiler is focusing on gen tile cpp the console gives me those errors g dhave config h i i usr include pthread i usr include d reentrant i usr local include g mt gen tile o md mp mf deps gen tile tpo c o gen tile o gen tile cpp in file included from gen tile cpp usr local include mapnik filter factory hpp warning warning filter factory hpp is now called expression hpp gen tile cpp in function protocmd render mapnik map char mapnik projection int int int unsigned int metatile gen tile cpp error variable mapnik image buf has initializer but incomplete type gen tile cpp error image view was not declared in this scope gen tile cpp error expected primary expression before token gen tile cpp error vw was not declared in this scope make error make leaving directory home direst osm mod tile make error make leaving directory home direst osm mod tile make error apparently i can t instantiante the value buf and can t understand why my os is debian squeeze and i ve installed mapnik properly according to this following site the libboost version i have is and i ve also the latest icu version in order to install mod tile i ve performed those operations cd osm svn co cd mod tile autogen sh configure make sudo make install sudo make install mod tile but the make and sudo make install commands are not the working sudo make install mod tile is working actually if i can t compile i will not be able to perform this commande renderd f so if anyone has a solution thanks
| 1
|
6,421
| 2,610,242,771
|
IssuesEvent
|
2015-02-26 19:17:15
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
台州割包皮包茎手术价钱
|
auto-migrated Priority-Medium Type-Defect
|
```
台州割包皮包茎手术价钱【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 31 May 2014 at 1:24
|
1.0
|
台州割包皮包茎手术价钱 - ```
台州割包皮包茎手术价钱【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 31 May 2014 at 1:24
|
defect
|
台州割包皮包茎手术价钱 台州割包皮包茎手术价钱【台州五洲生殖医院】 询热线 微信号tzwzszyy 医院地址 台州 (枫南大转盘旁)乘车线路 、 � �� 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 1
|
83,819
| 24,152,699,494
|
IssuesEvent
|
2022-09-22 03:35:42
|
habitat-sh/builder
|
https://api.github.com/repos/habitat-sh/builder
|
closed
|
RBAC & base packages
|
Focus:Builder Type:Feature Type:DevX Type:Front End Focus: UI Stale
|
Looking at some of the conversations in habislack recently, it looks like there may be a good opportunity to add some rbac around base packages to ensure they aren't promoted at the wrong time.
Idea:
- tag a package as a `base` plan, and be able to pull this tag into the Builder UI
- only a subset of people can promote a `base` plan (this is already the case for core plans maintainers, but perhaps we want to mirror the approver structure in github to have a more restricted list of promoters within the UI)
- when a person with permission to promote a `base` plan wants to do so, we pop up a confirmation dialog and require password re-entry (this could be a per-origin option) to ensure that this is being done with full awareness
We could also start to explore this for on-prem builder and using tagging to allow greater restrictions on who can promote what packages to which channels.
cc @ryankeairns for ideas :)
|
1.0
|
RBAC & base packages - Looking at some of the conversations in habislack recently, it looks like there may be a good opportunity to add some rbac around base packages to ensure they aren't promoted at the wrong time.
Idea:
- tag a package as a `base` plan, and be able to pull this tag into the Builder UI
- only a subset of people can promote a `base` plan (this is already the case for core plans maintainers, but perhaps we want to mirror the approver structure in github to have a more restricted list of promoters within the UI)
- when a person with permission to promote a `base` plan wants to do so, we pop up a confirmation dialog and require password re-entry (this could be a per-origin option) to ensure that this is being done with full awareness
We could also start to explore this for on-prem builder and using tagging to allow greater restrictions on who can promote what packages to which channels.
cc @ryankeairns for ideas :)
|
non_defect
|
rbac base packages looking at some of the conversations in habislack recently it looks like there may be a good opportunity to add some rbac around base packages to ensure they aren t promoted at the wrong time idea tag a package as a base plan and be able to pull this tag into the builder ui only a subset of people can promote a base plan this is already the case for core plans maintainers but perhaps we want to mirror the approver structure in github to have a more restricted list of promoters within the ui when a person with permission to promote a base plan wants to do so we pop up a confirmation dialog and require password re entry this could be a per origin option to ensure that this is being done with full awareness we could also start to explore this for on prem builder and using tagging to allow greater restrictions on who can promote what packages to which channels cc ryankeairns for ideas
| 0
|
28,055
| 8,069,118,507
|
IssuesEvent
|
2018-08-06 03:30:31
|
openshiftio/openshift.io
|
https://api.github.com/repos/openshiftio/openshift.io
|
opened
|
jenkins-idler: No need to unidle/idle `content-reposiory` once it is removed
|
team/build-cd type/task
|
There is an effort to remove content repository with these PRs
https://github.com/fabric8io/fabric8-pipeline-library/pull/409
https://github.com/fabric8-services/fabric8-tenant-jenkins/pull/108
Idler has code to idle and unidle content repository. This will have to be changed
|
1.0
|
jenkins-idler: No need to unidle/idle `content-reposiory` once it is removed - There is an effort to remove content repository with these PRs
https://github.com/fabric8io/fabric8-pipeline-library/pull/409
https://github.com/fabric8-services/fabric8-tenant-jenkins/pull/108
Idler has code to idle and unidle content repository. This will have to be changed
|
non_defect
|
jenkins idler no need to unidle idle content reposiory once it is removed there is an effort to remove content repository with these prs idler has code to idle and unidle content repository this will have to be changed
| 0
|
369,835
| 25,871,028,146
|
IssuesEvent
|
2022-12-14 02:39:47
|
ifm/ifm3d
|
https://api.github.com/repos/ifm/ifm3d
|
closed
|
Improve documentation and integrate documentation with readthedocs.org
|
documentation stale
|
Documentation for ifm3d could be significantly improved -- both the content and quality as well as the presentation. On the latter topic, this issue proposes integrating our documentation with the [Read The Docs](https://readthedocs.org) website. Putting the infrastructure in place to automatically build updated documentation upon release or commit to a particular branch would be ideal.
Also, let's start a list of what documentation needs to be made available:
- [x] Supported platforms
- [x] Building the software
- [x] Building and installing on Windows (should be treated separately from above)
- [x] How to implement a new image container
- [x] How to configure the imager using the JSON interface
- [ ] Related to above, all parameters need to be documented by the firmware/systems team
- [x] Using the `ifm3d` command-line tool
Please help me add to this list!
(NOTE: Is there a good way to make this check list shared so others can add to it? This is beyond my github skills).
We note that, per @graugans [comment](https://github.com/lovepark/ifm3d/issues/35#issuecomment-364398823) in #35 we will continue to provide code examples on using the library. The side benefit of making the examples a separate project also provides a nice way to illustrate how to integrate `ifm3d` into your project to include the build chain.
|
1.0
|
Improve documentation and integrate documentation with readthedocs.org - Documentation for ifm3d could be significantly improved -- both the content and quality as well as the presentation. On the latter topic, this issue proposes integrating our documentation with the [Read The Docs](https://readthedocs.org) website. Putting the infrastructure in place to automatically build updated documentation upon release or commit to a particular branch would be ideal.
Also, let's start a list of what documentation needs to be made available:
- [x] Supported platforms
- [x] Building the software
- [x] Building and installing on Windows (should be treated separately from above)
- [x] How to implement a new image container
- [x] How to configure the imager using the JSON interface
- [ ] Related to above, all parameters need to be documented by the firmware/systems team
- [x] Using the `ifm3d` command-line tool
Please help me add to this list!
(NOTE: Is there a good way to make this check list shared so others can add to it? This is beyond my github skills).
We note that, per @graugans [comment](https://github.com/lovepark/ifm3d/issues/35#issuecomment-364398823) in #35 we will continue to provide code examples on using the library. The side benefit of making the examples a separate project also provides a nice way to illustrate how to integrate `ifm3d` into your project to include the build chain.
|
non_defect
|
improve documentation and integrate documentation with readthedocs org documentation for could be significantly improved both the content and quality as well as the presentation on the latter topic this issue proposes integrating our documentation with the website putting the infrastructure in place to automatically build updated documentation upon release or commit to a particular branch would be ideal also let s start a list of what documentation needs to be made available supported platforms building the software building and installing on windows should be treated separately from above how to implement a new image container how to configure the imager using the json interface related to above all parameters need to be documented by the firmware systems team using the command line tool please help me add to this list note is there a good way to make this check list shared so others can add to it this is beyond my github skills we note that per graugans in we will continue to provide code examples on using the library the side benefit of making the examples a separate project also provides a nice way to illustrate how to integrate into your project to include the build chain
| 0
|
38,471
| 8,848,146,275
|
IssuesEvent
|
2019-01-08 05:34:49
|
prettydiff/prettydiff
|
https://api.github.com/repos/prettydiff/prettydiff
|
opened
|
Some markup beautification trivialities
|
Beautification Defect
|
The code sample is the HTML of https://en.wikipedia.org/wiki/Functional_programming
1. attribute wrap - I need to evaluate the algorithm for when attributes are indented, because it does not appear to be working when a tag is wider than the value of options.width
1. comment indent - Near the bottom of the code sample there are two HTML comments that span multiple lines. The first of these is indented different than the second.
1. error - types[bb] undefined - This error appeared when editing the source sample using the web tool, but it isn't related to the webtool.
|
1.0
|
Some markup beautification trivialities - The code sample is the HTML of https://en.wikipedia.org/wiki/Functional_programming
1. attribute wrap - I need to evaluate the algorithm for when attributes are indented, because it does not appear to be working when a tag is wider than the value of options.width
1. comment indent - Near the bottom of the code sample there are two HTML comments that span multiple lines. The first of these is indented different than the second.
1. error - types[bb] undefined - This error appeared when editing the source sample using the web tool, but it isn't related to the webtool.
|
defect
|
some markup beautification trivialities the code sample is the html of attribute wrap i need to evaluate the algorithm for when attributes are indented because it does not appear to be working when a tag is wider than the value of options width comment indent near the bottom of the code sample there are two html comments that span multiple lines the first of these is indented different than the second error types undefined this error appeared when editing the source sample using the web tool but it isn t related to the webtool
| 1
|
60,457
| 17,023,429,382
|
IssuesEvent
|
2021-07-03 01:59:29
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Distance calculation shows up as NaN in the website
|
Component: yours Priority: minor Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 1.21pm, Thursday, 25th June 2009]**
The route distance is shown as 'NaN' in the website when calculating a route.
|
1.0
|
Distance calculation shows up as NaN in the website - **[Submitted to the original trac issue database at 1.21pm, Thursday, 25th June 2009]**
The route distance is shown as 'NaN' in the website when calculating a route.
|
defect
|
distance calculation shows up as nan in the website the route distance is shown as nan in the website when calculating a route
| 1
|
11,689
| 2,660,585,481
|
IssuesEvent
|
2015-03-19 08:44:56
|
miscalencu/Google-Chrome-Simple-Weather
|
https://api.github.com/repos/miscalencu/Google-Chrome-Simple-Weather
|
closed
|
characters not displaying correctly on Linux systems
|
auto-migrated Priority-Medium Type-Defect
|
```
On some systems like Linux, native alphabet is just not shown when the
specified font cannot handle native alphabet.
TO DO:
1. Add options for different fonts.
```
Original issue reported on code.google.com by `george.m...@gmail.com` on 29 Jan 2010 at 2:12
|
1.0
|
characters not displaying correctly on Linux systems - ```
On some systems like Linux, native alphabet is just not shown when the
specified font cannot handle native alphabet.
TO DO:
1. Add options for different fonts.
```
Original issue reported on code.google.com by `george.m...@gmail.com` on 29 Jan 2010 at 2:12
|
defect
|
characters not displaying correctly on linux systems on some systems like linux native alphabet is just not shown when the specified font cannot handle native alphabet to do add options for different fonts original issue reported on code google com by george m gmail com on jan at
| 1
|
179,950
| 13,912,801,036
|
IssuesEvent
|
2020-10-20 19:26:07
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
Nodes are left in AWS/DO after the cluster is deleted in Rancher:v2.4-head
|
[zube]: To Test kind/bug-qa status/blocker
|
<!--
Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue
For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase.
-->
**What kind of request is this (question/bug/enhancement/feature request):**
bug
**Steps to reproduce (least amount of steps as possible):**
- run rancher:`v2.4-3997-head` / `v2.4.9-rc1` docker install
- provision a cluster using either AWS or DO
- wait for the cluster to be active
- delete the cluster
**Result:**
The cluster is deleted in Rancher UI, but the nodes are left in AWS/DO.
Rancher's logs
```
2020/10/19 19:33:38 [INFO] Deleting nodePool [np-xg2lh]
2020/10/19 19:33:38 [INFO] Removing node jiaqi-test-delete1
2020/10/19 19:33:38 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
2020/10/19 19:33:38 [INFO] [mgmt-auth-crtb-controller] Deleting roleBinding clusterrolebinding-gsqtz
2020/10/19 19:33:38 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
2020/10/19 19:33:38 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
2020/10/19 19:33:38 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
2020/10/19 19:33:42 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
2020/10/19 19:33:43 [INFO] [mgmt-auth-prtb-controller] Updating owner label for roleBinding clusterrolebinding-csrjf
2020/10/19 19:33:43 [INFO] [mgmt-auth-prtb-controller] Deleting roleBinding clusterrolebinding-csrjf
2020/10/19 19:33:50 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
I1019 19:33:54.053455 26 namespace_controller.go:185] Namespace has been deleted p-s4jt4
I1019 19:33:54.062783 26 namespace_controller.go:185] Namespace has been deleted p-gxpk6
2020/10/19 19:34:06 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
2020/10/19 19:34:38 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
```
|
1.0
|
Nodes are left in AWS/DO after the cluster is deleted in Rancher:v2.4-head - <!--
Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue
For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase.
-->
**What kind of request is this (question/bug/enhancement/feature request):**
bug
**Steps to reproduce (least amount of steps as possible):**
- run rancher:`v2.4-3997-head` / `v2.4.9-rc1` docker install
- provision a cluster using either AWS or DO
- wait for the cluster to be active
- delete the cluster
**Result:**
The cluster is deleted in Rancher UI, but the nodes are left in AWS/DO.
Rancher's logs
```
2020/10/19 19:33:38 [INFO] Deleting nodePool [np-xg2lh]
2020/10/19 19:33:38 [INFO] Removing node jiaqi-test-delete1
2020/10/19 19:33:38 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
2020/10/19 19:33:38 [INFO] [mgmt-auth-crtb-controller] Deleting roleBinding clusterrolebinding-gsqtz
2020/10/19 19:33:38 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
2020/10/19 19:33:38 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
2020/10/19 19:33:38 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
2020/10/19 19:33:42 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
2020/10/19 19:33:43 [INFO] [mgmt-auth-prtb-controller] Updating owner label for roleBinding clusterrolebinding-csrjf
2020/10/19 19:33:43 [INFO] [mgmt-auth-prtb-controller] Deleting roleBinding clusterrolebinding-csrjf
2020/10/19 19:33:50 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
I1019 19:33:54.053455 26 namespace_controller.go:185] Namespace has been deleted p-s4jt4
I1019 19:33:54.062783 26 namespace_controller.go:185] Namespace has been deleted p-gxpk6
2020/10/19 19:34:06 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
2020/10/19 19:34:38 [ERROR] NodeController c-dldns/m-scsng [node-controller] failed with : cluster.management.cattle.io "c-dldns" not found
```
|
non_defect
|
nodes are left in aws do after the cluster is deleted in rancher head please search for existing issues first then read to see what we expect in an issue for security issues please email security rancher com instead of posting a public issue in github you may but are not required to use the gpg key located on keybase what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible run rancher head docker install provision a cluster using either aws or do wait for the cluster to be active delete the cluster result the cluster is deleted in rancher ui but the nodes are left in aws do rancher s logs deleting nodepool removing node jiaqi test nodecontroller c dldns m scsng failed with cluster management cattle io c dldns not found deleting rolebinding clusterrolebinding gsqtz nodecontroller c dldns m scsng failed with cluster management cattle io c dldns not found nodecontroller c dldns m scsng failed with cluster management cattle io c dldns not found nodecontroller c dldns m scsng failed with cluster management cattle io c dldns not found nodecontroller c dldns m scsng failed with cluster management cattle io c dldns not found updating owner label for rolebinding clusterrolebinding csrjf deleting rolebinding clusterrolebinding csrjf nodecontroller c dldns m scsng failed with cluster management cattle io c dldns not found namespace controller go namespace has been deleted p namespace controller go namespace has been deleted p nodecontroller c dldns m scsng failed with cluster management cattle io c dldns not found nodecontroller c dldns m scsng failed with cluster management cattle io c dldns not found
| 0
|
314,284
| 26,990,356,198
|
IssuesEvent
|
2023-02-09 19:21:54
|
acikkaynak/deprem-yardim-frontend
|
https://api.github.com/repos/acikkaynak/deprem-yardim-frontend
|
closed
|
bug: Konum detaylarının sayfa sınırlarını aştığı kısımlar görüntülenemiyor
|
bug approved test-failed
|
## Hata Tanımı
Konum detaylarının sayfa sınırlarını aştığı kısımlar görüntülenemiyor. Web üzerinde Edge, Opera, Chrome, Mobilde Safari ve Chrome'da hata reproduce edildi. Özellikle mobilde ciddi sorun oluşturduğu gözlemlenmiştir.
-- bu bilgiyi ekleyiniz
notDepresseDeveloper#0919
## Hata Ortamı
rc.afetharita.com
afetharita.com
## Hatayı nasıl aldığınızı adım adım açıklayın
1. Bir lokasyondaki içerik kontol edilmeye çalışılır.
2. Yardım içeriğinin tamamının son satıra kadar görüntülenip görüntülenmediği kontrol edilir.
3. Uzun içerikler için de 2. işlem kontrol edilir.
## Beklenen Davranış
Beklediğiniz şeyin açık ve kısa bir açıklaması.
Yardım içeriğinin tamamının Twitter fonksiyonlarıyla birlikte (paylaş butonu vs.) görüntülenebilmesi/scrollanabilmesi.
## Ekran Görüntüleri

## Masaüstü Bilgileri
- İşletim Sistemi: Windows
- Tarayıcı : Edge, Opera, Chrome
- Sürüm : En güncel sürümler
## Akıllı Telefon Bilgileri
- Cihaz: iOS (Ipad 6th Gen), Android
- İşletim Sistemi: iOS 14.2
- Tarayıcı : IOS'ta Safari ve Chrome, Android'de Chrome
- Sürüm : En güncel sürümler
|
1.0
|
bug: Konum detaylarının sayfa sınırlarını aştığı kısımlar görüntülenemiyor - ## Hata Tanımı
Konum detaylarının sayfa sınırlarını aştığı kısımlar görüntülenemiyor. Web üzerinde Edge, Opera, Chrome, Mobilde Safari ve Chrome'da hata reproduce edildi. Özellikle mobilde ciddi sorun oluşturduğu gözlemlenmiştir.
-- bu bilgiyi ekleyiniz
notDepresseDeveloper#0919
## Hata Ortamı
rc.afetharita.com
afetharita.com
## Hatayı nasıl aldığınızı adım adım açıklayın
1. Bir lokasyondaki içerik kontol edilmeye çalışılır.
2. Yardım içeriğinin tamamının son satıra kadar görüntülenip görüntülenmediği kontrol edilir.
3. Uzun içerikler için de 2. işlem kontrol edilir.
## Beklenen Davranış
Beklediğiniz şeyin açık ve kısa bir açıklaması.
Yardım içeriğinin tamamının Twitter fonksiyonlarıyla birlikte (paylaş butonu vs.) görüntülenebilmesi/scrollanabilmesi.
## Ekran Görüntüleri

## Masaüstü Bilgileri
- İşletim Sistemi: Windows
- Tarayıcı : Edge, Opera, Chrome
- Sürüm : En güncel sürümler
## Akıllı Telefon Bilgileri
- Cihaz: iOS (Ipad 6th Gen), Android
- İşletim Sistemi: iOS 14.2
- Tarayıcı : IOS'ta Safari ve Chrome, Android'de Chrome
- Sürüm : En güncel sürümler
|
non_defect
|
bug konum detaylarının sayfa sınırlarını aştığı kısımlar görüntülenemiyor hata tanımı konum detaylarının sayfa sınırlarını aştığı kısımlar görüntülenemiyor web üzerinde edge opera chrome mobilde safari ve chrome da hata reproduce edildi özellikle mobilde ciddi sorun oluşturduğu gözlemlenmiştir bu bilgiyi ekleyiniz notdepressedeveloper hata ortamı rc afetharita com afetharita com hatayı nasıl aldığınızı adım adım açıklayın bir lokasyondaki içerik kontol edilmeye çalışılır yardım içeriğinin tamamının son satıra kadar görüntülenip görüntülenmediği kontrol edilir uzun içerikler için de işlem kontrol edilir beklenen davranış beklediğiniz şeyin açık ve kısa bir açıklaması yardım içeriğinin tamamının twitter fonksiyonlarıyla birlikte paylaş butonu vs görüntülenebilmesi scrollanabilmesi ekran görüntüleri masaüstü bilgileri i̇şletim sistemi windows tarayıcı edge opera chrome sürüm en güncel sürümler akıllı telefon bilgileri cihaz ios ipad gen android i̇şletim sistemi ios tarayıcı ios ta safari ve chrome android de chrome sürüm en güncel sürümler
| 0
|
767,670
| 26,935,882,627
|
IssuesEvent
|
2023-02-07 20:36:38
|
googleapis/google-api-java-client
|
https://api.github.com/repos/googleapis/google-api-java-client
|
closed
|
Initializing Google Drive API using builder giving Exception
|
type: bug priority: p3
|
#### Environment details
1. Specify the API at the beginning of the title. For example, "BigQuery: ...").
General, Core, and Other are also allowed as types
2. OS type and version: Android OS 10/11/12/13
3. Java version: depending on Android devices
4. Version(s): com.google.api-client:google-api-client:1.31.1
#### Steps to reproduce
It is not reproducible at our end and getting reported on firebase.
#### Code example
```java
Drive.Builder(netHttpTransport, jacksonFactory, setHttpTimeout(accountCredential)).build()
```
#### Stack trace
```
Caused by java.lang.IllegalStateException: No successful match so far
at java.util.regex.Matcher.ensureMatch(Matcher.java:1189)
at java.util.regex.Matcher.group(Matcher.java:456)
at com.google.api.client.googleapis.GoogleUtils.<clinit>(GoogleUtils.java:66)
at com.google.api.services.drive.Drive.<clinit>(Drive.java:44)
at com.google.api.services.drive.Drive$Builder.build(Drive.java:9813)
```
#### Any additional information below
This is similar to https://github.com/googleapis/google-api-java-client/issues/1451 which is marked closed but is still reproducible on google api client 1.31.1.
Please let me know for any additional details.
Following these steps guarantees the quickest resolution possible.
Thanks!
|
1.0
|
Initializing Google Drive API using builder giving Exception - #### Environment details
1. Specify the API at the beginning of the title. For example, "BigQuery: ...").
General, Core, and Other are also allowed as types
2. OS type and version: Android OS 10/11/12/13
3. Java version: depending on Android devices
4. Version(s): com.google.api-client:google-api-client:1.31.1
#### Steps to reproduce
It is not reproducible at our end and getting reported on firebase.
#### Code example
```java
Drive.Builder(netHttpTransport, jacksonFactory, setHttpTimeout(accountCredential)).build()
```
#### Stack trace
```
Caused by java.lang.IllegalStateException: No successful match so far
at java.util.regex.Matcher.ensureMatch(Matcher.java:1189)
at java.util.regex.Matcher.group(Matcher.java:456)
at com.google.api.client.googleapis.GoogleUtils.<clinit>(GoogleUtils.java:66)
at com.google.api.services.drive.Drive.<clinit>(Drive.java:44)
at com.google.api.services.drive.Drive$Builder.build(Drive.java:9813)
```
#### Any additional information below
This is similar to https://github.com/googleapis/google-api-java-client/issues/1451 which is marked closed but is still reproducible on google api client 1.31.1.
Please let me know for any additional details.
Following these steps guarantees the quickest resolution possible.
Thanks!
|
non_defect
|
initializing google drive api using builder giving exception environment details specify the api at the beginning of the title for example bigquery general core and other are also allowed as types os type and version android os java version depending on android devices version s com google api client google api client steps to reproduce it is not reproducible at our end and getting reported on firebase code example java drive builder nethttptransport jacksonfactory sethttptimeout accountcredential build stack trace caused by java lang illegalstateexception no successful match so far at java util regex matcher ensurematch matcher java at java util regex matcher group matcher java at com google api client googleapis googleutils googleutils java at com google api services drive drive drive java at com google api services drive drive builder build drive java any additional information below this is similar to which is marked closed but is still reproducible on google api client please let me know for any additional details following these steps guarantees the quickest resolution possible thanks
| 0
|
135,954
| 12,695,601,605
|
IssuesEvent
|
2020-06-22 08:41:47
|
ocaml/ocaml
|
https://api.github.com/repos/ocaml/ocaml
|
reopened
|
Feature: Recommend ocamlc as driver for compiling C sources for stublibs
|
documentation feature-wish newcomer-job
|
**Original bug ID:** 5733
**Reporter:** gerd
**Status:** confirmed (set by @damiendoligez on 2012-11-15T14:30:52Z)
**Resolution:** open
**Priority:** normal
**Severity:** feature
**Version:** 4.00.0
**Category:** documentation
**Tags:** junior_job
**Monitored by:** @ygrek @hcarty @dbuenzli
## Bug description
Basically, it is up to the user how to call the C compiler for compiling stublibs. However, in many build systems I see that users do it the wrong way (e.g. hardcode the compiler "gcc", or forget to pass switches like -fPIC which are sometimes required). My wish is that the section "Interfacing C with OCaml" includes examples how to do it right - which is absolutely simple, because you just need to use ocamlc as driver even for C sources.
There could e.g. be a new section before "Statically linking C code with OCaml code" about "Compiling C code". Contents:
- How to use ocamlc as driver for C sources: Sample command
- Mention that the stdlib is automatically in the include path
- Mention that the switches are ok for dynamically loadable code
- The created .o files can be args of ocamlmklib
- use -I, -ccopt, -cclib to pass args
- Mention that ocamlc -config can be used to extract the compiler path and switches if lower-level calls are needed
|
1.0
|
Feature: Recommend ocamlc as driver for compiling C sources for stublibs - **Original bug ID:** 5733
**Reporter:** gerd
**Status:** confirmed (set by @damiendoligez on 2012-11-15T14:30:52Z)
**Resolution:** open
**Priority:** normal
**Severity:** feature
**Version:** 4.00.0
**Category:** documentation
**Tags:** junior_job
**Monitored by:** @ygrek @hcarty @dbuenzli
## Bug description
Basically, it is up to the user how to call the C compiler for compiling stublibs. However, in many build systems I see that users do it the wrong way (e.g. hardcode the compiler "gcc", or forget to pass switches like -fPIC which are sometimes required). My wish is that the section "Interfacing C with OCaml" includes examples how to do it right - which is absolutely simple, because you just need to use ocamlc as driver even for C sources.
There could e.g. be a new section before "Statically linking C code with OCaml code" about "Compiling C code". Contents:
- How to use ocamlc as driver for C sources: Sample command
- Mention that the stdlib is automatically in the include path
- Mention that the switches are ok for dynamically loadable code
- The created .o files can be args of ocamlmklib
- use -I, -ccopt, -cclib to pass args
- Mention that ocamlc -config can be used to extract the compiler path and switches if lower-level calls are needed
|
non_defect
|
feature recommend ocamlc as driver for compiling c sources for stublibs original bug id reporter gerd status confirmed set by damiendoligez on resolution open priority normal severity feature version category documentation tags junior job monitored by ygrek hcarty dbuenzli bug description basically it is up to the user how to call the c compiler for compiling stublibs however in many build systems i see that users do it the wrong way e g hardcode the compiler gcc or forget to pass switches like fpic which are sometimes required my wish is that the section interfacing c with ocaml includes examples how to do it right which is absolutely simple because you just need to use ocamlc as driver even for c sources there could e g be a new section before statically linking c code with ocaml code about compiling c code contents how to use ocamlc as driver for c sources sample command mention that the stdlib is automatically in the include path mention that the switches are ok for dynamically loadable code the created o files can be args of ocamlmklib use i ccopt cclib to pass args mention that ocamlc config can be used to extract the compiler path and switches if lower level calls are needed
| 0
|
438,460
| 30,643,078,909
|
IssuesEvent
|
2023-07-25 00:42:42
|
tyler-technologies-oss/forge
|
https://api.github.com/repos/tyler-technologies-oss/forge
|
opened
|
[radio] Storybook incorrect CSS property listed
|
bug documentation complexity: low
|
https://github.com/tyler-technologies-oss/forge/blob/b38e597366ad33a46b4b1dc267fb96e9b86d4a33/src/stories/src/components/radio-button/radio-button.mdx#L104
The current value listed in Storybook appears to have no styling impact, the correct variable is `--forge-theme-label-disabled-on-background`
|
1.0
|
[radio] Storybook incorrect CSS property listed - https://github.com/tyler-technologies-oss/forge/blob/b38e597366ad33a46b4b1dc267fb96e9b86d4a33/src/stories/src/components/radio-button/radio-button.mdx#L104
The current value listed in Storybook appears to have no styling impact, the correct variable is `--forge-theme-label-disabled-on-background`
|
non_defect
|
storybook incorrect css property listed the current value listed in storybook appears to have no styling impact the correct variable is forge theme label disabled on background
| 0
|
88,813
| 10,580,512,960
|
IssuesEvent
|
2019-10-08 06:58:38
|
Jogans/Gruppe1Semester4
|
https://api.github.com/repos/Jogans/Gruppe1Semester4
|
closed
|
Oprettelse af første udkast til arkitektur - Arbejdstegning
|
documentation
|
Her tænkes et overordnet view af vores arkitektur, som skal give overblik over vores projekt
|
1.0
|
Oprettelse af første udkast til arkitektur - Arbejdstegning - Her tænkes et overordnet view af vores arkitektur, som skal give overblik over vores projekt
|
non_defect
|
oprettelse af første udkast til arkitektur arbejdstegning her tænkes et overordnet view af vores arkitektur som skal give overblik over vores projekt
| 0
|
52,201
| 13,211,406,547
|
IssuesEvent
|
2020-08-15 22:55:14
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
[dst] I3DSTExtractor16 tries to split P frames (Trac #1843)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1843">https://code.icecube.wisc.edu/projects/icecube/ticket/1843</a>, reported by jvansantenand owned by juancarlos</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:38",
"_ts": "1550067158057333",
"description": "test dst16.py fails because it operates on P frames, and attempts to split them into subframes. This failed cryptically before, but now fails explicitly.",
"reporter": "jvansanten",
"cc": "",
"resolution": "fixed",
"time": "2016-08-31T08:52:40",
"component": "combo reconstruction",
"summary": "[dst] I3DSTExtractor16 tries to split P frames",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "juancarlos",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[dst] I3DSTExtractor16 tries to split P frames (Trac #1843) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1843">https://code.icecube.wisc.edu/projects/icecube/ticket/1843</a>, reported by jvansantenand owned by juancarlos</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:38",
"_ts": "1550067158057333",
"description": "test dst16.py fails because it operates on P frames, and attempts to split them into subframes. This failed cryptically before, but now fails explicitly.",
"reporter": "jvansanten",
"cc": "",
"resolution": "fixed",
"time": "2016-08-31T08:52:40",
"component": "combo reconstruction",
"summary": "[dst] I3DSTExtractor16 tries to split P frames",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "juancarlos",
"type": "defect"
}
```
</p>
</details>
|
defect
|
tries to split p frames trac migrated from json status closed changetime ts description test py fails because it operates on p frames and attempts to split them into subframes this failed cryptically before but now fails explicitly reporter jvansanten cc resolution fixed time component combo reconstruction summary tries to split p frames priority normal keywords milestone owner juancarlos type defect
| 1
|
37,773
| 8,517,419,124
|
IssuesEvent
|
2018-11-01 07:54:57
|
GoldenSoftwareLtd/gedemin
|
https://api.github.com/repos/GoldenSoftwareLtd/gedemin
|
closed
|
Ошибка и пожелание по утилите конвертации БД
|
FDBConvert Priority-Medium Type-Defect
|
Originally reported on Google Code with ID 2538
```
1. При конвертации получаю:
```
Внимание
Произошла ошибка при компиляции триггера:
Input parameter mismatch for procedure AGG_TUPLE_CHANGED
```
Вот вызов самой процедуры в триггере:
execute procedure agg_tuple_changed('+', new.tuple_id);
А сама процедура описана с параметром по умолчанию:
create or alter procedure AGG_TUPLE_CHANGED (
OPERATION_SIGN varchar(1),
TUPLE_ID integer,
ASSOCIATION_ID integer = null)
Т.е., ошибки не должно быть. При перекачке с помощью скрипта (и т.д.) все ОК.
2. Возможно, есть смысл запоминать используемый утилитой язык диалога, ну или по умолчанию
использовать системный.
```
Reported by `gs1994` on 2011-07-23 18:28:57
|
1.0
|
Ошибка и пожелание по утилите конвертации БД - Originally reported on Google Code with ID 2538
```
1. При конвертации получаю:
```
Внимание
Произошла ошибка при компиляции триггера:
Input parameter mismatch for procedure AGG_TUPLE_CHANGED
```
Вот вызов самой процедуры в триггере:
execute procedure agg_tuple_changed('+', new.tuple_id);
А сама процедура описана с параметром по умолчанию:
create or alter procedure AGG_TUPLE_CHANGED (
OPERATION_SIGN varchar(1),
TUPLE_ID integer,
ASSOCIATION_ID integer = null)
Т.е., ошибки не должно быть. При перекачке с помощью скрипта (и т.д.) все ОК.
2. Возможно, есть смысл запоминать используемый утилитой язык диалога, ну или по умолчанию
использовать системный.
```
Reported by `gs1994` on 2011-07-23 18:28:57
|
defect
|
ошибка и пожелание по утилите конвертации бд originally reported on google code with id при конвертации получаю внимание произошла ошибка при компиляции триггера input parameter mismatch for procedure agg tuple changed вот вызов самой процедуры в триггере execute procedure agg tuple changed new tuple id а сама процедура описана с параметром по умолчанию create or alter procedure agg tuple changed operation sign varchar tuple id integer association id integer null т е ошибки не должно быть при перекачке с помощью скрипта и т д все ок возможно есть смысл запоминать используемый утилитой язык диалога ну или по умолчанию использовать системный reported by on
| 1
|
171,598
| 14,332,503,215
|
IssuesEvent
|
2020-11-27 02:42:08
|
justinzliu/myNeighbourhood
|
https://api.github.com/repos/justinzliu/myNeighbourhood
|
opened
|
Prepare application to showcase basic design and functionality
|
documentation
|
- Project overview including: purpose, initialization procedure, technology stack, application features
- Application framework: project distribution system structure, project file structure, valuation of project feasibility
|
1.0
|
Prepare application to showcase basic design and functionality - - Project overview including: purpose, initialization procedure, technology stack, application features
- Application framework: project distribution system structure, project file structure, valuation of project feasibility
|
non_defect
|
prepare application to showcase basic design and functionality project overview including purpose initialization procedure technology stack application features application framework project distribution system structure project file structure valuation of project feasibility
| 0
|
7,975
| 4,116,526,949
|
IssuesEvent
|
2016-06-08 01:08:34
|
MRtrix3/mrtrix3
|
https://api.github.com/repos/MRtrix3/mrtrix3
|
opened
|
build: Detect and handle RAM errors
|
build wishlist
|
Spawned from #655.
If a compilation task fails due to running out of RAM, add it to a list of tasks that should be run single-threaded. Once all multi-threaded tasks that can be run have been run, execute the single-threaded tasks sequentially. Then, go back to multi-threading. Repeat until there are no tasks left.
Tagged as wishlist, but it seems more and more people are running into memory issues during compilation.
|
1.0
|
build: Detect and handle RAM errors - Spawned from #655.
If a compilation task fails due to running out of RAM, add it to a list of tasks that should be run single-threaded. Once all multi-threaded tasks that can be run have been run, execute the single-threaded tasks sequentially. Then, go back to multi-threading. Repeat until there are no tasks left.
Tagged as wishlist, but it seems more and more people are running into memory issues during compilation.
|
non_defect
|
build detect and handle ram errors spawned from if a compilation task fails due to running out of ram add it to a list of tasks that should be run single threaded once all multi threaded tasks that can be run have been run execute the single threaded tasks sequentially then go back to multi threading repeat until there are no tasks left tagged as wishlist but it seems more and more people are running into memory issues during compilation
| 0
|
26,232
| 4,634,601,159
|
IssuesEvent
|
2016-09-29 02:03:44
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
Check if an item belongs to a hasMany association before updating in CakePHP 3.3.4
|
Defect Enhancement On hold ORM validation
|
This is a (multiple allowed):
* [x] bug
* [x] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.3.4
* Platform and Target: Ubuntu 14.04.1 LTS, Plesk 12.5.30, PHP 7+, MySQL
### What you did
**What I am trying to do:**
I have Estimates and Estimates have items "EstimateItems". When updating a Estimate the EstimateItems changed should update. (using patchEntity)
This is working with my current code, my only problem is that other users can edit the Estimate Items of other users when changing the primary key of a EstimateItem in the edit form, because when patching the existing EstimateItems CakePHP only looks at the primary key of the EstimateItem and doesn't take the association in consideration.
**_Also it's still possible to edit the estimate_id of a EstimateItem while $_accessible estimate_id is set to false._**
So what I need is CakePHP to validate that this EstimateItem belongs to the current association before updating or while trying to update.
**Current Query**
```
UPDATE
estimate_items
SET
data = 'Test Query 1',
amount = 123456789,
tax_id = 3
WHERE
id = 3
```
**Expected Query**
```
UPDATE
estimate_items
SET
data = 'Test Query 1',
amount = 123456789,
tax_id = 3
WHERE
id = 3 AND estimate_id = 1
```
**Current code:**
**Estimates -> Edit.ctp**
```
<?php $this->Form->templates($formTemplates['default']); ?>
<?= $this->Form->create($estimate, ['enctype' => 'multipart/form-data']) ?>
<fieldset>
<legend><?= __('Offerte') ?></legend>
<?= $this->Form->input('reference', ['label' => __('#Referentie'), 'autocomplete' => 'off']) ?>
<?= $this->Form->input('client_id',
[
'type' => 'select',
'empty' => true,
'label' => __('Klant'),
'options' => $clients
]
)
?>
<?php
foreach($estimate->estimate_items as $key => $item){
?>
<div class="item">
<legend>Item</legend>
<?= $this->Form->hidden('estimate_items.'. $key .'.id') ?>
<?= $this->Form->input('estimate_items.'. $key .'.data', ['type' => 'text', 'label' => __('Beschrijving')]) ?>
<?= $this->Form->input('estimate_items.'. $key .'.amount', ['type' => 'text', 'label' => __('Bedrag'), 'class' => 'input-date']) ?>
<?= $this->Form->input('estimate_items.'. $key .'.tax_id',
[
'type' => 'select',
'empty' => true,
'label' => __('Belasting type'),
'options' => $taxes
]
)
?>
</div>
<?php
}
?>
<legend>Informatie</legend>
<?= $this->Form->input('date', ['type' => 'text', 'label' => __('Offerte datum'), 'autocomplete' => 'off']) ?>
<?= $this->Form->input('expiration', ['type' => 'text', 'label' => __('Verloop datum'), 'autocomplete' => 'off']) ?>
</fieldset>
<?= $this->Form->button(__('Save')); ?>
<?= $this->Form->end() ?>
```
**Estimates Controller**
```
namespace App\Controller;
use App\Controller\AppController;
use Cake\Event\Event;
use Cake\ORM\TableRegistry;
class EstimatesController extends AppController
{
public function edit($id){
$associated = ['EstimateItems'];
$estimate = $this->Estimates->get($id, ['contain' => $associated]);
$this->log($estimate);
if($this->request->is(['patch', 'post', 'put'])) {
$estimate = $this->Estimates->patchEntity($estimate, $this->request->data, [
'associated' => $associated
]);
$estimate->total = '0';
$this->log($estimate);
$this->log($this->request->data);
if($this->Estimates->save($estimate, ['associated' => $associated])){
$this->Flash->success(__('De offerte is bijgewerkt'));
return $this->redirect(['action' => 'index']);
}
}
$this->set('taxes', $this->Estimates->Taxes->find('list', [ 'keyField' => 'id', 'valueField' => 'tax_name' ]));
$this->set('clients', $this->Estimates->Clients->find('list', [ 'keyField' => 'id', 'valueField' => 'companyname' ]));
$this->set('estimate', $estimate);
}
}
```
**EstimatesTable**
```
<?php
namespace App\Model\Table;
use Cake\ORM\Query;
use Cake\ORM\Table;
use Cake\Validation\Validator;
use Cake\ORM\RulesChecker;
use Cake\ORM\Rule\IsUnique;
class EstimatesTable extends Table
{
public function initialize(array $config)
{
$this->addAssociations([
'hasOne' => ['Taxes'],
'belongsTo' => ['Companies', 'Clients'],
'hasMany' => ['EstimateItems' => [
'foreignKey' => 'estimate_id'
]]
]);
}
public function buildRules(RulesChecker $rules){
// A Node however should in addition also always reference a Site.
$rules->add($rules->existsIn(['estimate_id'], 'EstimateItems'));
return $rules;
}
}
```
**EstimateItem Entity**
```
<?php
namespace App\Model\Entity;
use Cake\ORM\Entity;
class EstimateItem extends Entity
{
protected $_accessible = [
'*' => false,
'data' => true,
'amount' => true,
'tax_id' => true,
'unit_id' => true
];
}
```
**EstimateItemsTable**
```
<?php
namespace App\Model\Table;
use Cake\ORM\Entity;
use Cake\ORM\Table;
use Cake\Validation\Validator;
use Cake\ORM\RulesChecker;
use Cake\ORM\Rule\IsUnique;
use Cake\ORM\Query;
class EstimateItemsTable extends Table
{
public function initialize(array $config)
{
$this->addAssociations([
'belongsTo' => ['Estimates' => ['foreignKey' => 'estimate_id']],
'hasOne' => ['Taxes' => ['foreignKey' => 'tax_id']]
]);
}
}
```
**Estimate Entity**
```
<?php
namespace App\Model\Entity;
use Cake\ORM\Entity;
class Estimate extends Entity
{
/**
* Fields that can be mass assigned using newEntity() or patchEntity().
*
* Note that when '*' is set to true, this allows all unspecified fields to
* be mass assigned. For security purposes, it is advised to set '*' to false
* (or remove it), and explicitly make individual fields accessible as needed.
*
* @var array
*/
protected $_accessible = [
'*' => false,
'id' => false,
];
}
```
### Expected Behavior
When a belongsto entity gets updated by a hasmany association it should only update when the entity being updated actually belongsto the assocation which called the update.
### Actual Behavior
Entity that belongsto a hasmany association still gets updated while the update gets called by the belongsto association while it does not belong to the updated entity. What happens is that the association automaticly changes the "estimate_id" (foreign key) and makes it did not belong to the hasmany in the first place. As seen in the example code.
|
1.0
|
Check if an item belongs to a hasMany association before updating in CakePHP 3.3.4 - This is a (multiple allowed):
* [x] bug
* [x] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.3.4
* Platform and Target: Ubuntu 14.04.1 LTS, Plesk 12.5.30, PHP 7+, MySQL
### What you did
**What I am trying to do:**
I have Estimates and Estimates have items "EstimateItems". When updating a Estimate the EstimateItems changed should update. (using patchEntity)
This is working with my current code, my only problem is that other users can edit the Estimate Items of other users when changing the primary key of a EstimateItem in the edit form, because when patching the existing EstimateItems CakePHP only looks at the primary key of the EstimateItem and doesn't take the association in consideration.
**_Also it's still possible to edit the estimate_id of a EstimateItem while $_accessible estimate_id is set to false._**
So what I need is CakePHP to validate that this EstimateItem belongs to the current association before updating or while trying to update.
**Current Query**
```
UPDATE
estimate_items
SET
data = 'Test Query 1',
amount = 123456789,
tax_id = 3
WHERE
id = 3
```
**Expected Query**
```
UPDATE
estimate_items
SET
data = 'Test Query 1',
amount = 123456789,
tax_id = 3
WHERE
id = 3 AND estimate_id = 1
```
**Current code:**
**Estimates -> Edit.ctp**
```
<?php $this->Form->templates($formTemplates['default']); ?>
<?= $this->Form->create($estimate, ['enctype' => 'multipart/form-data']) ?>
<fieldset>
<legend><?= __('Offerte') ?></legend>
<?= $this->Form->input('reference', ['label' => __('#Referentie'), 'autocomplete' => 'off']) ?>
<?= $this->Form->input('client_id',
[
'type' => 'select',
'empty' => true,
'label' => __('Klant'),
'options' => $clients
]
)
?>
<?php
foreach($estimate->estimate_items as $key => $item){
?>
<div class="item">
<legend>Item</legend>
<?= $this->Form->hidden('estimate_items.'. $key .'.id') ?>
<?= $this->Form->input('estimate_items.'. $key .'.data', ['type' => 'text', 'label' => __('Beschrijving')]) ?>
<?= $this->Form->input('estimate_items.'. $key .'.amount', ['type' => 'text', 'label' => __('Bedrag'), 'class' => 'input-date']) ?>
<?= $this->Form->input('estimate_items.'. $key .'.tax_id',
[
'type' => 'select',
'empty' => true,
'label' => __('Belasting type'),
'options' => $taxes
]
)
?>
</div>
<?php
}
?>
<legend>Informatie</legend>
<?= $this->Form->input('date', ['type' => 'text', 'label' => __('Offerte datum'), 'autocomplete' => 'off']) ?>
<?= $this->Form->input('expiration', ['type' => 'text', 'label' => __('Verloop datum'), 'autocomplete' => 'off']) ?>
</fieldset>
<?= $this->Form->button(__('Save')); ?>
<?= $this->Form->end() ?>
```
**Estimates Controller**
```
namespace App\Controller;
use App\Controller\AppController;
use Cake\Event\Event;
use Cake\ORM\TableRegistry;
class EstimatesController extends AppController
{
public function edit($id){
$associated = ['EstimateItems'];
$estimate = $this->Estimates->get($id, ['contain' => $associated]);
$this->log($estimate);
if($this->request->is(['patch', 'post', 'put'])) {
$estimate = $this->Estimates->patchEntity($estimate, $this->request->data, [
'associated' => $associated
]);
$estimate->total = '0';
$this->log($estimate);
$this->log($this->request->data);
if($this->Estimates->save($estimate, ['associated' => $associated])){
$this->Flash->success(__('De offerte is bijgewerkt'));
return $this->redirect(['action' => 'index']);
}
}
$this->set('taxes', $this->Estimates->Taxes->find('list', [ 'keyField' => 'id', 'valueField' => 'tax_name' ]));
$this->set('clients', $this->Estimates->Clients->find('list', [ 'keyField' => 'id', 'valueField' => 'companyname' ]));
$this->set('estimate', $estimate);
}
}
```
**EstimatesTable**
```
<?php
namespace App\Model\Table;
use Cake\ORM\Query;
use Cake\ORM\Table;
use Cake\Validation\Validator;
use Cake\ORM\RulesChecker;
use Cake\ORM\Rule\IsUnique;
class EstimatesTable extends Table
{
public function initialize(array $config)
{
$this->addAssociations([
'hasOne' => ['Taxes'],
'belongsTo' => ['Companies', 'Clients'],
'hasMany' => ['EstimateItems' => [
'foreignKey' => 'estimate_id'
]]
]);
}
public function buildRules(RulesChecker $rules){
// A Node however should in addition also always reference a Site.
$rules->add($rules->existsIn(['estimate_id'], 'EstimateItems'));
return $rules;
}
}
```
**EstimateItem Entity**
```
<?php
namespace App\Model\Entity;
use Cake\ORM\Entity;
class EstimateItem extends Entity
{
protected $_accessible = [
'*' => false,
'data' => true,
'amount' => true,
'tax_id' => true,
'unit_id' => true
];
}
```
**EstimateItemsTable**
```
<?php
namespace App\Model\Table;
use Cake\ORM\Entity;
use Cake\ORM\Table;
use Cake\Validation\Validator;
use Cake\ORM\RulesChecker;
use Cake\ORM\Rule\IsUnique;
use Cake\ORM\Query;
class EstimateItemsTable extends Table
{
public function initialize(array $config)
{
$this->addAssociations([
'belongsTo' => ['Estimates' => ['foreignKey' => 'estimate_id']],
'hasOne' => ['Taxes' => ['foreignKey' => 'tax_id']]
]);
}
}
```
**Estimate Entity**
```
<?php
namespace App\Model\Entity;
use Cake\ORM\Entity;
class Estimate extends Entity
{
/**
* Fields that can be mass assigned using newEntity() or patchEntity().
*
* Note that when '*' is set to true, this allows all unspecified fields to
* be mass assigned. For security purposes, it is advised to set '*' to false
* (or remove it), and explicitly make individual fields accessible as needed.
*
* @var array
*/
protected $_accessible = [
'*' => false,
'id' => false,
];
}
```
### Expected Behavior
When a belongsto entity gets updated by a hasmany association it should only update when the entity being updated actually belongsto the assocation which called the update.
### Actual Behavior
Entity that belongsto a hasmany association still gets updated while the update gets called by the belongsto association while it does not belong to the updated entity. What happens is that the association automaticly changes the "estimate_id" (foreign key) and makes it did not belong to the hasmany in the first place. As seen in the example code.
|
defect
|
check if an item belongs to a hasmany association before updating in cakephp this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target ubuntu lts plesk php mysql what you did what i am trying to do i have estimates and estimates have items estimateitems when updating a estimate the estimateitems changed should update using patchentity this is working with my current code my only problem is that other users can edit the estimate items of other users when changing the primary key of a estimateitem in the edit form because when patching the existing estimateitems cakephp only looks at the primary key of the estimateitem and doesn t take the association in consideration also it s still possible to edit the estimate id of a estimateitem while accessible estimate id is set to false so what i need is cakephp to validate that this estimateitem belongs to the current association before updating or while trying to update current query update estimate items set data test query amount tax id where id expected query update estimate items set data test query amount tax id where id and estimate id current code estimates edit ctp form templates formtemplates form create estimate form input reference form input client id type select empty true label klant options clients php foreach estimate estimate items as key item item form hidden estimate items key id form input estimate items key data form input estimate items key amount form input estimate items key tax id type select empty true label belasting type options taxes php informatie form input date form input expiration form button save form end estimates controller namespace app controller use app controller appcontroller use cake event event use cake orm tableregistry class estimatescontroller extends appcontroller public function edit id associated estimate this estimates get id this log estimate if this request is estimate this estimates patchentity estimate this request data associated associated estimate total this log estimate this log this request data if this estimates save estimate this flash success de offerte is bijgewerkt return this redirect this set taxes this estimates taxes find list this set clients this estimates clients find list this set estimate estimate estimatestable php namespace app model table use cake orm query use cake orm table use cake validation validator use cake orm ruleschecker use cake orm rule isunique class estimatestable extends table public function initialize array config this addassociations hasone belongsto hasmany estimateitems foreignkey estimate id public function buildrules ruleschecker rules a node however should in addition also always reference a site rules add rules existsin estimateitems return rules estimateitem entity php namespace app model entity use cake orm entity class estimateitem extends entity protected accessible false data true amount true tax id true unit id true estimateitemstable php namespace app model table use cake orm entity use cake orm table use cake validation validator use cake orm ruleschecker use cake orm rule isunique use cake orm query class estimateitemstable extends table public function initialize array config this addassociations belongsto hasone estimate entity php namespace app model entity use cake orm entity class estimate extends entity fields that can be mass assigned using newentity or patchentity note that when is set to true this allows all unspecified fields to be mass assigned for security purposes it is advised to set to false or remove it and explicitly make individual fields accessible as needed var array protected accessible false id false expected behavior when a belongsto entity gets updated by a hasmany association it should only update when the entity being updated actually belongsto the assocation which called the update actual behavior entity that belongsto a hasmany association still gets updated while the update gets called by the belongsto association while it does not belong to the updated entity what happens is that the association automaticly changes the estimate id foreign key and makes it did not belong to the hasmany in the first place as seen in the example code
| 1
|
262,908
| 8,272,598,264
|
IssuesEvent
|
2018-09-16 21:56:39
|
javaee/glassfish
|
https://api.github.com/repos/javaee/glassfish
|
closed
|
Add details to cluster creation failure SEVERE log message
|
3_1_1-scrubbed Component: configuration ERR: Assignee Priority: Major Type: Improvement
|
After running thousands of devtests I did a sanity check by grepping for SEVERE in the server log. I saw this:
=================
[#|2011-05-18T12:09:43.473-0700|SEVERE|glassfish3.1|javax.enterprise.system.tools.admin.org.glassfish.config.support|_ThreadID=17;_ThreadName=Thread-1;|Exception while adding the new configuration : Constraints for this Cluster configuration have been violated: on property [ name ] violation reason [ Invalid cluster name. The name must start with a letter, number or underscore and may contain only letters, numbers, and these characters: hyphen, period, underscore, and semicolon. ]|#]
=================
What was the name of the cluster? That info must be right at the fingertips of the code that created this message. Why not include it?
#### Affected Versions
[3.1]
|
1.0
|
Add details to cluster creation failure SEVERE log message - After running thousands of devtests I did a sanity check by grepping for SEVERE in the server log. I saw this:
=================
[#|2011-05-18T12:09:43.473-0700|SEVERE|glassfish3.1|javax.enterprise.system.tools.admin.org.glassfish.config.support|_ThreadID=17;_ThreadName=Thread-1;|Exception while adding the new configuration : Constraints for this Cluster configuration have been violated: on property [ name ] violation reason [ Invalid cluster name. The name must start with a letter, number or underscore and may contain only letters, numbers, and these characters: hyphen, period, underscore, and semicolon. ]|#]
=================
What was the name of the cluster? That info must be right at the fingertips of the code that created this message. Why not include it?
#### Affected Versions
[3.1]
|
non_defect
|
add details to cluster creation failure severe log message after running thousands of devtests i did a sanity check by grepping for severe in the server log i saw this violation reason what was the name of the cluster that info must be right at the fingertips of the code that created this message why not include it affected versions
| 0
|
104,030
| 4,188,666,443
|
IssuesEvent
|
2016-06-23 21:30:37
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
Create the required namespaces in all the clusters in the federation e2e tests
|
area/cluster-federation priority/P1
|
Spinning off from #27909
cc @kubernetes/sig-cluster-federation @matchstick
|
1.0
|
Create the required namespaces in all the clusters in the federation e2e tests - Spinning off from #27909
cc @kubernetes/sig-cluster-federation @matchstick
|
non_defect
|
create the required namespaces in all the clusters in the federation tests spinning off from cc kubernetes sig cluster federation matchstick
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.