Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
5,635 | 2,610,192,348 | IssuesEvent | 2015-02-26 19:00:46 | chrsmith/quchuseban | https://api.github.com/repos/chrsmith/quchuseban | opened | 分享色斑的治疗偏方有哪些 | auto-migrated Priority-Medium Type-Defect | ```
《摘要》
我不知道它从哪里来,要到哪里去,也不知道它因何而来,��
�何而去。时光的隧道里,它存在了千年,亦穿越了千年。期�
��的冷暖,个中的磨难,无论是深重,还是沧桑,都由它独自
承担。宛若前世预约的邂逅,只消一个照面,便刻在心尖,��
�此相互缠绕,相依相偎,彼此的世界都不再孤单。风起的日�
��,静静聆听它的浅吟低唱,风落的日子,默默细数风过的划
痕。任微风撩起岁月的裙幔,趟过飘香的温暖,漫过飞雪的��
�寒,直至生命烟消云散。但是岁月的脚步,真的无法更改吗�
��脸上的雀斑真的无法去除吗!色斑的治疗偏方有哪些,
《客户案例》
刘女士 26岁<br>
做妈妈很伟大,但是不能否认处于孕期的女人是最丑的��
�不但全身臃肿走路不利索,而且皮肤会变差,还容易长斑,�
��就是最佳一例。怀孕三四个月后,我脸上就长出来很多斑,
每天照镜子的时候,我都忍不住去抠那些斑,可是皮肤都抠��
�血来了,斑还是顽固的长在那里,就是下不去。迷信的公公�
��婆说生男孩脸上就长斑,生女孩就不长,老公也说脸上长斑
不要紧,可是我看书上说很多人在孕期时候长斑,孩子出生��
�斑也不能去掉,心里比较焦急。听人说橄榄油可以祛斑,又�
��心对胎儿有影响,就一直没敢尝试,一直到生完孩子好几个
月了才开始到处寻找祛斑产品,可是又怕把皮肤搞坏所以也��
�敢轻易尝试,直到遇到了「黛芙薇尔精华液」。因为「黛芙�
��尔精华液」是植物提取的,性质很温和。<br>
使用之前跟专家交流过,专家说「黛芙薇尔精华液」的��
�用机理是是运用舌下黏膜毛细血管最为丰富的原理,使人生�
��素通过舌下黏膜进入血液循环,迅速刺激和活化脑垂体分泌
HGH。,从而达到祛斑的效果。我听了觉得应该对身体没有伤��
�,就开始使用产品。第一个星期的时候没什么反应,脸上的�
��一点变化都没有,我着急起来了,虽然着急但是没有放弃使
用产品,这样坚持到第九天的时候,我发现脸颊旁的色斑颜��
�变淡了不少,而且脸上没那么干燥了,摸上去滑滑的,感觉�
��湿润,这个发现让我很开心,我使用产品积极多了。一个周
期的产品下来,我的身体免疫力得到了很大提高,以前经常��
�感冒,使用「黛芙薇尔精华液」的三个月里都没有再感冒,�
��喷嚏都少打。脸上和脖子上的斑大部分都消失了,下巴下面
还有一些没有消退,但是颜色已经变得很淡,不仔细看看不��
�来。
后来我又使用了一个周期的巩固效果,现在斑点已经完全去��
�了,「黛芙薇尔精华液」祛斑专家说我吸收好,体质也还可�
��,所以祛斑的效果比较好,此外,我皮肤还变好了,也变白
了不少。老公都夸我比以前漂亮了,我好高兴,感谢「黛芙��
�尔精华液」让我做没有斑点的漂亮妈妈。
阅读了色斑的治疗偏方有哪些,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
色斑的治疗偏方有哪些,同时为您分享祛斑小方法
去斑方法,柠檬黄瓜汁:洗面后,抹上柠檬和黄瓜汁,30分钟��
�用水洗去,再涂抹护肤霜,
连续20天,有助于驱除斑点,增白皮肤。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 5:25 | 1.0 | 分享色斑的治疗偏方有哪些 - ```
《摘要》
我不知道它从哪里来,要到哪里去,也不知道它因何而来,��
�何而去。时光的隧道里,它存在了千年,亦穿越了千年。期�
��的冷暖,个中的磨难,无论是深重,还是沧桑,都由它独自
承担。宛若前世预约的邂逅,只消一个照面,便刻在心尖,��
�此相互缠绕,相依相偎,彼此的世界都不再孤单。风起的日�
��,静静聆听它的浅吟低唱,风落的日子,默默细数风过的划
痕。任微风撩起岁月的裙幔,趟过飘香的温暖,漫过飞雪的��
�寒,直至生命烟消云散。但是岁月的脚步,真的无法更改吗�
��脸上的雀斑真的无法去除吗!色斑的治疗偏方有哪些,
《客户案例》
刘女士 26岁<br>
做妈妈很伟大,但是不能否认处于孕期的女人是最丑的��
�不但全身臃肿走路不利索,而且皮肤会变差,还容易长斑,�
��就是最佳一例。怀孕三四个月后,我脸上就长出来很多斑,
每天照镜子的时候,我都忍不住去抠那些斑,可是皮肤都抠��
�血来了,斑还是顽固的长在那里,就是下不去。迷信的公公�
��婆说生男孩脸上就长斑,生女孩就不长,老公也说脸上长斑
不要紧,可是我看书上说很多人在孕期时候长斑,孩子出生��
�斑也不能去掉,心里比较焦急。听人说橄榄油可以祛斑,又�
��心对胎儿有影响,就一直没敢尝试,一直到生完孩子好几个
月了才开始到处寻找祛斑产品,可是又怕把皮肤搞坏所以也��
�敢轻易尝试,直到遇到了「黛芙薇尔精华液」。因为「黛芙�
��尔精华液」是植物提取的,性质很温和。<br>
使用之前跟专家交流过,专家说「黛芙薇尔精华液」的��
�用机理是是运用舌下黏膜毛细血管最为丰富的原理,使人生�
��素通过舌下黏膜进入血液循环,迅速刺激和活化脑垂体分泌
HGH。,从而达到祛斑的效果。我听了觉得应该对身体没有伤��
�,就开始使用产品。第一个星期的时候没什么反应,脸上的�
��一点变化都没有,我着急起来了,虽然着急但是没有放弃使
用产品,这样坚持到第九天的时候,我发现脸颊旁的色斑颜��
�变淡了不少,而且脸上没那么干燥了,摸上去滑滑的,感觉�
��湿润,这个发现让我很开心,我使用产品积极多了。一个周
期的产品下来,我的身体免疫力得到了很大提高,以前经常��
�感冒,使用「黛芙薇尔精华液」的三个月里都没有再感冒,�
��喷嚏都少打。脸上和脖子上的斑大部分都消失了,下巴下面
还有一些没有消退,但是颜色已经变得很淡,不仔细看看不��
�来。
后来我又使用了一个周期的巩固效果,现在斑点已经完全去��
�了,「黛芙薇尔精华液」祛斑专家说我吸收好,体质也还可�
��,所以祛斑的效果比较好,此外,我皮肤还变好了,也变白
了不少。老公都夸我比以前漂亮了,我好高兴,感谢「黛芙��
�尔精华液」让我做没有斑点的漂亮妈妈。
阅读了色斑的治疗偏方有哪些,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
色斑的治疗偏方有哪些,同时为您分享祛斑小方法
去斑方法,柠檬黄瓜汁:洗面后,抹上柠檬和黄瓜汁,30分钟��
�用水洗去,再涂抹护肤霜,
连续20天,有助于驱除斑点,增白皮肤。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 5:25 | defect | 分享色斑的治疗偏方有哪些 《摘要》 我不知道它从哪里来,要到哪里去,也不知道它因何而来,�� �何而去。时光的隧道里,它存在了千年,亦穿越了千年。期� ��的冷暖,个中的磨难,无论是深重,还是沧桑,都由它独自 承担。宛若前世预约的邂逅,只消一个照面,便刻在心尖,�� �此相互缠绕,相依相偎,彼此的世界都不再孤单。风起的日� ��,静静聆听它的浅吟低唱,风落的日子,默默细数风过的划 痕。任微风撩起岁月的裙幔,趟过飘香的温暖,漫过飞雪的�� �寒,直至生命烟消云散。但是岁月的脚步,真的无法更改吗� ��脸上的雀斑真的无法去除吗!色斑的治疗偏方有哪些, 《客户案例》 刘女士 做妈妈很伟大,但是不能否认处于孕期的女人是最丑的�� �不但全身臃肿走路不利索,而且皮肤会变差,还容易长斑,� ��就是最佳一例。怀孕三四个月后,我脸上就长出来很多斑, 每天照镜子的时候,我都忍不住去抠那些斑,可是皮肤都抠�� �血来了,斑还是顽固的长在那里,就是下不去。迷信的公公� ��婆说生男孩脸上就长斑,生女孩就不长,老公也说脸上长斑 不要紧,可是我看书上说很多人在孕期时候长斑,孩子出生�� �斑也不能去掉,心里比较焦急。听人说橄榄油可以祛斑,又� ��心对胎儿有影响,就一直没敢尝试,一直到生完孩子好几个 月了才开始到处寻找祛斑产品,可是又怕把皮肤搞坏所以也�� �敢轻易尝试,直到遇到了「黛芙薇尔精华液」。因为「黛芙� ��尔精华液」是植物提取的,性质很温和。 使用之前跟专家交流过,专家说「黛芙薇尔精华液」的�� �用机理是是运用舌下黏膜毛细血管最为丰富的原理,使人生� ��素通过舌下黏膜进入血液循环,迅速刺激和活化脑垂体分泌 hgh。,从而达到祛斑的效果。我听了觉得应该对身体没有伤�� �,就开始使用产品。第一个星期的时候没什么反应,脸上的� ��一点变化都没有,我着急起来了,虽然着急但是没有放弃使 用产品,这样坚持到第九天的时候,我发现脸颊旁的色斑颜�� �变淡了不少,而且脸上没那么干燥了,摸上去滑滑的,感觉� ��湿润,这个发现让我很开心,我使用产品积极多了。一个周 期的产品下来,我的身体免疫力得到了很大提高,以前经常�� �感冒,使用「黛芙薇尔精华液」的三个月里都没有再感冒,� ��喷嚏都少打。脸上和脖子上的斑大部分都消失了,下巴下面 还有一些没有消退,但是颜色已经变得很淡,不仔细看看不�� �来。 后来我又使用了一个周期的巩固效果,现在斑点已经完全去�� �了,「黛芙薇尔精华液」祛斑专家说我吸收好,体质也还可� ��,所以祛斑的效果比较好,此外,我皮肤还变好了,也变白 了不少。老公都夸我比以前漂亮了,我好高兴,感谢「黛芙�� �尔精华液」让我做没有斑点的漂亮妈妈。 阅读了色斑的治疗偏方有哪些,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 色斑的治疗偏方有哪些,同时为您分享祛斑小方法 去斑方法 柠檬黄瓜汁:洗面后,抹上柠檬和黄瓜汁, �� �用水洗去,再涂抹护肤霜, ,有助于驱除斑点,增白皮肤。 original issue reported on code google com by additive gmail com on jul at | 1 |
899 | 2,658,854,386 | IssuesEvent | 2015-03-18 17:42:36 | ReactiveX/RxJava | https://api.github.com/repos/ReactiveX/RxJava | closed | Travis-CI build failures | Build | Lately, both the main and PR builds fail with error 137, which is an indication the process was killed by SIGKILL and likely due to resource exhaustion (disk or memory?). I thought reverting one of my test changes will fix it (#2831) but strangely, only the PR's build succeeded, the main did not. | 1.0 | Travis-CI build failures - Lately, both the main and PR builds fail with error 137, which is an indication the process was killed by SIGKILL and likely due to resource exhaustion (disk or memory?). I thought reverting one of my test changes will fix it (#2831) but strangely, only the PR's build succeeded, the main did not. | non_defect | travis ci build failures lately both the main and pr builds fail with error which is an indication the process was killed by sigkill and likely due to resource exhaustion disk or memory i thought reverting one of my test changes will fix it but strangely only the pr s build succeeded the main did not | 0 |
296,384 | 25,547,256,166 | IssuesEvent | 2022-11-29 19:58:05 | sandialabs/Albany | https://api.github.com/repos/sandialabs/Albany | closed | Tests in new clang build on CEE do not run | Testing | See https://sems-cdash-son.sandia.gov/cdash/viewTest.php?onlyfailed&buildid=42446 . I verified that this happens when building the code from scratch using cmake rather than using the nightly cdash scripts. @jewatkins , can you please have a look? | 1.0 | Tests in new clang build on CEE do not run - See https://sems-cdash-son.sandia.gov/cdash/viewTest.php?onlyfailed&buildid=42446 . I verified that this happens when building the code from scratch using cmake rather than using the nightly cdash scripts. @jewatkins , can you please have a look? | non_defect | tests in new clang build on cee do not run see i verified that this happens when building the code from scratch using cmake rather than using the nightly cdash scripts jewatkins can you please have a look | 0 |
69,294 | 22,319,741,796 | IssuesEvent | 2022-06-14 04:31:58 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Loading on startup gets stuck on spinner | T-Defect X-Cannot-Reproduce S-Critical A-Electron O-Uncommon | ### Description
If I set Element Desktop to load on startup (using the Gnome Tweaks tool), it usually (though not always) gets up to a spinning loading icon and just stays there. Closing or quitting the app than relaunching does not change the behavior. If I kill all element process than relaunch the app, then it loads normally. If it is not set to load on startup, I have not encountered the issue.
Though I don't have a solid timeline, this has been happening for several months now.
### Steps to reproduce
- Have Element start on Startup
It should finish loading and bring up the full application
Logs being sent: yes

### Version information
- **Platform**: Desktop
- **OS**: Fedora 33 (Gnome 3.38.4 with Wayland)
- **Version**: 1.7.21
| 1.0 | Loading on startup gets stuck on spinner - ### Description
If I set Element Desktop to load on startup (using the Gnome Tweaks tool), it usually (though not always) gets up to a spinning loading icon and just stays there. Closing or quitting the app than relaunching does not change the behavior. If I kill all element process than relaunch the app, then it loads normally. If it is not set to load on startup, I have not encountered the issue.
Though I don't have a solid timeline, this has been happening for several months now.
### Steps to reproduce
- Have Element start on Startup
It should finish loading and bring up the full application
Logs being sent: yes

### Version information
- **Platform**: Desktop
- **OS**: Fedora 33 (Gnome 3.38.4 with Wayland)
- **Version**: 1.7.21
| defect | loading on startup gets stuck on spinner description if i set element desktop to load on startup using the gnome tweaks tool it usually though not always gets up to a spinning loading icon and just stays there closing or quitting the app than relaunching does not change the behavior if i kill all element process than relaunch the app then it loads normally if it is not set to load on startup i have not encountered the issue though i don t have a solid timeline this has been happening for several months now steps to reproduce have element start on startup it should finish loading and bring up the full application logs being sent yes version information platform desktop os fedora gnome with wayland version | 1 |
303,840 | 26,232,156,560 | IssuesEvent | 2023-01-05 01:52:25 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | acceptance: TestDockerNodeJS failed | C-test-failure O-robot branch-master | [(acceptance).TestDockerNodeJS failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661095&tab=buildLog) on [master@355fe0b831bb9b682ac1200428b20a0bdfe03065](https://github.com/cockroachdb/cockroach/commits/355fe0b831bb9b682ac1200428b20a0bdfe03065):
```
{"foo": 123}
✓ can be selected directly
✓ can be inserted into a table and then retrieved
JSONB
✓ gives the right error code on invalid JSON
round-tripping a value
123
✓ can be selected directly
✓ can be inserted into a table and then retrieved
"hello"
✓ can be selected directly
✓ can be inserted into a table and then retrieved
{}
✓ can be selected directly
✓ can be inserted into a table and then retrieved
[]
✓ can be selected directly
✓ can be inserted into a table and then retrieved
0
✓ can be selected directly
✓ can be inserted into a table and then retrieved
0.0000
✓ can be selected directly
✓ can be inserted into a table and then retrieved
""
✓ can be selected directly
✓ can be inserted into a table and then retrieved
"🚀"
✓ can be selected directly
✓ can be inserted into a table and then retrieved
{"🚀": "hello"}
✓ can be selected directly
✓ can be inserted into a table and then retrieved
[1, 2, 3]
✓ can be selected directly
✓ can be inserted into a table and then retrieved
{"foo": 123}
✓ can be selected directly
✓ can be inserted into a table and then retrieved
sequelize
✓ can create a model with a json field (129ms)
✓ can create a model with an inverted index (238ms)
56 passing (1s)
dockercluster.go:687: unexpected extra event &{0 die} (after [])
--- FAIL: TestDockerNodeJS/runMode=docker (7.58s)
```
<details><summary>More</summary><p>
Parameters:
- GOFLAGS=-json
```
make stressrace TESTS=TestDockerNodeJS PKG=./pkg/acceptance TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
Related:
- #58353 acceptance: TestDockerNodeJS failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.2)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestDockerNodeJS.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
Jira issue: CRDB-3193 | 1.0 | acceptance: TestDockerNodeJS failed - [(acceptance).TestDockerNodeJS failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661095&tab=buildLog) on [master@355fe0b831bb9b682ac1200428b20a0bdfe03065](https://github.com/cockroachdb/cockroach/commits/355fe0b831bb9b682ac1200428b20a0bdfe03065):
```
{"foo": 123}
✓ can be selected directly
✓ can be inserted into a table and then retrieved
JSONB
✓ gives the right error code on invalid JSON
round-tripping a value
123
✓ can be selected directly
✓ can be inserted into a table and then retrieved
"hello"
✓ can be selected directly
✓ can be inserted into a table and then retrieved
{}
✓ can be selected directly
✓ can be inserted into a table and then retrieved
[]
✓ can be selected directly
✓ can be inserted into a table and then retrieved
0
✓ can be selected directly
✓ can be inserted into a table and then retrieved
0.0000
✓ can be selected directly
✓ can be inserted into a table and then retrieved
""
✓ can be selected directly
✓ can be inserted into a table and then retrieved
"🚀"
✓ can be selected directly
✓ can be inserted into a table and then retrieved
{"🚀": "hello"}
✓ can be selected directly
✓ can be inserted into a table and then retrieved
[1, 2, 3]
✓ can be selected directly
✓ can be inserted into a table and then retrieved
{"foo": 123}
✓ can be selected directly
✓ can be inserted into a table and then retrieved
sequelize
✓ can create a model with a json field (129ms)
✓ can create a model with an inverted index (238ms)
56 passing (1s)
dockercluster.go:687: unexpected extra event &{0 die} (after [])
--- FAIL: TestDockerNodeJS/runMode=docker (7.58s)
```
<details><summary>More</summary><p>
Parameters:
- GOFLAGS=-json
```
make stressrace TESTS=TestDockerNodeJS PKG=./pkg/acceptance TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
Related:
- #58353 acceptance: TestDockerNodeJS failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.2)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestDockerNodeJS.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
Jira issue: CRDB-3193 | non_defect | acceptance testdockernodejs failed on foo ✓ can be selected directly ✓ can be inserted into a table and then retrieved jsonb ✓ gives the right error code on invalid json round tripping a value ✓ can be selected directly ✓ can be inserted into a table and then retrieved hello ✓ can be selected directly ✓ can be inserted into a table and then retrieved ✓ can be selected directly ✓ can be inserted into a table and then retrieved ✓ can be selected directly ✓ can be inserted into a table and then retrieved ✓ can be selected directly ✓ can be inserted into a table and then retrieved ✓ can be selected directly ✓ can be inserted into a table and then retrieved ✓ can be selected directly ✓ can be inserted into a table and then retrieved 🚀 ✓ can be selected directly ✓ can be inserted into a table and then retrieved 🚀 hello ✓ can be selected directly ✓ can be inserted into a table and then retrieved ✓ can be selected directly ✓ can be inserted into a table and then retrieved foo ✓ can be selected directly ✓ can be inserted into a table and then retrieved sequelize ✓ can create a model with a json field ✓ can create a model with an inverted index passing dockercluster go unexpected extra event die after fail testdockernodejs runmode docker more parameters goflags json make stressrace tests testdockernodejs pkg pkg acceptance testtimeout stressflags timeout related acceptance testdockernodejs failed powered by jira issue crdb | 0 |
16,430 | 4,052,460,914 | IssuesEvent | 2016-05-24 02:44:26 | boostorg/compute | https://api.github.com/repos/boostorg/compute | closed | Update documentation before next Boost release | documentation | Now that Boost.Compute is an official Boost library and will be included in Boost 1.61, we need to update the documentation to reflect this.
Currently there are a few locations where we describe Boost.Compute as ["not yet an offical Boost library"](http://www.boost.org/doc/libs/master/libs/compute/doc/html/boost_compute/getting_started.html).
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/31578143-update-documentation-before-next-boost-release?utm_campaign=plugin&utm_content=tracker%2F402515&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F402515&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | Update documentation before next Boost release - Now that Boost.Compute is an official Boost library and will be included in Boost 1.61, we need to update the documentation to reflect this.
Currently there are a few locations where we describe Boost.Compute as ["not yet an offical Boost library"](http://www.boost.org/doc/libs/master/libs/compute/doc/html/boost_compute/getting_started.html).
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/31578143-update-documentation-before-next-boost-release?utm_campaign=plugin&utm_content=tracker%2F402515&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F402515&utm_medium=issues&utm_source=github).
</bountysource-plugin> | non_defect | update documentation before next boost release now that boost compute is an official boost library and will be included in boost we need to update the documentation to reflect this currently there are a few locations where we describe boost compute as want to back this issue we accept bounties via | 0 |
6,880 | 2,610,301,407 | IssuesEvent | 2015-02-26 19:36:53 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | Recaptcha | auto-migrated Priority-High Type-Defect | ```
Recaptcha have changed there request thing, see
https://groups.google.com/forum/#!topic/recaptcha/V7qswqBnA1o
Signup is therefore impossible atm.
```
-----
Original issue reported on code.google.com by `eugene....@gmail.com` on 29 Apr 2011 at 4:51 | 1.0 | Recaptcha - ```
Recaptcha have changed there request thing, see
https://groups.google.com/forum/#!topic/recaptcha/V7qswqBnA1o
Signup is therefore impossible atm.
```
-----
Original issue reported on code.google.com by `eugene....@gmail.com` on 29 Apr 2011 at 4:51 | defect | recaptcha recaptcha have changed there request thing see signup is therefore impossible atm original issue reported on code google com by eugene gmail com on apr at | 1 |
35,892 | 7,822,176,718 | IssuesEvent | 2018-06-14 00:49:48 | StrikeNP/trac_test | https://api.github.com/repos/StrikeNP/trac_test | closed | TRAC appears to convert character sets incorrectly (Trac #6) | Migrated from Trac defect senkbeil@uwm.edu utilities | Some of our source files (e.g. microphys_driver.F90) contain some non-ascii characters. According to the TRAC website it uses Unicode for character encoding, so I think these should display properly, but they do not.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/6
```json
{
"status": "closed",
"changetime": "2009-05-13T18:15:45",
"description": "Some of our source files (e.g. microphys_driver.F90) contain some non-ascii characters. According to the TRAC website it uses Unicode for character encoding, so I think these should display properly, but they do not.",
"reporter": "dschanen@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1242238545000000",
"component": "utilities",
"summary": "TRAC appears to convert character sets incorrectly",
"priority": "minor",
"keywords": "",
"time": "2009-05-04T20:44:26",
"milestone": "",
"owner": "senkbeil@uwm.edu",
"type": "defect"
}
```
| 1.0 | TRAC appears to convert character sets incorrectly (Trac #6) - Some of our source files (e.g. microphys_driver.F90) contain some non-ascii characters. According to the TRAC website it uses Unicode for character encoding, so I think these should display properly, but they do not.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/6
```json
{
"status": "closed",
"changetime": "2009-05-13T18:15:45",
"description": "Some of our source files (e.g. microphys_driver.F90) contain some non-ascii characters. According to the TRAC website it uses Unicode for character encoding, so I think these should display properly, but they do not.",
"reporter": "dschanen@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1242238545000000",
"component": "utilities",
"summary": "TRAC appears to convert character sets incorrectly",
"priority": "minor",
"keywords": "",
"time": "2009-05-04T20:44:26",
"milestone": "",
"owner": "senkbeil@uwm.edu",
"type": "defect"
}
```
| defect | trac appears to convert character sets incorrectly trac some of our source files e g microphys driver contain some non ascii characters according to the trac website it uses unicode for character encoding so i think these should display properly but they do not attachments migrated from json status closed changetime description some of our source files e g microphys driver contain some non ascii characters according to the trac website it uses unicode for character encoding so i think these should display properly but they do not reporter dschanen uwm edu cc resolution verified by v larson ts component utilities summary trac appears to convert character sets incorrectly priority minor keywords time milestone owner senkbeil uwm edu type defect | 1 |
178,474 | 29,831,544,131 | IssuesEvent | 2023-06-18 10:36:38 | bigbluebutton/bigbluebutton | https://api.github.com/repos/bigbluebutton/bigbluebutton | closed | [2.3] 2 Upload areas? | status: accepted component: presentation module: client type: bug design: user interface | <!--PLEASE DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS.
This issue tracker is only for bbb development related issues.-->
**To Reproduce**
Steps to reproduce the behavior:
Open manage presentations
**Expected behavior**
Only one upload area.
**Actual behavior**
two fields with different text but the same behavior are available
**Screenshots**

BBB version (optional):
2.3 beta 3(1588)
Smartphone (please complete the following information):
Device: iPhone 12 mini
OS: iOS 14.5 beta
Browser Safari | 1.0 | [2.3] 2 Upload areas? - <!--PLEASE DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS.
This issue tracker is only for bbb development related issues.-->
**To Reproduce**
Steps to reproduce the behavior:
Open manage presentations
**Expected behavior**
Only one upload area.
**Actual behavior**
two fields with different text but the same behavior are available
**Screenshots**

BBB version (optional):
2.3 beta 3(1588)
Smartphone (please complete the following information):
Device: iPhone 12 mini
OS: iOS 14.5 beta
Browser Safari | non_defect | upload areas please do not file issues for general support questions this issue tracker is only for bbb development related issues to reproduce steps to reproduce the behavior open manage presentations expected behavior only one upload area actual behavior two fields with different text but the same behavior are available screenshots bbb version optional beta smartphone please complete the following information device iphone mini os ios beta browser safari | 0 |
19,779 | 3,254,792,971 | IssuesEvent | 2015-10-20 03:10:15 | numpy/numpy | https://api.github.com/repos/numpy/numpy | closed | Segmentation fault using ma.masked_all with Python 3.3 | component: numpy.ma Defect priority: normal | Using `ma.masked_all()` with a unicode string on linux-64 with python 3.3 and numpy 1.7.0, there is a segmentation fault shown below. This is not present in python 3.2 or python 2.7.
```
ccosmos$ ipython3
Python 3.3.0 |Anaconda 1.4.0 (64-bit)| (default, Feb 25 2013, 15:57:27)
...
In [1]: import numpy.ma as ma
In [2]: t = ma.masked_all(3, dtype=[('a', '<U3')])
In [3]: print(t)
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-3-d6c4f5caf31e> in <module>()
----> 1 print(t)
/home/aldcroft/soft/anaconda/envs/py3/lib/python3.3/site-packages/numpy/ma/core.py in __str__(self)
3538 else:
3539 rdtype = _recursive_make_descr(self.dtype, "O")
-> 3540 res = self._data.astype(rdtype)
3541 _recursive_printoption(res, m, f)
3542 else:
ValueError: character U+53000464 is not in range [U+0000; U+10ffff]
In [4]: print(repr(t))
Segmentation fault
``` | 1.0 | Segmentation fault using ma.masked_all with Python 3.3 - Using `ma.masked_all()` with a unicode string on linux-64 with python 3.3 and numpy 1.7.0, there is a segmentation fault shown below. This is not present in python 3.2 or python 2.7.
```
ccosmos$ ipython3
Python 3.3.0 |Anaconda 1.4.0 (64-bit)| (default, Feb 25 2013, 15:57:27)
...
In [1]: import numpy.ma as ma
In [2]: t = ma.masked_all(3, dtype=[('a', '<U3')])
In [3]: print(t)
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-3-d6c4f5caf31e> in <module>()
----> 1 print(t)
/home/aldcroft/soft/anaconda/envs/py3/lib/python3.3/site-packages/numpy/ma/core.py in __str__(self)
3538 else:
3539 rdtype = _recursive_make_descr(self.dtype, "O")
-> 3540 res = self._data.astype(rdtype)
3541 _recursive_printoption(res, m, f)
3542 else:
ValueError: character U+53000464 is not in range [U+0000; U+10ffff]
In [4]: print(repr(t))
Segmentation fault
``` | defect | segmentation fault using ma masked all with python using ma masked all with a unicode string on linux with python and numpy there is a segmentation fault shown below this is not present in python or python ccosmos python anaconda bit default feb in import numpy ma as ma in t ma masked all dtype in print t valueerror traceback most recent call last in print t home aldcroft soft anaconda envs lib site packages numpy ma core py in str self else rdtype recursive make descr self dtype o res self data astype rdtype recursive printoption res m f else valueerror character u is not in range in print repr t segmentation fault | 1 |
690,784 | 23,672,152,336 | IssuesEvent | 2022-08-27 14:24:51 | Gogo1951/Groupie-LFG | https://api.github.com/repos/Gogo1951/Groupie-LFG | opened | Sort Current Spec vs Other Spec | Priority - 3 Average Type - Enhancement | Once we can safely send the current spec, we'll need to update the message (and remove option to send 2nd Spec from Right Click Menu and Bulletin Board Window). We'll just send details on the current spec and current gear equipped.
/w Person "{rt3} Groupie : {Current Role} LFG! Level {Level} {Current Spec} {Class} wearing {Average Item level} average item-level gear. Other Spec is {Other Spec} ({Other Role}). {Locale}-speaking Player."
/w Person "{rt3} Groupie : Tank LFG! Level 80 Blood Death Knight wearing 193 average item-level gear. Other Spec is Frost (Melee DPS). English-speaking Player." | 1.0 | Sort Current Spec vs Other Spec - Once we can safely send the current spec, we'll need to update the message (and remove option to send 2nd Spec from Right Click Menu and Bulletin Board Window). We'll just send details on the current spec and current gear equipped.
/w Person "{rt3} Groupie : {Current Role} LFG! Level {Level} {Current Spec} {Class} wearing {Average Item level} average item-level gear. Other Spec is {Other Spec} ({Other Role}). {Locale}-speaking Player."
/w Person "{rt3} Groupie : Tank LFG! Level 80 Blood Death Knight wearing 193 average item-level gear. Other Spec is Frost (Melee DPS). English-speaking Player." | non_defect | sort current spec vs other spec once we can safely send the current spec we ll need to update the message and remove option to send spec from right click menu and bulletin board window we ll just send details on the current spec and current gear equipped w person groupie current role lfg level level current spec class wearing average item level average item level gear other spec is other spec other role locale speaking player w person groupie tank lfg level blood death knight wearing average item level gear other spec is frost melee dps english speaking player | 0 |
12,548 | 2,707,971,766 | IssuesEvent | 2015-04-08 04:17:00 | FreeRADIUS/freeradius-server | https://api.github.com/repos/FreeRADIUS/freeradius-server | opened | FreeRADIUS shouldn't use gettimeofday in the event code, should use monotonic clock | defect v3.0.x v3.1.x | Should use clock_gettime() on most POSIX systems http://pubs.opengroup.org/onlinepubs/9699919799/
or mach_absolute_time on OSX
Examples here:
https://github.com/ThomasHabets/monotonic_clock | 1.0 | FreeRADIUS shouldn't use gettimeofday in the event code, should use monotonic clock - Should use clock_gettime() on most POSIX systems http://pubs.opengroup.org/onlinepubs/9699919799/
or mach_absolute_time on OSX
Examples here:
https://github.com/ThomasHabets/monotonic_clock | defect | freeradius shouldn t use gettimeofday in the event code should use monotonic clock should use clock gettime on most posix systems or mach absolute time on osx examples here | 1 |
224,927 | 24,800,139,600 | IssuesEvent | 2022-10-24 20:53:54 | devzom/Vue-SimpleSelect | https://api.github.com/repos/devzom/Vue-SimpleSelect | closed | WS-2022-0008 (Medium) detected in node-forge-0.10.0.tgz - autoclosed | security vulnerability | ## WS-2022-0008 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.11.2.tgz (Root Library)
- selfsigned-1.10.11.tgz
- :x: **node-forge-0.10.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The forge.debug API had a potential prototype pollution issue if called with untrusted input. The API was only used for internal debug purposes in a safe way and never documented or advertised. It is suspected that uses of this API, if any exist, would likely not have used untrusted inputs in a vulnerable way.
<p>Publish Date: 2022-01-08
<p>URL: <a href=https://github.com/digitalbazaar/forge/commit/51228083550dde97701ac8e06c629a5184117562>WS-2022-0008</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-5rrq-pxf6-6jx5">https://github.com/advisories/GHSA-5rrq-pxf6-6jx5</a></p>
<p>Release Date: 2022-01-08</p>
<p>Fix Resolution (node-forge): 1.0.0</p>
<p>Direct dependency fix Resolution (webpack-dev-server): 4.7.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2022-0008 (Medium) detected in node-forge-0.10.0.tgz - autoclosed - ## WS-2022-0008 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.11.2.tgz (Root Library)
- selfsigned-1.10.11.tgz
- :x: **node-forge-0.10.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The forge.debug API had a potential prototype pollution issue if called with untrusted input. The API was only used for internal debug purposes in a safe way and never documented or advertised. It is suspected that uses of this API, if any exist, would likely not have used untrusted inputs in a vulnerable way.
<p>Publish Date: 2022-01-08
<p>URL: <a href=https://github.com/digitalbazaar/forge/commit/51228083550dde97701ac8e06c629a5184117562>WS-2022-0008</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-5rrq-pxf6-6jx5">https://github.com/advisories/GHSA-5rrq-pxf6-6jx5</a></p>
<p>Release Date: 2022-01-08</p>
<p>Fix Resolution (node-forge): 1.0.0</p>
<p>Direct dependency fix Resolution (webpack-dev-server): 4.7.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | ws medium detected in node forge tgz autoclosed ws medium severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file package json path to vulnerable library node modules node forge package json dependency hierarchy webpack dev server tgz root library selfsigned tgz x node forge tgz vulnerable library found in base branch master vulnerability details the forge debug api had a potential prototype pollution issue if called with untrusted input the api was only used for internal debug purposes in a safe way and never documented or advertised it is suspected that uses of this api if any exist would likely not have used untrusted inputs in a vulnerable way publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge direct dependency fix resolution webpack dev server step up your open source security game with mend | 0 |
440,637 | 30,753,363,338 | IssuesEvent | 2023-07-28 21:49:12 | hwchase17/langchainjs | https://api.github.com/repos/hwchase17/langchainjs | closed | Update chroma integration to note that it has node support | auto:improvement auto:documentation | Chroma's client is isomorphic, but the current sidebar implies it does not work in Node. Suggestions for how to fix this? @nfcampos Thanks!
<img width="693" alt="Screenshot 2023-07-26 at 10 36 44 PM" src="https://github.com/hwchase17/langchainjs/assets/891664/10d29873-ed7a-4a01-b51d-e38c27500fec">
| 1.0 | Update chroma integration to note that it has node support - Chroma's client is isomorphic, but the current sidebar implies it does not work in Node. Suggestions for how to fix this? @nfcampos Thanks!
<img width="693" alt="Screenshot 2023-07-26 at 10 36 44 PM" src="https://github.com/hwchase17/langchainjs/assets/891664/10d29873-ed7a-4a01-b51d-e38c27500fec">
| non_defect | update chroma integration to note that it has node support chroma s client is isomorphic but the current sidebar implies it does not work in node suggestions for how to fix this nfcampos thanks img width alt screenshot at pm src | 0 |
151,890 | 19,667,698,932 | IssuesEvent | 2022-01-11 01:24:23 | 123tarunanand/Signal-Server | https://api.github.com/repos/123tarunanand/Signal-Server | closed | WS-2021-0491 (Medium) detected in logback-classic-1.2.3.jar - autoclosed | security vulnerability | ## WS-2021-0491 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>logback-classic-1.2.3.jar</b></p></summary>
<p>logback-classic module</p>
<p>Library home page: <a href="http://logback.qos.ch">http://logback.qos.ch</a></p>
<p>Path to dependency file: Signal-Server/service/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/ch/qos/logback/logback-classic/1.2.3/logback-classic-1.2.3.jar,canner/.m2/repository/ch/qos/logback/logback-classic/1.2.3/logback-classic-1.2.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **logback-classic-1.2.3.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LOGBack before 1.2.8 is vulnerable to Remote-Code-Execution (RCE) when the write access to 'logback.xml' and JNDI lookup are enabled.
<p>Publish Date: 2021-12-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/WS-2021-0491>WS-2021-0491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>0.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2021-0491 (Medium) detected in logback-classic-1.2.3.jar - autoclosed - ## WS-2021-0491 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>logback-classic-1.2.3.jar</b></p></summary>
<p>logback-classic module</p>
<p>Library home page: <a href="http://logback.qos.ch">http://logback.qos.ch</a></p>
<p>Path to dependency file: Signal-Server/service/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/ch/qos/logback/logback-classic/1.2.3/logback-classic-1.2.3.jar,canner/.m2/repository/ch/qos/logback/logback-classic/1.2.3/logback-classic-1.2.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **logback-classic-1.2.3.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LOGBack before 1.2.8 is vulnerable to Remote-Code-Execution (RCE) when the write access to 'logback.xml' and JNDI lookup are enabled.
<p>Publish Date: 2021-12-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/WS-2021-0491>WS-2021-0491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>0.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | ws medium detected in logback classic jar autoclosed ws medium severity vulnerability vulnerable library logback classic jar logback classic module library home page a href path to dependency file signal server service pom xml path to vulnerable library canner repository ch qos logback logback classic logback classic jar canner repository ch qos logback logback classic logback classic jar dependency hierarchy x logback classic jar vulnerable library found in base branch master vulnerability details logback before is vulnerable to remote code execution rce when the write access to logback xml and jndi lookup are enabled publish date url a href cvss score details base score metrics not available step up your open source security game with whitesource | 0 |
80,477 | 30,302,252,102 | IssuesEvent | 2023-07-10 07:01:24 | gperftools/gperftools | https://api.github.com/repos/gperftools/gperftools | closed | pprof can't work on vs2010 | Type-Defect Priority-Medium Status-New | Originally reported on Google Code with ID 392
```
What steps will reproduce the problem?
1. I have a program(xxx.exe) used by vs2010, and use pprof --text xxx.exe test.txt
2. and then output info:
Total: 1.0 MB
1.0 96.3% 96.3% 1.0 96.3% sidestep::MiniDisassembler::s_ia32_opcode_map_
0.0 3.7% 100.0% 0.0 3.7% `string'
0.0 0.0% 100.0% 0.0 0.4% 0x0000000000ed1495
0.0 0.0% 100.0% 0.0 0.8% 0x0000000000ed14b1
0.0 0.0% 100.0% 0.0 1.5% 0x0000000000ed14cd
0.0 0.0% 100.0% 0.0 0.4% 0x0000000000ed14e9
0.0 0.0% 100.0% 0.0 0.7% 0x0000000000ed1505
0.0 0.0% 100.0% 1.0 96.3% 0x0000000000ed151f
0.0 0.0% 100.0% 1.0 100.0% 0x0000000000ed19fe
0.0 0.0% 100.0% 1.0 100.0% 0x0000000000ed1bce
0.0 0.0% 100.0% 1.0 100.0% BaseThreadInitThunk
0.0 0.0% 100.0% 1.0 100.0% RtlpLowFragHeapAllocFromContext
0.0 0.0% 100.0% 1.0 96.3% std::_Debug_lt
3. test.txt:
heap profile: 6: 1089248 [ 7: 1090272] @ heapprofile
1: 1048576 [ 1: 1048576] @ 0x66adff3a 0x66aae8db 0x00ed1520 0x00ed1bcf 0x00ed19ff
0x7701ed6c 0x7739377b 0x7739374e
1: 16384 [ 1: 16384] @ 0x66ae016c 0x00ed14ce 0x00ed1bcf 0x00ed19ff 0x7701ed6c
0x7739377b 0x7739374e
1: 8192 [ 1: 8192] @ 0x66ae016c 0x00ed14b2 0x00ed1bcf 0x00ed19ff 0x7701ed6c
0x7739377b 0x7739374e
1: 8000 [ 1: 8000] @ 0x66ae016c 0x00ed1506 0x00ed1bcf 0x00ed19ff 0x7701ed6c
0x7739377b 0x7739374e
1: 4096 [ 1: 4096] @ 0x66ae016c 0x00ed1496 0x00ed1bcf 0x00ed19ff 0x7701ed6c
0x7739377b 0x7739374e
1: 4000 [ 1: 4000] @ 0x66ae016c 0x00ed14ea 0x00ed1bcf 0x00ed19ff 0x7701ed6c
0x7739377b 0x7739374e
0: 0 [ 1: 1024] @ 0x66ae016c 0x00ed1462 0x00ed1bcf 0x00ed19ff 0x7701ed6c
0x7739377b 0x7739374e
MAPPED_LIBRARIES:
00ec0000-00edb000 r-xp 00000000 00:00 0 D:\test\win32tcmalloc\Debug\win32tcmalloc.exe
77330000-7746c000 r-xp 00000000 00:00 0 C:\Windows\SYSTEM32\ntdll.dll
76fd0000-770a4000 r-xp 00000000 00:00 0 C:\Windows\system32\kernel32.dll
756e0000-7572a000 r-xp 00000000 00:00 0 C:\Windows\system32\KERNELBASE.dll
66a60000-66b48000 r-xp 00000000 00:00 0 D:\test\win32tcmalloc\Debug\libtcmalloc_minimal-debug.dll
669a0000-66a57000 r-xp 00000000 00:00 0 C:\Windows\system32\MSVCP100D.dll
66820000-66993000 r-xp 00000000 00:00 0 C:\Windows\system32\MSVCR100D.dll
77480000-77485000 r-xp 00000000 00:00 0 C:\Windows\system32\PSAPI.DLL
75860000-758b7000 r-xp 00000000 00:00 0 C:\Windows\system32\SHLWAPI.dll
76b00000-76b4e000 r-xp 00000000 00:00 0 C:\Windows\system32\GDI32.dll
76f00000-76fc9000 r-xp 00000000 00:00 0 C:\Windows\system32\USER32.dll
77490000-7749a000 r-xp 00000000 00:00 0 C:\Windows\system32\LPK.dll
759d0000-75a6d000 r-xp 00000000 00:00 0 C:\Windows\system32\USP10.dll
75a70000-75b1c000 r-xp 00000000 00:00 0 C:\Windows\system32\msvcrt.dll
76ee0000-76eff000 r-xp 00000000 00:00 0 C:\Windows\system32\IMM32.DLL
77260000-7732c000 r-xp 00000000 00:00 0 C:\Windows\system32\MSCTF.dll
6d590000-6d596000 r-xp 00000000 00:00 0 C:\PROGRA~1\KASPER~1\KASPER~1.0FO\kloehk.dll
757c0000-75860000 r-xp 00000000 00:00 0 C:\Windows\system32\ADVAPI32.dll
768e0000-768f9000 r-xp 00000000 00:00 0 C:\Windows\SYSTEM32\sechost.dll
774b0000-77551000 r-xp 00000000 00:00 0 C:\Windows\system32\RPCRT4.dll
I use tcmalloc 1.9
```
Reported by `iceryeah` on 2012-01-11 08:43:37
| 1.0 | pprof can't work on vs2010 - Originally reported on Google Code with ID 392
```
What steps will reproduce the problem?
1. I have a program(xxx.exe) used by vs2010, and use pprof --text xxx.exe test.txt
2. and then output info:
Total: 1.0 MB
1.0 96.3% 96.3% 1.0 96.3% sidestep::MiniDisassembler::s_ia32_opcode_map_
0.0 3.7% 100.0% 0.0 3.7% `string'
0.0 0.0% 100.0% 0.0 0.4% 0x0000000000ed1495
0.0 0.0% 100.0% 0.0 0.8% 0x0000000000ed14b1
0.0 0.0% 100.0% 0.0 1.5% 0x0000000000ed14cd
0.0 0.0% 100.0% 0.0 0.4% 0x0000000000ed14e9
0.0 0.0% 100.0% 0.0 0.7% 0x0000000000ed1505
0.0 0.0% 100.0% 1.0 96.3% 0x0000000000ed151f
0.0 0.0% 100.0% 1.0 100.0% 0x0000000000ed19fe
0.0 0.0% 100.0% 1.0 100.0% 0x0000000000ed1bce
0.0 0.0% 100.0% 1.0 100.0% BaseThreadInitThunk
0.0 0.0% 100.0% 1.0 100.0% RtlpLowFragHeapAllocFromContext
0.0 0.0% 100.0% 1.0 96.3% std::_Debug_lt
3. test.txt:
heap profile: 6: 1089248 [ 7: 1090272] @ heapprofile
1: 1048576 [ 1: 1048576] @ 0x66adff3a 0x66aae8db 0x00ed1520 0x00ed1bcf 0x00ed19ff
0x7701ed6c 0x7739377b 0x7739374e
1: 16384 [ 1: 16384] @ 0x66ae016c 0x00ed14ce 0x00ed1bcf 0x00ed19ff 0x7701ed6c
0x7739377b 0x7739374e
1: 8192 [ 1: 8192] @ 0x66ae016c 0x00ed14b2 0x00ed1bcf 0x00ed19ff 0x7701ed6c
0x7739377b 0x7739374e
1: 8000 [ 1: 8000] @ 0x66ae016c 0x00ed1506 0x00ed1bcf 0x00ed19ff 0x7701ed6c
0x7739377b 0x7739374e
1: 4096 [ 1: 4096] @ 0x66ae016c 0x00ed1496 0x00ed1bcf 0x00ed19ff 0x7701ed6c
0x7739377b 0x7739374e
1: 4000 [ 1: 4000] @ 0x66ae016c 0x00ed14ea 0x00ed1bcf 0x00ed19ff 0x7701ed6c
0x7739377b 0x7739374e
0: 0 [ 1: 1024] @ 0x66ae016c 0x00ed1462 0x00ed1bcf 0x00ed19ff 0x7701ed6c
0x7739377b 0x7739374e
MAPPED_LIBRARIES:
00ec0000-00edb000 r-xp 00000000 00:00 0 D:\test\win32tcmalloc\Debug\win32tcmalloc.exe
77330000-7746c000 r-xp 00000000 00:00 0 C:\Windows\SYSTEM32\ntdll.dll
76fd0000-770a4000 r-xp 00000000 00:00 0 C:\Windows\system32\kernel32.dll
756e0000-7572a000 r-xp 00000000 00:00 0 C:\Windows\system32\KERNELBASE.dll
66a60000-66b48000 r-xp 00000000 00:00 0 D:\test\win32tcmalloc\Debug\libtcmalloc_minimal-debug.dll
669a0000-66a57000 r-xp 00000000 00:00 0 C:\Windows\system32\MSVCP100D.dll
66820000-66993000 r-xp 00000000 00:00 0 C:\Windows\system32\MSVCR100D.dll
77480000-77485000 r-xp 00000000 00:00 0 C:\Windows\system32\PSAPI.DLL
75860000-758b7000 r-xp 00000000 00:00 0 C:\Windows\system32\SHLWAPI.dll
76b00000-76b4e000 r-xp 00000000 00:00 0 C:\Windows\system32\GDI32.dll
76f00000-76fc9000 r-xp 00000000 00:00 0 C:\Windows\system32\USER32.dll
77490000-7749a000 r-xp 00000000 00:00 0 C:\Windows\system32\LPK.dll
759d0000-75a6d000 r-xp 00000000 00:00 0 C:\Windows\system32\USP10.dll
75a70000-75b1c000 r-xp 00000000 00:00 0 C:\Windows\system32\msvcrt.dll
76ee0000-76eff000 r-xp 00000000 00:00 0 C:\Windows\system32\IMM32.DLL
77260000-7732c000 r-xp 00000000 00:00 0 C:\Windows\system32\MSCTF.dll
6d590000-6d596000 r-xp 00000000 00:00 0 C:\PROGRA~1\KASPER~1\KASPER~1.0FO\kloehk.dll
757c0000-75860000 r-xp 00000000 00:00 0 C:\Windows\system32\ADVAPI32.dll
768e0000-768f9000 r-xp 00000000 00:00 0 C:\Windows\SYSTEM32\sechost.dll
774b0000-77551000 r-xp 00000000 00:00 0 C:\Windows\system32\RPCRT4.dll
I use tcmalloc 1.9
```
Reported by `iceryeah` on 2012-01-11 08:43:37
| defect | pprof can t work on originally reported on google code with id what steps will reproduce the problem i have a program xxx exe used by and use pprof text xxx exe test txt and then output info total mb sidestep minidisassembler s opcode map string basethreadinitthunk rtlplowfragheapallocfromcontext std debug lt test txt heap profile heapprofile mapped libraries r xp d test debug exe r xp c windows ntdll dll r xp c windows dll r xp c windows kernelbase dll r xp d test debug libtcmalloc minimal debug dll r xp c windows dll r xp c windows dll r xp c windows psapi dll r xp c windows shlwapi dll r xp c windows dll r xp c windows dll r xp c windows lpk dll r xp c windows dll r xp c windows msvcrt dll r xp c windows dll r xp c windows msctf dll r xp c progra kasper kasper kloehk dll r xp c windows dll r xp c windows sechost dll r xp c windows dll i use tcmalloc reported by iceryeah on | 1 |
32,261 | 6,750,242,812 | IssuesEvent | 2017-10-23 03:12:04 | line/armeria | https://api.github.com/repos/line/armeria | closed | NullPointerException if an exception is thrown from one way thrift method | defect | If a thrift service throws an exception on one-way method, it causes NullPointerException because ThriftFunction does not have a value if the method is one-way.
```java
@Mock
private OnewayHelloService.Iface serviceHandler = mock(OnewayHelloService.Iface.class);
@Rule
public final ServerRule server = new ServerRule() {
@Override
protected void configure(ServerBuilder sb) throws Exception {
sb.service("/thrift", THttpService.of(serviceHandler));
}
};
@Test
public void execute_oneway() throws Exception {
OnewayHelloService.Iface client = new ClientBuilder(server.uri(BINARY, "/thrift"))
.build(OnewayHelloService.Iface.class);
doThrow(new IllegalArgumentException())
.doThrow(new IllegalArgumentException())
.doNothing()
.when(onewayServiceHandler).hello(anyString());
client.hello("hello");
verify(onewayServiceHandler, times(3)).hello("hello");
}
```
```
20:52:32.367 [armeria-common-blocking-tasks-2-1] WARN c.l.a.common.util.CompletionActions - Unexpected exception from a completion action:
java.util.concurrent.CompletionException: java.lang.NullPointerException
at java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)
at java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280)
at java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:824)
at java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:797)
at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
at com.linecorp.armeria.common.DefaultRpcResponse.completeExceptionally(DefaultRpcResponse.java:75)
at com.linecorp.armeria.server.thrift.ThriftCallService.lambda$invokeSynchronously$0(ThriftCallService.java:176)
at com.linecorp.armeria.common.AbstractRequestContext.lambda$makeContextAware$1(AbstractRequestContext.java:72)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:138)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.NullPointerException: null
at com.linecorp.armeria.internal.thrift.ThriftFunction.newResult(ThriftFunction.java:219)
at com.linecorp.armeria.server.thrift.THttpService.handleException(THttpService.java:661)
at com.linecorp.armeria.server.thrift.THttpService.lambda$invoke$4(THttpService.java:595)
at com.linecorp.armeria.common.util.Functions.lambda$voidFunction$1(Functions.java:192)
at java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822)
... 10 common frames omitted
``` | 1.0 | NullPointerException if an exception is thrown from one way thrift method - If a thrift service throws an exception on one-way method, it causes NullPointerException because ThriftFunction does not have a value if the method is one-way.
```java
@Mock
private OnewayHelloService.Iface serviceHandler = mock(OnewayHelloService.Iface.class);
@Rule
public final ServerRule server = new ServerRule() {
@Override
protected void configure(ServerBuilder sb) throws Exception {
sb.service("/thrift", THttpService.of(serviceHandler));
}
};
@Test
public void execute_oneway() throws Exception {
OnewayHelloService.Iface client = new ClientBuilder(server.uri(BINARY, "/thrift"))
.build(OnewayHelloService.Iface.class);
doThrow(new IllegalArgumentException())
.doThrow(new IllegalArgumentException())
.doNothing()
.when(onewayServiceHandler).hello(anyString());
client.hello("hello");
verify(onewayServiceHandler, times(3)).hello("hello");
}
```
```
20:52:32.367 [armeria-common-blocking-tasks-2-1] WARN c.l.a.common.util.CompletionActions - Unexpected exception from a completion action:
java.util.concurrent.CompletionException: java.lang.NullPointerException
at java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)
at java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280)
at java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:824)
at java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:797)
at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
at com.linecorp.armeria.common.DefaultRpcResponse.completeExceptionally(DefaultRpcResponse.java:75)
at com.linecorp.armeria.server.thrift.ThriftCallService.lambda$invokeSynchronously$0(ThriftCallService.java:176)
at com.linecorp.armeria.common.AbstractRequestContext.lambda$makeContextAware$1(AbstractRequestContext.java:72)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:138)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.NullPointerException: null
at com.linecorp.armeria.internal.thrift.ThriftFunction.newResult(ThriftFunction.java:219)
at com.linecorp.armeria.server.thrift.THttpService.handleException(THttpService.java:661)
at com.linecorp.armeria.server.thrift.THttpService.lambda$invoke$4(THttpService.java:595)
at com.linecorp.armeria.common.util.Functions.lambda$voidFunction$1(Functions.java:192)
at java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822)
... 10 common frames omitted
``` | defect | nullpointerexception if an exception is thrown from one way thrift method if a thrift service throws an exception on one way method it causes nullpointerexception because thriftfunction does not have a value if the method is one way java mock private onewayhelloservice iface servicehandler mock onewayhelloservice iface class rule public final serverrule server new serverrule override protected void configure serverbuilder sb throws exception sb service thrift thttpservice of servicehandler test public void execute oneway throws exception onewayhelloservice iface client new clientbuilder server uri binary thrift build onewayhelloservice iface class dothrow new illegalargumentexception dothrow new illegalargumentexception donothing when onewayservicehandler hello anystring client hello hello verify onewayservicehandler times hello hello warn c l a common util completionactions unexpected exception from a completion action java util concurrent completionexception java lang nullpointerexception at java util concurrent completablefuture encodethrowable completablefuture java at java util concurrent completablefuture completethrowable completablefuture java at java util concurrent completablefuture unihandle completablefuture java at java util concurrent completablefuture unihandle tryfire completablefuture java at java util concurrent completablefuture postcomplete completablefuture java at java util concurrent completablefuture completeexceptionally completablefuture java at com linecorp armeria common defaultrpcresponse completeexceptionally defaultrpcresponse java at com linecorp armeria server thrift thriftcallservice lambda invokesynchronously thriftcallservice java at com linecorp armeria common abstractrequestcontext lambda makecontextaware abstractrequestcontext java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at io netty util concurrent defaultthreadfactory defaultrunnabledecorator run defaultthreadfactory java at java lang thread run thread java caused by java lang nullpointerexception null at com linecorp armeria internal thrift thriftfunction newresult thriftfunction java at com linecorp armeria server thrift thttpservice handleexception thttpservice java at com linecorp armeria server thrift thttpservice lambda invoke thttpservice java at com linecorp armeria common util functions lambda voidfunction functions java at java util concurrent completablefuture unihandle completablefuture java common frames omitted | 1 |
209,613 | 7,177,840,853 | IssuesEvent | 2018-01-31 14:52:58 | strapi/strapi | https://api.github.com/repos/strapi/strapi | closed | Deprecated gulp-util | os: linux priority: medium type: bug | **Node.js version**:
v8.9.4
**npm version**:
5.6.0
**Strapi version**:
v3.0.0 alpha 9
**Operating system**:
Ubuntu 16.04
## On fresh clone `npm run setup` :
```
📦 Linking strapi-helper-plugin...
npm WARN deprecated gulp-util@3.0.8: gulp-util is deprecated - replace it, following the guidelines at https://medium.com/gulpjs/gulp-util-ca3b1f9f9ac5
npm ERR! Unexpected end of input at 1:21655
npm ERR! "^6.23.0","babel-helper-define-map":"^6.23.0","babel-messages":"^6.23
npm ERR! ^
```
Well let's see https://medium.com/gulpjs/gulp-util-ca3b1f9f9ac5 :
> Help us fix the ecosystem
> Using these steps, you can help plugin authors migrate away from gulp-util.
>
> Run `npm ls gulp-util` to get the list of plugins dependent on it.
> For each dependent plugin, run `npm issues {PLUGIN NAME}` which will open their issue tracker.
> Open an issue or pull request to remove gulp-util using the following API replacements:
> * gutil.File => https://www.npmjs.com/package/vinyl
> * gutil.replaceExtension => The .extname property on Vinyl objects or https://www.npmjs.com/package/replace-ext
> * gutil.colors => https://www.npmjs.com/package/ansi-colors
> * gutil.date => https://www.npmjs.com/package/date-format
> * gutil.log => https://www.npmjs.com/package/fancy-log
> * gutil.template => https://www.npmjs.com/package/lodash.template
> * gutil.env => https://www.npmjs.com/package/minimist
> * gutil.beep => https://www.npmjs.com/package/beeper
> * gutil.noop => https://www.npmjs.com/package/through2
> * gutil.isStream => Use the .isStream() method on Vinyl objects
> * gutil.isBuffer => Use the .isBuffer() method on Vinyl objects
> * gutil.isNull => Use the .isNull() method on Vinyl objects
> * gutil.linefeed => Use the string '\n' in your code
> * gutil.combine => https://www.npmjs.com/package/multipipe
> * gutil.buffer => https://www.npmjs.com/package/list-stream
> * gutil.PluginError => https://www.npmjs.com/package/plugin-error
Okay ... I'm not going to fix this alone x)
Did you already planed to fix this or not ? | 1.0 | Deprecated gulp-util - **Node.js version**:
v8.9.4
**npm version**:
5.6.0
**Strapi version**:
v3.0.0 alpha 9
**Operating system**:
Ubuntu 16.04
## On fresh clone `npm run setup` :
```
📦 Linking strapi-helper-plugin...
npm WARN deprecated gulp-util@3.0.8: gulp-util is deprecated - replace it, following the guidelines at https://medium.com/gulpjs/gulp-util-ca3b1f9f9ac5
npm ERR! Unexpected end of input at 1:21655
npm ERR! "^6.23.0","babel-helper-define-map":"^6.23.0","babel-messages":"^6.23
npm ERR! ^
```
Well let's see https://medium.com/gulpjs/gulp-util-ca3b1f9f9ac5 :
> Help us fix the ecosystem
> Using these steps, you can help plugin authors migrate away from gulp-util.
>
> Run `npm ls gulp-util` to get the list of plugins dependent on it.
> For each dependent plugin, run `npm issues {PLUGIN NAME}` which will open their issue tracker.
> Open an issue or pull request to remove gulp-util using the following API replacements:
> * gutil.File => https://www.npmjs.com/package/vinyl
> * gutil.replaceExtension => The .extname property on Vinyl objects or https://www.npmjs.com/package/replace-ext
> * gutil.colors => https://www.npmjs.com/package/ansi-colors
> * gutil.date => https://www.npmjs.com/package/date-format
> * gutil.log => https://www.npmjs.com/package/fancy-log
> * gutil.template => https://www.npmjs.com/package/lodash.template
> * gutil.env => https://www.npmjs.com/package/minimist
> * gutil.beep => https://www.npmjs.com/package/beeper
> * gutil.noop => https://www.npmjs.com/package/through2
> * gutil.isStream => Use the .isStream() method on Vinyl objects
> * gutil.isBuffer => Use the .isBuffer() method on Vinyl objects
> * gutil.isNull => Use the .isNull() method on Vinyl objects
> * gutil.linefeed => Use the string '\n' in your code
> * gutil.combine => https://www.npmjs.com/package/multipipe
> * gutil.buffer => https://www.npmjs.com/package/list-stream
> * gutil.PluginError => https://www.npmjs.com/package/plugin-error
Okay ... I'm not going to fix this alone x)
Did you already planed to fix this or not ? | non_defect | deprecated gulp util node js version npm version strapi version alpha operating system ubuntu on fresh clone npm run setup 📦 linking strapi helper plugin npm warn deprecated gulp util gulp util is deprecated replace it following the guidelines at npm err unexpected end of input at npm err babel helper define map babel messages npm err well let s see help us fix the ecosystem using these steps you can help plugin authors migrate away from gulp util run npm ls gulp util to get the list of plugins dependent on it for each dependent plugin run npm issues plugin name which will open their issue tracker open an issue or pull request to remove gulp util using the following api replacements gutil file gutil replaceextension the extname property on vinyl objects or gutil colors gutil date gutil log gutil template gutil env gutil beep gutil noop gutil isstream use the isstream method on vinyl objects gutil isbuffer use the isbuffer method on vinyl objects gutil isnull use the isnull method on vinyl objects gutil linefeed use the string n in your code gutil combine gutil buffer gutil pluginerror okay i m not going to fix this alone x did you already planed to fix this or not | 0 |
282,476 | 8,706,872,806 | IssuesEvent | 2018-12-06 05:10:08 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | reopened | shop.lego.com - site is not usable | browser-firefox-mobile priority-normal | <!-- @browser: Firefox Mobile 64.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:64.0) Gecko/64.0 Firefox/64.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://shop.lego.com/de-DE/Schachspiel-2017-40174?cmp=afc-AffiliateDE
**Browser / Version**: Firefox Mobile 64.0
**Operating System**: Android 8.0.0
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Nothing is loading
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | shop.lego.com - site is not usable - <!-- @browser: Firefox Mobile 64.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:64.0) Gecko/64.0 Firefox/64.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://shop.lego.com/de-DE/Schachspiel-2017-40174?cmp=afc-AffiliateDE
**Browser / Version**: Firefox Mobile 64.0
**Operating System**: Android 8.0.0
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Nothing is loading
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_defect | shop lego com site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description nothing is loading steps to reproduce browser configuration none from with ❤️ | 0 |
182,919 | 14,170,385,973 | IssuesEvent | 2020-11-12 14:28:41 | rust-lang/rust | https://api.github.com/repos/rust-lang/rust | reopened | Build hang on Linux | E-hard E-needs-test I-hang P-medium T-compiler | **Problem**
I am seeing the build process hang at the last few steps when attempting to build [this particular repository](https://github.com/danieldeankon/hypermine) on Linux using either the nightly or the beta toolchain. I've seen this happen for the repository both when building locally (Linux Mint) and in Github's CI (Ubuntu). This does not happen with the stable toolchain, and I have not seen it happen with any toolchain on Windows.
I expect to see the build process finish in approximately 20 minutes like it does with the stable toolchain, but I observe a hanging build. If there is a problem with the build process, I expect to see an indication of what it is, especially if I pass a verbose flag, but I observe no further output once it gets into the hanging state---the only output I see is from steps that have already finished, which provides no insight into why it hangs.
**Steps**
Switch to nightly toolchain:
```
$ rustup default nightly
info: using existing install for 'nightly-x86_64-unknown-linux-gnu'
info: default toolchain set to 'nightly-x86_64-unknown-linux-gnu'
nightly-x86_64-unknown-linux-gnu unchanged - rustc 1.48.0-nightly (bbc677480 2020-09-18)
```
Fetch the following repository:
```
$ git clone https://github.com/danieldeankon/hypermine.git
$ cd hypermine/
$ git lfs pull
```
Attempt to build the binaries:
```
$ cargo build --release -v
```
The last thing I see before it gets into a hanging state:
```
Compiling client v0.1.0 (/home/daniel/git/testinghypermine/hypermine/client)
Running `rustc --crate-name client --edition=2018 client/src/lib.rs --error-format=json --json=diagnostic-rendered-ansi --crate-type lib --emit=dep-info,metadata,link -C opt-level=3 -C embed-bitcode=no --cfg 'feature="default"' --cfg 'feature="use-repo-assets"' -C metadata=c38d85b37e3e2e40 -C extra-filename=-c38d85b37e3e2e40 --out-dir /home/daniel/git/testinghypermine/hypermine/target/release/deps -L dependency=/home/daniel/git/testinghypermine/hypermine/target/release/deps --extern anyhow=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libanyhow-43ee26a5a4ae988b.rmeta --extern ash=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libash-1d91e46c2845798f.rmeta --extern ash_window=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libash_window-e581873b9287fab4.rmeta --extern common=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libcommon-bdb19128efefc960.rmeta --extern directories=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libdirectories-4a64043fdaf679b9.rmeta --extern downcast_rs=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libdowncast_rs-be493f8657cf1f42.rmeta --extern futures_util=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libfutures_util-79782c1686738829.rmeta --extern fxhash=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libfxhash-aa9a2aa3814ab312.rmeta --extern gltf=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libgltf-ae78e76833b667c0.rmeta --extern hdrhistogram=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libhdrhistogram-1c5607d77dc61a11.rmeta --extern hecs=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libhecs-7ceed138b109be7b.rmeta --extern lahar=/home/daniel/git/testinghypermine/hypermine/target/release/deps/liblahar-c3a4855d049ab353.rmeta --extern memoffset=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libmemoffset-33ada65033678cfe.rmeta --extern metrics=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libmetrics-79a79d3342c8701e.rmeta --extern metrics_core=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libmetrics_core-7e657b2c60f0cf49.rmeta --extern na=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libnalgebra-615d0479649eed30.rmeta --extern png=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libpng-089627b494fcd05d.rmeta --extern quinn=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libquinn-a3a9da2f9e31ef73.rmeta --extern rcgen=/home/daniel/git/testinghypermine/hypermine/target/release/deps/librcgen-be435a028b2dfce2.rmeta --extern rustls=/home/daniel/git/testinghypermine/hypermine/target/release/deps/librustls-825cbe6b34245b7a.rmeta --extern serde=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libserde-fbaa982a65688d4a.rmeta --extern server=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libserver-8fd1f97c8fb84524.rmeta --extern tokio=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libtokio-aa180fba706594fa.rmeta --extern toml=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libtoml-f7f6190f25a07c8e.rmeta --extern tracing=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libtracing-e0bf1ed6aae74c7d.rmeta --extern vk_shader_macros=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libvk_shader_macros-d2fb7777a36a4f79.rmeta --extern webpki=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libwebpki-66dacb92629212d5.rmeta --extern whoami=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libwhoami-05f81a39d2c15f6f.rmeta --extern winit=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libwinit-0172bcc7807380da.rmeta -L native=/home/daniel/git/testinghypermine/hypermine/target/release/build/ring-86d47ac3f01cd916/out`
Building [=====================================================> ] 255/257: client
```
**Notes**
I was able to reproduce this behaviour in Github CI here: https://github.com/danieldeankon/hypermine/pull/2/checks?check_run_id=1138809646
CI / test (ubuntu-latest, stable) finished in 17m 39s. CI / test (ubuntu-latest, beta) has been going for two hours and counting at time of writing, and similarly for CI / test (ubuntu-latest, nightly).
Output of `cargo version`:
cargo 1.48.0-nightly (8777a6b1e 2020-09-15)
My operating system is Linux Mint 19.2 Cinnamon.
| 1.0 | Build hang on Linux - **Problem**
I am seeing the build process hang at the last few steps when attempting to build [this particular repository](https://github.com/danieldeankon/hypermine) on Linux using either the nightly or the beta toolchain. I've seen this happen for the repository both when building locally (Linux Mint) and in Github's CI (Ubuntu). This does not happen with the stable toolchain, and I have not seen it happen with any toolchain on Windows.
I expect to see the build process finish in approximately 20 minutes like it does with the stable toolchain, but I observe a hanging build. If there is a problem with the build process, I expect to see an indication of what it is, especially if I pass a verbose flag, but I observe no further output once it gets into the hanging state---the only output I see is from steps that have already finished, which provides no insight into why it hangs.
**Steps**
Switch to nightly toolchain:
```
$ rustup default nightly
info: using existing install for 'nightly-x86_64-unknown-linux-gnu'
info: default toolchain set to 'nightly-x86_64-unknown-linux-gnu'
nightly-x86_64-unknown-linux-gnu unchanged - rustc 1.48.0-nightly (bbc677480 2020-09-18)
```
Fetch the following repository:
```
$ git clone https://github.com/danieldeankon/hypermine.git
$ cd hypermine/
$ git lfs pull
```
Attempt to build the binaries:
```
$ cargo build --release -v
```
The last thing I see before it gets into a hanging state:
```
Compiling client v0.1.0 (/home/daniel/git/testinghypermine/hypermine/client)
Running `rustc --crate-name client --edition=2018 client/src/lib.rs --error-format=json --json=diagnostic-rendered-ansi --crate-type lib --emit=dep-info,metadata,link -C opt-level=3 -C embed-bitcode=no --cfg 'feature="default"' --cfg 'feature="use-repo-assets"' -C metadata=c38d85b37e3e2e40 -C extra-filename=-c38d85b37e3e2e40 --out-dir /home/daniel/git/testinghypermine/hypermine/target/release/deps -L dependency=/home/daniel/git/testinghypermine/hypermine/target/release/deps --extern anyhow=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libanyhow-43ee26a5a4ae988b.rmeta --extern ash=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libash-1d91e46c2845798f.rmeta --extern ash_window=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libash_window-e581873b9287fab4.rmeta --extern common=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libcommon-bdb19128efefc960.rmeta --extern directories=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libdirectories-4a64043fdaf679b9.rmeta --extern downcast_rs=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libdowncast_rs-be493f8657cf1f42.rmeta --extern futures_util=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libfutures_util-79782c1686738829.rmeta --extern fxhash=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libfxhash-aa9a2aa3814ab312.rmeta --extern gltf=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libgltf-ae78e76833b667c0.rmeta --extern hdrhistogram=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libhdrhistogram-1c5607d77dc61a11.rmeta --extern hecs=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libhecs-7ceed138b109be7b.rmeta --extern lahar=/home/daniel/git/testinghypermine/hypermine/target/release/deps/liblahar-c3a4855d049ab353.rmeta --extern memoffset=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libmemoffset-33ada65033678cfe.rmeta --extern metrics=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libmetrics-79a79d3342c8701e.rmeta --extern metrics_core=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libmetrics_core-7e657b2c60f0cf49.rmeta --extern na=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libnalgebra-615d0479649eed30.rmeta --extern png=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libpng-089627b494fcd05d.rmeta --extern quinn=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libquinn-a3a9da2f9e31ef73.rmeta --extern rcgen=/home/daniel/git/testinghypermine/hypermine/target/release/deps/librcgen-be435a028b2dfce2.rmeta --extern rustls=/home/daniel/git/testinghypermine/hypermine/target/release/deps/librustls-825cbe6b34245b7a.rmeta --extern serde=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libserde-fbaa982a65688d4a.rmeta --extern server=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libserver-8fd1f97c8fb84524.rmeta --extern tokio=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libtokio-aa180fba706594fa.rmeta --extern toml=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libtoml-f7f6190f25a07c8e.rmeta --extern tracing=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libtracing-e0bf1ed6aae74c7d.rmeta --extern vk_shader_macros=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libvk_shader_macros-d2fb7777a36a4f79.rmeta --extern webpki=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libwebpki-66dacb92629212d5.rmeta --extern whoami=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libwhoami-05f81a39d2c15f6f.rmeta --extern winit=/home/daniel/git/testinghypermine/hypermine/target/release/deps/libwinit-0172bcc7807380da.rmeta -L native=/home/daniel/git/testinghypermine/hypermine/target/release/build/ring-86d47ac3f01cd916/out`
Building [=====================================================> ] 255/257: client
```
**Notes**
I was able to reproduce this behaviour in Github CI here: https://github.com/danieldeankon/hypermine/pull/2/checks?check_run_id=1138809646
CI / test (ubuntu-latest, stable) finished in 17m 39s. CI / test (ubuntu-latest, beta) has been going for two hours and counting at time of writing, and similarly for CI / test (ubuntu-latest, nightly).
Output of `cargo version`:
cargo 1.48.0-nightly (8777a6b1e 2020-09-15)
My operating system is Linux Mint 19.2 Cinnamon.
| non_defect | build hang on linux problem i am seeing the build process hang at the last few steps when attempting to build on linux using either the nightly or the beta toolchain i ve seen this happen for the repository both when building locally linux mint and in github s ci ubuntu this does not happen with the stable toolchain and i have not seen it happen with any toolchain on windows i expect to see the build process finish in approximately minutes like it does with the stable toolchain but i observe a hanging build if there is a problem with the build process i expect to see an indication of what it is especially if i pass a verbose flag but i observe no further output once it gets into the hanging state the only output i see is from steps that have already finished which provides no insight into why it hangs steps switch to nightly toolchain rustup default nightly info using existing install for nightly unknown linux gnu info default toolchain set to nightly unknown linux gnu nightly unknown linux gnu unchanged rustc nightly fetch the following repository git clone cd hypermine git lfs pull attempt to build the binaries cargo build release v the last thing i see before it gets into a hanging state compiling client home daniel git testinghypermine hypermine client running rustc crate name client edition client src lib rs error format json json diagnostic rendered ansi crate type lib emit dep info metadata link c opt level c embed bitcode no cfg feature default cfg feature use repo assets c metadata c extra filename out dir home daniel git testinghypermine hypermine target release deps l dependency home daniel git testinghypermine hypermine target release deps extern anyhow home daniel git testinghypermine hypermine target release deps libanyhow rmeta extern ash home daniel git testinghypermine hypermine target release deps libash rmeta extern ash window home daniel git testinghypermine hypermine target release deps libash window rmeta extern common home daniel git testinghypermine hypermine target release deps libcommon rmeta extern directories home daniel git testinghypermine hypermine target release deps libdirectories rmeta extern downcast rs home daniel git testinghypermine hypermine target release deps libdowncast rs rmeta extern futures util home daniel git testinghypermine hypermine target release deps libfutures util rmeta extern fxhash home daniel git testinghypermine hypermine target release deps libfxhash rmeta extern gltf home daniel git testinghypermine hypermine target release deps libgltf rmeta extern hdrhistogram home daniel git testinghypermine hypermine target release deps libhdrhistogram rmeta extern hecs home daniel git testinghypermine hypermine target release deps libhecs rmeta extern lahar home daniel git testinghypermine hypermine target release deps liblahar rmeta extern memoffset home daniel git testinghypermine hypermine target release deps libmemoffset rmeta extern metrics home daniel git testinghypermine hypermine target release deps libmetrics rmeta extern metrics core home daniel git testinghypermine hypermine target release deps libmetrics core rmeta extern na home daniel git testinghypermine hypermine target release deps libnalgebra rmeta extern png home daniel git testinghypermine hypermine target release deps libpng rmeta extern quinn home daniel git testinghypermine hypermine target release deps libquinn rmeta extern rcgen home daniel git testinghypermine hypermine target release deps librcgen rmeta extern rustls home daniel git testinghypermine hypermine target release deps librustls rmeta extern serde home daniel git testinghypermine hypermine target release deps libserde rmeta extern server home daniel git testinghypermine hypermine target release deps libserver rmeta extern tokio home daniel git testinghypermine hypermine target release deps libtokio rmeta extern toml home daniel git testinghypermine hypermine target release deps libtoml rmeta extern tracing home daniel git testinghypermine hypermine target release deps libtracing rmeta extern vk shader macros home daniel git testinghypermine hypermine target release deps libvk shader macros rmeta extern webpki home daniel git testinghypermine hypermine target release deps libwebpki rmeta extern whoami home daniel git testinghypermine hypermine target release deps libwhoami rmeta extern winit home daniel git testinghypermine hypermine target release deps libwinit rmeta l native home daniel git testinghypermine hypermine target release build ring out building client notes i was able to reproduce this behaviour in github ci here ci test ubuntu latest stable finished in ci test ubuntu latest beta has been going for two hours and counting at time of writing and similarly for ci test ubuntu latest nightly output of cargo version cargo nightly my operating system is linux mint cinnamon | 0 |
39,640 | 9,600,650,158 | IssuesEvent | 2019-05-10 09:54:29 | jOOQ/jOOR | https://api.github.com/repos/jOOQ/jOOR | opened | ClassFormatError when processors passed to Reflect.compile() produce new source files | P: Medium T: Defect | When a processor passed to `Reflect.compile()` attempts to produce new source files, we get the following error:
```
Exception in thread "main" java.lang.ClassFormatError: Incompatible magic value 1885430635 in class file test/Data
at java.base/java.lang.ClassLoader.defineClass1(Native Method)
at java.base/java.lang.ClassLoader.defineClass(ClassLoader.java:1016)
at java.base/java.lang.ClassLoader.defineClass(ClassLoader.java:877)
at org.joor.Compile$ByteArrayClassLoader.findClass(Compile.java:169)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:588)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:521)
at org.joor.Compile.lambda$compile$3(Compile.java:142)
at org.joor.Compile$ClassFileManager.loadAndReturnMainClass(Compile.java:232)
at org.joor.Compile.compile(Compile.java:141)
at org.joor.Reflect.compile(Reflect.java:102)
at org.jooq.annotation.processor.Test.main(Test.java:9)
``` | 1.0 | ClassFormatError when processors passed to Reflect.compile() produce new source files - When a processor passed to `Reflect.compile()` attempts to produce new source files, we get the following error:
```
Exception in thread "main" java.lang.ClassFormatError: Incompatible magic value 1885430635 in class file test/Data
at java.base/java.lang.ClassLoader.defineClass1(Native Method)
at java.base/java.lang.ClassLoader.defineClass(ClassLoader.java:1016)
at java.base/java.lang.ClassLoader.defineClass(ClassLoader.java:877)
at org.joor.Compile$ByteArrayClassLoader.findClass(Compile.java:169)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:588)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:521)
at org.joor.Compile.lambda$compile$3(Compile.java:142)
at org.joor.Compile$ClassFileManager.loadAndReturnMainClass(Compile.java:232)
at org.joor.Compile.compile(Compile.java:141)
at org.joor.Reflect.compile(Reflect.java:102)
at org.jooq.annotation.processor.Test.main(Test.java:9)
``` | defect | classformaterror when processors passed to reflect compile produce new source files when a processor passed to reflect compile attempts to produce new source files we get the following error exception in thread main java lang classformaterror incompatible magic value in class file test data at java base java lang classloader native method at java base java lang classloader defineclass classloader java at java base java lang classloader defineclass classloader java at org joor compile bytearrayclassloader findclass compile java at java base java lang classloader loadclass classloader java at java base java lang classloader loadclass classloader java at org joor compile lambda compile compile java at org joor compile classfilemanager loadandreturnmainclass compile java at org joor compile compile compile java at org joor reflect compile reflect java at org jooq annotation processor test main test java | 1 |
89,096 | 25,575,414,865 | IssuesEvent | 2022-11-30 21:37:26 | spack/spack | https://api.github.com/repos/spack/spack | opened | Installation issue: py-torch 1.12.1 (LSF error?) | build-error | ### Steps to reproduce the issue
```console
Input spec
--------------------------------
py-torch@1.12.1%clang@10.0.1+caffe2+cuda cuda_arch=70
Concretized
--------------------------------
py-torch@1.12.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +caffe2+cuda+cudnn~debug+distributed+fbgemm+gloo+kineto~magma~metal+mkldnn+mpi~nccl+nnpack+numa+numpy+onnx_ml+openmp+qnnpack~rocm+tensorpipe~test+valgrind+xnnpack build_system=python_pip cuda_arch=70 patches=2fac519 arch=linux-rhel7-power9le
^cmake@3.21.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~doc+ncurses+ownlibs~qt build_system=generic build_type=Release arch=linux-rhel7-power9le
^cuda@11.8.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~allow-unsupported-compilers~dev build_system=generic arch=linux-rhel7-power9le
^cudnn@8.4.0.27-11.6%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=generic arch=linux-rhel7-power9le
^eigen@3.4.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ipo build_system=cmake build_type=RelWithDebInfo arch=linux-rhel7-power9le
^fxdiv@2020-04-17%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ipo build_system=cmake build_type=RelWithDebInfo arch=linux-rhel7-power9le
^intel-mkl@2020.4.304%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ilp64+shared build_system=generic threads=none arch=linux-rhel7-power9le
^cpio@2.11%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^ninja@1.9.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=generic arch=linux-rhel7-power9le
^numactl@2.0.14%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools patches=4e1d78c,62fc8a8,ff37630 arch=linux-rhel7-power9le
^autoconf@2.69%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools patches=7793209 arch=linux-rhel7-power9le
^automake@1.13.4%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^gnuconfig@2021-08-14%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=generic arch=linux-rhel7-power9le
^libtool@2.4.7%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^m4@1.4.16%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +sigsegv build_system=autotools arch=linux-rhel7-power9le
^protobuf@3.21.7%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ipo+shared build_system=cmake build_type=Release arch=linux-rhel7-power9le
^zlib@1.2.7%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +optimize+pic+shared build_system=makefile arch=linux-rhel7-power9le
^psimd@2020-05-17%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ipo build_system=cmake build_type=RelWithDebInfo arch=linux-rhel7-power9le
^pthreadpool@2021-04-13%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ipo build_system=cmake build_type=RelWithDebInfo arch=linux-rhel7-power9le
^py-cffi@1.15.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^libffi@3.4.2%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^pkgconf@1.8.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^py-pycparser@2.21%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-future@0.18.2%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-numpy@1.23.4%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +blas+lapack build_system=python_pip patches=873745d arch=linux-rhel7-power9le
^py-cython@0.29.32%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-pip@22.2.2%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=generic arch=linux-rhel7-power9le
^py-protobuf@3.20.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~cpp build_system=python_pip arch=linux-rhel7-power9le
^py-six@1.16.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-pybind11@2.6.2%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ipo build_system=cmake build_type=RelWithDebInfo arch=linux-rhel7-power9le
^py-pyyaml@6.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +libyaml build_system=python_pip arch=linux-rhel7-power9le
^libyaml@0.2.5%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^py-setuptools@59.4.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=generic arch=linux-rhel7-power9le
^py-tqdm@4.64.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~notebook~telegram build_system=python_pip arch=linux-rhel7-power9le
^py-setuptools-scm@7.0.5%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +toml build_system=python_pip arch=linux-rhel7-power9le
^py-packaging@21.3%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-pyparsing@3.0.9%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-tomli@2.0.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-typing-extensions@4.3.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-flit-core@3.7.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-wheel@0.37.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=generic arch=linux-rhel7-power9le
^python@3.8.3%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +bz2+ctypes+dbm~debug+libxml2+lzma~nis~optimizations+pic+pyexpat+pythoncmd+readline+shared+sqlite3+ssl~tix~tkinter~ucs4+uuid+zlib build_system=generic patches=0d98e93,4c24573,f2fd060 arch=linux-rhel7-power9le
^bzip2@1.0.6%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~debug~pic+shared build_system=generic arch=linux-rhel7-power9le
^expat@2.4.8%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +libbsd build_system=autotools arch=linux-rhel7-power9le
^libbsd@0.11.5%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^libmd@1.0.4%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^gdbm@1.23%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^gettext@0.21.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +bzip2+curses+git~libunistring+libxml2+tar+xz build_system=autotools arch=linux-rhel7-power9le
^libiconv@1.16%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools libs=shared,static arch=linux-rhel7-power9le
^libxml2@2.10.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~python build_system=autotools arch=linux-rhel7-power9le
^tar@1.26%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools patches=08921fc,5c314db,d428578 zip=pigz arch=linux-rhel7-power9le
^ncurses@5.9.20130511%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~symlinks+termlib abi=5 build_system=autotools patches=f84b270 arch=linux-rhel7-power9le
^openssl@1.0.2k-fips%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~docs~shared build_system=generic certs=mozilla arch=linux-rhel7-power9le
^readline@8.1.2%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^sqlite@3.39.4%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +column_metadata~dynamic_extensions+fts+rtree build_system=autotools arch=linux-rhel7-power9le
^util-linux-uuid@2.38.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^xz@5.2.2%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~pic build_system=autotools libs=shared,static arch=linux-rhel7-power9le
^spectrum-mpi@rolling-release%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=bundle arch=linux-rhel7-ppc64le
^valgrind@3.16.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +boost+mpi+only64bit~ubsan build_system=autotools libs=shared,static arch=linux-rhel7-power9le
```
### Error message
<details><summary>Error message (short)</summary><pre>
-- Found MPI_C: /usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-rolling-release/lib/libmpiprofilesupport.so (found version "3.1")
-- Found MPI_CXX: /usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-rolling-release/lib/libmpiprofilesupport.so (found version "3.1")
-- Found MPI: TRUE (found version "3.1")
-- MPI support found
-- MPI compile flags: -pthread
-- MPI include path: /usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-2020.08.19/include/usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-rolling-release/include
-- MPI LINK flags path: -L/usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-2020.08.19/lib -pthread
-- MPI libraries: /usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-rolling-release/lib/libmpiprofilesupport.so/usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-rolling-release/lib/libmpi_ibm.so
[lassen708:107831] mca_base_component_repository_open: unable to open mca_ess_lsf: libbat.so: cannot open shared object file: No such file or directory (ignored)
[lassen708:107831] mca_base_component_repository_open: unable to open mca_plm_lsf: libbat.so: cannot open shared object file: No such file or directory (ignored)
[lassen708:107831] mca_base_component_repository_open: unable to open mca_ras_lsf: libbat.so: cannot open shared object file: No such file or directory (ignored)
-- Found OpenMPI with CUDA support built.
</pre></details>
### Information on your system
* **Spack:** 0.19.0 (a86911246a8c5571ca210ff25fc990b8eb297a5c)
* **Python:** 3.8.2
* **Platform:** linux-rhel7-power9le
* **Concretizer:** clingo
### Additional information
@adamjstewart
[torch_build_log.txt](https://github.com/spack/spack/files/10127479/torch_build_log.txt)
[torch_env.txt](https://github.com/spack/spack/files/10127480/torch_env.txt)
### General information
- [X] I have run `spack debug report` and reported the version of Spack/Python/Platform
- [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers
- [X] I have uploaded the build log and environment files
- [X] I have searched the issues of this repo and believe this is not a duplicate | 1.0 | Installation issue: py-torch 1.12.1 (LSF error?) - ### Steps to reproduce the issue
```console
Input spec
--------------------------------
py-torch@1.12.1%clang@10.0.1+caffe2+cuda cuda_arch=70
Concretized
--------------------------------
py-torch@1.12.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +caffe2+cuda+cudnn~debug+distributed+fbgemm+gloo+kineto~magma~metal+mkldnn+mpi~nccl+nnpack+numa+numpy+onnx_ml+openmp+qnnpack~rocm+tensorpipe~test+valgrind+xnnpack build_system=python_pip cuda_arch=70 patches=2fac519 arch=linux-rhel7-power9le
^cmake@3.21.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~doc+ncurses+ownlibs~qt build_system=generic build_type=Release arch=linux-rhel7-power9le
^cuda@11.8.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~allow-unsupported-compilers~dev build_system=generic arch=linux-rhel7-power9le
^cudnn@8.4.0.27-11.6%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=generic arch=linux-rhel7-power9le
^eigen@3.4.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ipo build_system=cmake build_type=RelWithDebInfo arch=linux-rhel7-power9le
^fxdiv@2020-04-17%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ipo build_system=cmake build_type=RelWithDebInfo arch=linux-rhel7-power9le
^intel-mkl@2020.4.304%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ilp64+shared build_system=generic threads=none arch=linux-rhel7-power9le
^cpio@2.11%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^ninja@1.9.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=generic arch=linux-rhel7-power9le
^numactl@2.0.14%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools patches=4e1d78c,62fc8a8,ff37630 arch=linux-rhel7-power9le
^autoconf@2.69%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools patches=7793209 arch=linux-rhel7-power9le
^automake@1.13.4%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^gnuconfig@2021-08-14%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=generic arch=linux-rhel7-power9le
^libtool@2.4.7%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^m4@1.4.16%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +sigsegv build_system=autotools arch=linux-rhel7-power9le
^protobuf@3.21.7%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ipo+shared build_system=cmake build_type=Release arch=linux-rhel7-power9le
^zlib@1.2.7%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +optimize+pic+shared build_system=makefile arch=linux-rhel7-power9le
^psimd@2020-05-17%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ipo build_system=cmake build_type=RelWithDebInfo arch=linux-rhel7-power9le
^pthreadpool@2021-04-13%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ipo build_system=cmake build_type=RelWithDebInfo arch=linux-rhel7-power9le
^py-cffi@1.15.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^libffi@3.4.2%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^pkgconf@1.8.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^py-pycparser@2.21%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-future@0.18.2%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-numpy@1.23.4%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +blas+lapack build_system=python_pip patches=873745d arch=linux-rhel7-power9le
^py-cython@0.29.32%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-pip@22.2.2%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=generic arch=linux-rhel7-power9le
^py-protobuf@3.20.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~cpp build_system=python_pip arch=linux-rhel7-power9le
^py-six@1.16.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-pybind11@2.6.2%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~ipo build_system=cmake build_type=RelWithDebInfo arch=linux-rhel7-power9le
^py-pyyaml@6.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +libyaml build_system=python_pip arch=linux-rhel7-power9le
^libyaml@0.2.5%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^py-setuptools@59.4.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=generic arch=linux-rhel7-power9le
^py-tqdm@4.64.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~notebook~telegram build_system=python_pip arch=linux-rhel7-power9le
^py-setuptools-scm@7.0.5%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +toml build_system=python_pip arch=linux-rhel7-power9le
^py-packaging@21.3%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-pyparsing@3.0.9%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-tomli@2.0.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-typing-extensions@4.3.0%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-flit-core@3.7.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=python_pip arch=linux-rhel7-power9le
^py-wheel@0.37.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=generic arch=linux-rhel7-power9le
^python@3.8.3%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +bz2+ctypes+dbm~debug+libxml2+lzma~nis~optimizations+pic+pyexpat+pythoncmd+readline+shared+sqlite3+ssl~tix~tkinter~ucs4+uuid+zlib build_system=generic patches=0d98e93,4c24573,f2fd060 arch=linux-rhel7-power9le
^bzip2@1.0.6%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~debug~pic+shared build_system=generic arch=linux-rhel7-power9le
^expat@2.4.8%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +libbsd build_system=autotools arch=linux-rhel7-power9le
^libbsd@0.11.5%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^libmd@1.0.4%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^gdbm@1.23%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^gettext@0.21.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +bzip2+curses+git~libunistring+libxml2+tar+xz build_system=autotools arch=linux-rhel7-power9le
^libiconv@1.16%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools libs=shared,static arch=linux-rhel7-power9le
^libxml2@2.10.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~python build_system=autotools arch=linux-rhel7-power9le
^tar@1.26%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools patches=08921fc,5c314db,d428578 zip=pigz arch=linux-rhel7-power9le
^ncurses@5.9.20130511%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~symlinks+termlib abi=5 build_system=autotools patches=f84b270 arch=linux-rhel7-power9le
^openssl@1.0.2k-fips%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~docs~shared build_system=generic certs=mozilla arch=linux-rhel7-power9le
^readline@8.1.2%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^sqlite@3.39.4%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +column_metadata~dynamic_extensions+fts+rtree build_system=autotools arch=linux-rhel7-power9le
^util-linux-uuid@2.38.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=autotools arch=linux-rhel7-power9le
^xz@5.2.2%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" ~pic build_system=autotools libs=shared,static arch=linux-rhel7-power9le
^spectrum-mpi@rolling-release%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" build_system=bundle arch=linux-rhel7-ppc64le
^valgrind@3.16.1%clang@10.0.1 cflags="-Wno-unused-command-line-argument" cxxflags="-Wno-unused-command-line-argument" +boost+mpi+only64bit~ubsan build_system=autotools libs=shared,static arch=linux-rhel7-power9le
```
### Error message
<details><summary>Error message (short)</summary><pre>
-- Found MPI_C: /usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-rolling-release/lib/libmpiprofilesupport.so (found version "3.1")
-- Found MPI_CXX: /usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-rolling-release/lib/libmpiprofilesupport.so (found version "3.1")
-- Found MPI: TRUE (found version "3.1")
-- MPI support found
-- MPI compile flags: -pthread
-- MPI include path: /usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-2020.08.19/include/usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-rolling-release/include
-- MPI LINK flags path: -L/usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-2020.08.19/lib -pthread
-- MPI libraries: /usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-rolling-release/lib/libmpiprofilesupport.so/usr/tce/packages/spectrum-mpi/ibm/spectrum-mpi-rolling-release/lib/libmpi_ibm.so
[lassen708:107831] mca_base_component_repository_open: unable to open mca_ess_lsf: libbat.so: cannot open shared object file: No such file or directory (ignored)
[lassen708:107831] mca_base_component_repository_open: unable to open mca_plm_lsf: libbat.so: cannot open shared object file: No such file or directory (ignored)
[lassen708:107831] mca_base_component_repository_open: unable to open mca_ras_lsf: libbat.so: cannot open shared object file: No such file or directory (ignored)
-- Found OpenMPI with CUDA support built.
</pre></details>
### Information on your system
* **Spack:** 0.19.0 (a86911246a8c5571ca210ff25fc990b8eb297a5c)
* **Python:** 3.8.2
* **Platform:** linux-rhel7-power9le
* **Concretizer:** clingo
### Additional information
@adamjstewart
[torch_build_log.txt](https://github.com/spack/spack/files/10127479/torch_build_log.txt)
[torch_env.txt](https://github.com/spack/spack/files/10127480/torch_env.txt)
### General information
- [X] I have run `spack debug report` and reported the version of Spack/Python/Platform
- [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers
- [X] I have uploaded the build log and environment files
- [X] I have searched the issues of this repo and believe this is not a duplicate | non_defect | installation issue py torch lsf error steps to reproduce the issue console input spec py torch clang cuda cuda arch concretized py torch clang cflags wno unused command line argument cxxflags wno unused command line argument cuda cudnn debug distributed fbgemm gloo kineto magma metal mkldnn mpi nccl nnpack numa numpy onnx ml openmp qnnpack rocm tensorpipe test valgrind xnnpack build system python pip cuda arch patches arch linux cmake clang cflags wno unused command line argument cxxflags wno unused command line argument doc ncurses ownlibs qt build system generic build type release arch linux cuda clang cflags wno unused command line argument cxxflags wno unused command line argument allow unsupported compilers dev build system generic arch linux cudnn clang cflags wno unused command line argument cxxflags wno unused command line argument build system generic arch linux eigen clang cflags wno unused command line argument cxxflags wno unused command line argument ipo build system cmake build type relwithdebinfo arch linux fxdiv clang cflags wno unused command line argument cxxflags wno unused command line argument ipo build system cmake build type relwithdebinfo arch linux intel mkl clang cflags wno unused command line argument cxxflags wno unused command line argument shared build system generic threads none arch linux cpio clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools arch linux ninja clang cflags wno unused command line argument cxxflags wno unused command line argument build system generic arch linux numactl clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools patches arch linux autoconf clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools patches arch linux automake clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools arch linux gnuconfig clang cflags wno unused command line argument cxxflags wno unused command line argument build system generic arch linux libtool clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools arch linux clang cflags wno unused command line argument cxxflags wno unused command line argument sigsegv build system autotools arch linux protobuf clang cflags wno unused command line argument cxxflags wno unused command line argument ipo shared build system cmake build type release arch linux zlib clang cflags wno unused command line argument cxxflags wno unused command line argument optimize pic shared build system makefile arch linux psimd clang cflags wno unused command line argument cxxflags wno unused command line argument ipo build system cmake build type relwithdebinfo arch linux pthreadpool clang cflags wno unused command line argument cxxflags wno unused command line argument ipo build system cmake build type relwithdebinfo arch linux py cffi clang cflags wno unused command line argument cxxflags wno unused command line argument build system python pip arch linux libffi clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools arch linux pkgconf clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools arch linux py pycparser clang cflags wno unused command line argument cxxflags wno unused command line argument build system python pip arch linux py future clang cflags wno unused command line argument cxxflags wno unused command line argument build system python pip arch linux py numpy clang cflags wno unused command line argument cxxflags wno unused command line argument blas lapack build system python pip patches arch linux py cython clang cflags wno unused command line argument cxxflags wno unused command line argument build system python pip arch linux py pip clang cflags wno unused command line argument cxxflags wno unused command line argument build system generic arch linux py protobuf clang cflags wno unused command line argument cxxflags wno unused command line argument cpp build system python pip arch linux py six clang cflags wno unused command line argument cxxflags wno unused command line argument build system python pip arch linux py clang cflags wno unused command line argument cxxflags wno unused command line argument ipo build system cmake build type relwithdebinfo arch linux py pyyaml clang cflags wno unused command line argument cxxflags wno unused command line argument libyaml build system python pip arch linux libyaml clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools arch linux py setuptools clang cflags wno unused command line argument cxxflags wno unused command line argument build system generic arch linux py tqdm clang cflags wno unused command line argument cxxflags wno unused command line argument notebook telegram build system python pip arch linux py setuptools scm clang cflags wno unused command line argument cxxflags wno unused command line argument toml build system python pip arch linux py packaging clang cflags wno unused command line argument cxxflags wno unused command line argument build system python pip arch linux py pyparsing clang cflags wno unused command line argument cxxflags wno unused command line argument build system python pip arch linux py tomli clang cflags wno unused command line argument cxxflags wno unused command line argument build system python pip arch linux py typing extensions clang cflags wno unused command line argument cxxflags wno unused command line argument build system python pip arch linux py flit core clang cflags wno unused command line argument cxxflags wno unused command line argument build system python pip arch linux py wheel clang cflags wno unused command line argument cxxflags wno unused command line argument build system generic arch linux python clang cflags wno unused command line argument cxxflags wno unused command line argument ctypes dbm debug lzma nis optimizations pic pyexpat pythoncmd readline shared ssl tix tkinter uuid zlib build system generic patches arch linux clang cflags wno unused command line argument cxxflags wno unused command line argument debug pic shared build system generic arch linux expat clang cflags wno unused command line argument cxxflags wno unused command line argument libbsd build system autotools arch linux libbsd clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools arch linux libmd clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools arch linux gdbm clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools arch linux gettext clang cflags wno unused command line argument cxxflags wno unused command line argument curses git libunistring tar xz build system autotools arch linux libiconv clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools libs shared static arch linux clang cflags wno unused command line argument cxxflags wno unused command line argument python build system autotools arch linux tar clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools patches zip pigz arch linux ncurses clang cflags wno unused command line argument cxxflags wno unused command line argument symlinks termlib abi build system autotools patches arch linux openssl fips clang cflags wno unused command line argument cxxflags wno unused command line argument docs shared build system generic certs mozilla arch linux readline clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools arch linux sqlite clang cflags wno unused command line argument cxxflags wno unused command line argument column metadata dynamic extensions fts rtree build system autotools arch linux util linux uuid clang cflags wno unused command line argument cxxflags wno unused command line argument build system autotools arch linux xz clang cflags wno unused command line argument cxxflags wno unused command line argument pic build system autotools libs shared static arch linux spectrum mpi rolling release clang cflags wno unused command line argument cxxflags wno unused command line argument build system bundle arch linux valgrind clang cflags wno unused command line argument cxxflags wno unused command line argument boost mpi ubsan build system autotools libs shared static arch linux error message error message short found mpi c usr tce packages spectrum mpi ibm spectrum mpi rolling release lib libmpiprofilesupport so found version found mpi cxx usr tce packages spectrum mpi ibm spectrum mpi rolling release lib libmpiprofilesupport so found version found mpi true found version mpi support found mpi compile flags pthread mpi include path usr tce packages spectrum mpi ibm spectrum mpi include usr tce packages spectrum mpi ibm spectrum mpi rolling release include mpi link flags path l usr tce packages spectrum mpi ibm spectrum mpi lib pthread mpi libraries usr tce packages spectrum mpi ibm spectrum mpi rolling release lib libmpiprofilesupport so usr tce packages spectrum mpi ibm spectrum mpi rolling release lib libmpi ibm so mca base component repository open unable to open mca ess lsf libbat so cannot open shared object file no such file or directory ignored mca base component repository open unable to open mca plm lsf libbat so cannot open shared object file no such file or directory ignored mca base component repository open unable to open mca ras lsf libbat so cannot open shared object file no such file or directory ignored found openmpi with cuda support built information on your system spack python platform linux concretizer clingo additional information adamjstewart general information i have run spack debug report and reported the version of spack python platform i have run spack maintainers and mentioned any maintainers i have uploaded the build log and environment files i have searched the issues of this repo and believe this is not a duplicate | 0 |
301,665 | 22,768,820,253 | IssuesEvent | 2022-07-08 08:02:05 | envoyproxy/gateway | https://api.github.com/repos/envoyproxy/gateway | closed | Support Default GatewayClass and Gateway Management | documentation enhancement question area/config stale | To simplify the UX (especially for dev's), the project should support managing a default GatewayClass and Gateway. For example, when EG is started with an ENV var or CLI flag such as `--default-gateway-class=true` and `--default-gateway=true` it will create the default GatewayClass and default Gateway resources.
Open Questions:
1. What namespace should be used for the default Gateway? To support the most simple use case, the `default` namespace can be used? If the `Gateway` is created in the EG system namespace, then the default Gateway needs to support cross-namespace routing as this ns is not meant for proxied apps, [xref](https://gateway-api.sigs.k8s.io/v1alpha2/guides/multiple-ns/).
2. Should EG support active management of these resources, e.g. someone deletes the default Gateway and EG will recrate it.
3. Other considerations? | 1.0 | Support Default GatewayClass and Gateway Management - To simplify the UX (especially for dev's), the project should support managing a default GatewayClass and Gateway. For example, when EG is started with an ENV var or CLI flag such as `--default-gateway-class=true` and `--default-gateway=true` it will create the default GatewayClass and default Gateway resources.
Open Questions:
1. What namespace should be used for the default Gateway? To support the most simple use case, the `default` namespace can be used? If the `Gateway` is created in the EG system namespace, then the default Gateway needs to support cross-namespace routing as this ns is not meant for proxied apps, [xref](https://gateway-api.sigs.k8s.io/v1alpha2/guides/multiple-ns/).
2. Should EG support active management of these resources, e.g. someone deletes the default Gateway and EG will recrate it.
3. Other considerations? | non_defect | support default gatewayclass and gateway management to simplify the ux especially for dev s the project should support managing a default gatewayclass and gateway for example when eg is started with an env var or cli flag such as default gateway class true and default gateway true it will create the default gatewayclass and default gateway resources open questions what namespace should be used for the default gateway to support the most simple use case the default namespace can be used if the gateway is created in the eg system namespace then the default gateway needs to support cross namespace routing as this ns is not meant for proxied apps should eg support active management of these resources e g someone deletes the default gateway and eg will recrate it other considerations | 0 |
9,789 | 13,929,691,557 | IssuesEvent | 2020-10-22 00:20:11 | OpenDRR/boundaries | https://api.github.com/repos/OpenDRR/boundaries | opened | Incorrect attributes for some fields in Geometries_SAUID.gpkg | Priority: Must Have Requirement Severity: Minor | Some of the attributes for Geometries_SAUID.gpkg including CDUID, CDNAME, CSDUID, CSDNAME, ERUID, PRUID, PRNAME are not correct for the corresponding SAUID.
For example, BC has 10033 entries with CSDUID, CSDNAME that belong in other provinces (AB, YT, NT).
| 1.0 | Incorrect attributes for some fields in Geometries_SAUID.gpkg - Some of the attributes for Geometries_SAUID.gpkg including CDUID, CDNAME, CSDUID, CSDNAME, ERUID, PRUID, PRNAME are not correct for the corresponding SAUID.
For example, BC has 10033 entries with CSDUID, CSDNAME that belong in other provinces (AB, YT, NT).
| non_defect | incorrect attributes for some fields in geometries sauid gpkg some of the attributes for geometries sauid gpkg including cduid cdname csduid csdname eruid pruid prname are not correct for the corresponding sauid for example bc has entries with csduid csdname that belong in other provinces ab yt nt | 0 |
78,195 | 27,364,571,032 | IssuesEvent | 2023-02-27 18:08:45 | dotCMS/core | https://api.github.com/repos/dotCMS/core | opened | Favorite Page: Snapshots should be displayed when the language is changed | Type : Defect Team : Scout Triage | ### Parent Issue
#22343
### Problem Statement
If the language is changed in Edit Mode, existing favorite pages are loaded with blank snapshots.
### Steps to Reproduce
1. Using full starter create some favorite pages in English
2. Open an existing page in Edit Mode and change the language to Spanish
3. Go back to the Favorite Pages view http://localhost:8080/dotAdmin/#/pages and see how the existing favorite pages are loaded. Snapshot images are broken.
https://user-images.githubusercontent.com/8741395/221646026-9dc52209-bc76-4c6f-8aba-c1f9118a7520.mov
### Acceptance Criteria
Snapshots in favorite pages view should be displayed correctly no matter the language selected in Edit Mode.
### dotCMS Version
23.03
### Proposed Objective
Please Select
### Proposed Priority
Please Select
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
_No response_
### Assumptions & Initiation Needs
_No response_
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | 1.0 | Favorite Page: Snapshots should be displayed when the language is changed - ### Parent Issue
#22343
### Problem Statement
If the language is changed in Edit Mode, existing favorite pages are loaded with blank snapshots.
### Steps to Reproduce
1. Using full starter create some favorite pages in English
2. Open an existing page in Edit Mode and change the language to Spanish
3. Go back to the Favorite Pages view http://localhost:8080/dotAdmin/#/pages and see how the existing favorite pages are loaded. Snapshot images are broken.
https://user-images.githubusercontent.com/8741395/221646026-9dc52209-bc76-4c6f-8aba-c1f9118a7520.mov
### Acceptance Criteria
Snapshots in favorite pages view should be displayed correctly no matter the language selected in Edit Mode.
### dotCMS Version
23.03
### Proposed Objective
Please Select
### Proposed Priority
Please Select
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
_No response_
### Assumptions & Initiation Needs
_No response_
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | defect | favorite page snapshots should be displayed when the language is changed parent issue problem statement if the language is changed in edit mode existing favorite pages are loaded with blank snapshots steps to reproduce using full starter create some favorite pages in english open an existing page in edit mode and change the language to spanish go back to the favorite pages view and see how the existing favorite pages are loaded snapshot images are broken acceptance criteria snapshots in favorite pages view should be displayed correctly no matter the language selected in edit mode dotcms version proposed objective please select proposed priority please select external links slack conversations support tickets figma designs etc no response assumptions initiation needs no response quality assurance notes workarounds no response sub tasks estimates no response | 1 |
264,269 | 23,105,252,922 | IssuesEvent | 2022-07-27 08:15:10 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | [CI] GeoBoundingBoxQueryBuilderGeoShapeTests testToQuery failing | :Analytics/Geo >test-failure Team:Analytics | **Build scan:**
https://gradle-enterprise.elastic.co/s/wuixlj5txavzq/tests/:x-pack:plugin:spatial:test/org.elasticsearch.xpack.spatial.index.query.GeoBoundingBoxQueryBuilderGeoShapeTests/testToQuery
**Reproduction line:**
`./gradlew ':x-pack:plugin:spatial:test' --tests "org.elasticsearch.xpack.spatial.index.query.GeoBoundingBoxQueryBuilderGeoShapeTests.testToQuery" -Dtests.seed=DF3DEC11904590F2 -Dtests.locale=sl -Dtests.timezone=America/Thule -Druntime.java=17`
**Applicable branches:**
master
**Reproduces locally?:**
Didn't try
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.spatial.index.query.GeoBoundingBoxQueryBuilderGeoShapeTests&tests.test=testToQuery
**Failure excerpt:**
```
java.lang.AssertionError: expected:<class org.apache.lucene.search.IndexOrDocValuesQuery> but was:<class org.apache.lucene.document.LatLonShapeBoundingBoxQuery>
at __randomizedtesting.SeedInfo.seed([DF3DEC11904590F2:28C6EE2FE1C65518]:0)
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:144)
at org.elasticsearch.xpack.spatial.index.query.GeoBoundingBoxQueryBuilderGeoShapeTests.doAssertLuceneQuery(GeoBoundingBoxQueryBuilderGeoShapeTests.java:63)
at org.elasticsearch.xpack.spatial.index.query.GeoBoundingBoxQueryBuilderGeoShapeTests.doAssertLuceneQuery(GeoBoundingBoxQueryBuilderGeoShapeTests.java:30)
at org.elasticsearch.test.AbstractQueryTestCase.assertLuceneQuery(AbstractQueryTestCase.java:528)
at org.elasticsearch.test.AbstractQueryTestCase.testToQuery(AbstractQueryTestCase.java:422)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
``` | 1.0 | [CI] GeoBoundingBoxQueryBuilderGeoShapeTests testToQuery failing - **Build scan:**
https://gradle-enterprise.elastic.co/s/wuixlj5txavzq/tests/:x-pack:plugin:spatial:test/org.elasticsearch.xpack.spatial.index.query.GeoBoundingBoxQueryBuilderGeoShapeTests/testToQuery
**Reproduction line:**
`./gradlew ':x-pack:plugin:spatial:test' --tests "org.elasticsearch.xpack.spatial.index.query.GeoBoundingBoxQueryBuilderGeoShapeTests.testToQuery" -Dtests.seed=DF3DEC11904590F2 -Dtests.locale=sl -Dtests.timezone=America/Thule -Druntime.java=17`
**Applicable branches:**
master
**Reproduces locally?:**
Didn't try
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.spatial.index.query.GeoBoundingBoxQueryBuilderGeoShapeTests&tests.test=testToQuery
**Failure excerpt:**
```
java.lang.AssertionError: expected:<class org.apache.lucene.search.IndexOrDocValuesQuery> but was:<class org.apache.lucene.document.LatLonShapeBoundingBoxQuery>
at __randomizedtesting.SeedInfo.seed([DF3DEC11904590F2:28C6EE2FE1C65518]:0)
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:144)
at org.elasticsearch.xpack.spatial.index.query.GeoBoundingBoxQueryBuilderGeoShapeTests.doAssertLuceneQuery(GeoBoundingBoxQueryBuilderGeoShapeTests.java:63)
at org.elasticsearch.xpack.spatial.index.query.GeoBoundingBoxQueryBuilderGeoShapeTests.doAssertLuceneQuery(GeoBoundingBoxQueryBuilderGeoShapeTests.java:30)
at org.elasticsearch.test.AbstractQueryTestCase.assertLuceneQuery(AbstractQueryTestCase.java:528)
at org.elasticsearch.test.AbstractQueryTestCase.testToQuery(AbstractQueryTestCase.java:422)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
``` | non_defect | geoboundingboxquerybuildergeoshapetests testtoquery failing build scan reproduction line gradlew x pack plugin spatial test tests org elasticsearch xpack spatial index query geoboundingboxquerybuildergeoshapetests testtoquery dtests seed dtests locale sl dtests timezone america thule druntime java applicable branches master reproduces locally didn t try failure history failure excerpt java lang assertionerror expected but was at randomizedtesting seedinfo seed at org junit assert fail assert java at org junit assert failnotequals assert java at org junit assert assertequals assert java at org junit assert assertequals assert java at org elasticsearch xpack spatial index query geoboundingboxquerybuildergeoshapetests doassertlucenequery geoboundingboxquerybuildergeoshapetests java at org elasticsearch xpack spatial index query geoboundingboxquerybuildergeoshapetests doassertlucenequery geoboundingboxquerybuildergeoshapetests java at org elasticsearch test abstractquerytestcase assertlucenequery abstractquerytestcase java at org elasticsearch test abstractquerytestcase testtoquery abstractquerytestcase java at jdk internal reflect nativemethodaccessorimpl nativemethodaccessorimpl java at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java | 0 |
25,239 | 4,241,396,924 | IssuesEvent | 2016-07-06 16:11:17 | buildo/react-components | https://api.github.com/repos/buildo/react-components | closed | [TimePicker] Make TimePicker searchable | defect in review | ## description
- recent [changes](https://github.com/buildo/react-components/commit/c68f40548c50dfd78ae5df9906f8beb158b0e72b#diff-edd5be6eed72fd2e39bb04b199d23f29R35) to dropdown made `TimePicker` no more "searchable"
## how to reproduce
- {optional: describe steps to reproduce defect}
## specs
- add prop to `TimePicker` to handle "search-ability"
## misc
{optional: other useful info}
| 1.0 | [TimePicker] Make TimePicker searchable - ## description
- recent [changes](https://github.com/buildo/react-components/commit/c68f40548c50dfd78ae5df9906f8beb158b0e72b#diff-edd5be6eed72fd2e39bb04b199d23f29R35) to dropdown made `TimePicker` no more "searchable"
## how to reproduce
- {optional: describe steps to reproduce defect}
## specs
- add prop to `TimePicker` to handle "search-ability"
## misc
{optional: other useful info}
| defect | make timepicker searchable description recent to dropdown made timepicker no more searchable how to reproduce optional describe steps to reproduce defect specs add prop to timepicker to handle search ability misc optional other useful info | 1 |
43,563 | 11,755,555,261 | IssuesEvent | 2020-03-13 09:46:06 | mozilla-lockwise/lockwise-android | https://api.github.com/repos/mozilla-lockwise/lockwise-android | opened | Crash-When wanting to use search autofill functionality through Lockwise | defect | Pre-requisites:
- Have Lockwise installed with autofill option enabled for Lockwise and with an account signed in.
## Steps to reproduce
1. Open Fenix app.
2. Go to "mobile.twitter.com".
3. Tap on "Login" button.
4. Tap on the "Email" field.
5. When the autofill pop is displayed choose option "Search Firefox Lockwise".
### Expected behavior
The app opens.
### Actual behavior
A crash is encountered.
### Device & build information
* Device: Samsung Galaxy S8+ (Android 8)
* Build version: Lockwise 4.0.0 (6087) - taken from playstore, encountered on Firefox Preview Beta 4.1.0-beta.2 (Build # 20730532)
### Notes
Attachments:

| 1.0 | Crash-When wanting to use search autofill functionality through Lockwise - Pre-requisites:
- Have Lockwise installed with autofill option enabled for Lockwise and with an account signed in.
## Steps to reproduce
1. Open Fenix app.
2. Go to "mobile.twitter.com".
3. Tap on "Login" button.
4. Tap on the "Email" field.
5. When the autofill pop is displayed choose option "Search Firefox Lockwise".
### Expected behavior
The app opens.
### Actual behavior
A crash is encountered.
### Device & build information
* Device: Samsung Galaxy S8+ (Android 8)
* Build version: Lockwise 4.0.0 (6087) - taken from playstore, encountered on Firefox Preview Beta 4.1.0-beta.2 (Build # 20730532)
### Notes
Attachments:

| defect | crash when wanting to use search autofill functionality through lockwise pre requisites have lockwise installed with autofill option enabled for lockwise and with an account signed in steps to reproduce open fenix app go to mobile twitter com tap on login button tap on the email field when the autofill pop is displayed choose option search firefox lockwise expected behavior the app opens actual behavior a crash is encountered device build information device samsung galaxy android build version lockwise taken from playstore encountered on firefox preview beta beta build notes attachments | 1 |
83,806 | 24,150,549,234 | IssuesEvent | 2022-09-21 23:55:51 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | Generated linker scripts break when ZEPHYR_BASE and ZEPHYR_MODULES share structure that contains symlinks | bug priority: low area: Build System | **Describe the bug**
My zephyr workspace consists of symlinks to read-only copies of modules which exist in a programmatically managed filesystem.
The filesystem looks something like this.
```
-/store
\-- zephyr-base
\-- hal-nxp
\-- cmsis
\-- zcbor
- /workspace
-- zephyr -> /store/zephyr-base
-- modules/hal/nxp -> /store/hal-nxp
-- modules/hal/cmsis -> /store/cmsis
```
This command builds successfully:
```
source /store/zephyr/zephyr-env.sh
cmake $ZEPHYR_BASE/samples/hello_world \
-Bbuild -DBOARD=frdm_k64f \
-DZEPHYR_MODULES=/store/hal-nxp;/store/cmsis
make -Cbuild
```
This command fails to build:
```
source /workspace/zephyr/zephyr-env.sh
cmake $ZEPHYR_BASE/samples/hello_world \
-Bbuild -DBOARD=frdm_k64f \
-DZEPHYR_MODULES=/workspace/modules/hal/nxp;/workspace/modules/hal/cmsis
make -Cbuild
```
The reason is because the generated linker scripts in `build/zephyr/include/generated/` use relative paths to $ZEPHYR_BASE.
When the build is running it thinks its running from wherever it really is in the filesystem, i.e. /store/zephyr instead of ./workspace/zephyr.
The linker generates an include that looks like
```
/* Sort key: "default" */#include "../../../modules/hal/nxp/mcux/quick_access_code.ld"
```
But starting from /store/zephyr/ there is no ../../../modules. If any of the "../" are symlinks at configure time the build will fail.
If these files would embed absolute paths instead the symlinks would be resolved correctly at configure and build time.
Is there any reason to prefer relative paths over absolute paths? | 1.0 | Generated linker scripts break when ZEPHYR_BASE and ZEPHYR_MODULES share structure that contains symlinks - **Describe the bug**
My zephyr workspace consists of symlinks to read-only copies of modules which exist in a programmatically managed filesystem.
The filesystem looks something like this.
```
-/store
\-- zephyr-base
\-- hal-nxp
\-- cmsis
\-- zcbor
- /workspace
-- zephyr -> /store/zephyr-base
-- modules/hal/nxp -> /store/hal-nxp
-- modules/hal/cmsis -> /store/cmsis
```
This command builds successfully:
```
source /store/zephyr/zephyr-env.sh
cmake $ZEPHYR_BASE/samples/hello_world \
-Bbuild -DBOARD=frdm_k64f \
-DZEPHYR_MODULES=/store/hal-nxp;/store/cmsis
make -Cbuild
```
This command fails to build:
```
source /workspace/zephyr/zephyr-env.sh
cmake $ZEPHYR_BASE/samples/hello_world \
-Bbuild -DBOARD=frdm_k64f \
-DZEPHYR_MODULES=/workspace/modules/hal/nxp;/workspace/modules/hal/cmsis
make -Cbuild
```
The reason is because the generated linker scripts in `build/zephyr/include/generated/` use relative paths to $ZEPHYR_BASE.
When the build is running it thinks its running from wherever it really is in the filesystem, i.e. /store/zephyr instead of ./workspace/zephyr.
The linker generates an include that looks like
```
/* Sort key: "default" */#include "../../../modules/hal/nxp/mcux/quick_access_code.ld"
```
But starting from /store/zephyr/ there is no ../../../modules. If any of the "../" are symlinks at configure time the build will fail.
If these files would embed absolute paths instead the symlinks would be resolved correctly at configure and build time.
Is there any reason to prefer relative paths over absolute paths? | non_defect | generated linker scripts break when zephyr base and zephyr modules share structure that contains symlinks describe the bug my zephyr workspace consists of symlinks to read only copies of modules which exist in a programmatically managed filesystem the filesystem looks something like this store zephyr base hal nxp cmsis zcbor workspace zephyr store zephyr base modules hal nxp store hal nxp modules hal cmsis store cmsis this command builds successfully source store zephyr zephyr env sh cmake zephyr base samples hello world bbuild dboard frdm dzephyr modules store hal nxp store cmsis make cbuild this command fails to build source workspace zephyr zephyr env sh cmake zephyr base samples hello world bbuild dboard frdm dzephyr modules workspace modules hal nxp workspace modules hal cmsis make cbuild the reason is because the generated linker scripts in build zephyr include generated use relative paths to zephyr base when the build is running it thinks its running from wherever it really is in the filesystem i e store zephyr instead of workspace zephyr the linker generates an include that looks like sort key default include modules hal nxp mcux quick access code ld but starting from store zephyr there is no modules if any of the are symlinks at configure time the build will fail if these files would embed absolute paths instead the symlinks would be resolved correctly at configure and build time is there any reason to prefer relative paths over absolute paths | 0 |
82,055 | 31,884,420,923 | IssuesEvent | 2023-09-16 19:24:02 | zuzukin/whl2conda | https://api.github.com/repos/zuzukin/whl2conda | closed | whl2conda convert does not generate correct code for entry points | defect component: convert | The `link.json` file in the conda `info/` directory is generated incorrectly
for whl2conda <= 23.8.9. The entry was being generated in the top-level
of the json file instead of inside the `noarch` section.
| 1.0 | whl2conda convert does not generate correct code for entry points - The `link.json` file in the conda `info/` directory is generated incorrectly
for whl2conda <= 23.8.9. The entry was being generated in the top-level
of the json file instead of inside the `noarch` section.
| defect | convert does not generate correct code for entry points the link json file in the conda info directory is generated incorrectly for the entry was being generated in the top level of the json file instead of inside the noarch section | 1 |
76,238 | 26,327,495,468 | IssuesEvent | 2023-01-10 08:11:11 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | room not accessible | T-Defect X-Needs-Info | ### Steps to reproduce
Trying to add the following room
[#facebook:maunium.net](https://matrix.to/#/#facebook:maunium.net)
### Outcome
got this error
#facebook:maunium.net is not accessible at this time.
Try again later, or ask a room or space admin to check if you have access.
was returned while trying to access the room or space. If you think you're seeing this message in error, please [submit a bug report](https://github.com/vector-im/element-web/issues/new/choose).
### Operating system
MacOS
### Application version
Element v1.11.17
### How did you install the app?
element website, cant remember was ages ago
### Homeserver
matrix.optimusgreen.com
### Will you send logs?
Yes | 1.0 | room not accessible - ### Steps to reproduce
Trying to add the following room
[#facebook:maunium.net](https://matrix.to/#/#facebook:maunium.net)
### Outcome
got this error
#facebook:maunium.net is not accessible at this time.
Try again later, or ask a room or space admin to check if you have access.
was returned while trying to access the room or space. If you think you're seeing this message in error, please [submit a bug report](https://github.com/vector-im/element-web/issues/new/choose).
### Operating system
MacOS
### Application version
Element v1.11.17
### How did you install the app?
element website, cant remember was ages ago
### Homeserver
matrix.optimusgreen.com
### Will you send logs?
Yes | defect | room not accessible steps to reproduce trying to add the following room outcome got this error facebook maunium net is not accessible at this time try again later or ask a room or space admin to check if you have access was returned while trying to access the room or space if you think you re seeing this message in error please operating system macos application version element how did you install the app element website cant remember was ages ago homeserver matrix optimusgreen com will you send logs yes | 1 |
45,846 | 13,055,755,838 | IssuesEvent | 2020-07-30 02:38:26 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | pyroot plays nice with numpy (Trac #126) | Incomplete Migration Migrated from Trac defect documentation | Migrated from https://code.icecube.wisc.edu/ticket/126
```json
{
"status": "closed",
"changetime": "2012-10-31T18:49:16",
"description": "recently discovered. very easy to make root histograms\nfrom numpy data that has already had cuts/etc applied.\nneed examples/docs.",
"reporter": "troy",
"cc": "",
"resolution": "wont or cant fix",
"_ts": "1351709356000000",
"component": "documentation",
"summary": "pyroot plays nice with numpy",
"priority": "normal",
"keywords": "",
"time": "2008-09-07T13:40:39",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| 1.0 | pyroot plays nice with numpy (Trac #126) - Migrated from https://code.icecube.wisc.edu/ticket/126
```json
{
"status": "closed",
"changetime": "2012-10-31T18:49:16",
"description": "recently discovered. very easy to make root histograms\nfrom numpy data that has already had cuts/etc applied.\nneed examples/docs.",
"reporter": "troy",
"cc": "",
"resolution": "wont or cant fix",
"_ts": "1351709356000000",
"component": "documentation",
"summary": "pyroot plays nice with numpy",
"priority": "normal",
"keywords": "",
"time": "2008-09-07T13:40:39",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| defect | pyroot plays nice with numpy trac migrated from json status closed changetime description recently discovered very easy to make root histograms nfrom numpy data that has already had cuts etc applied nneed examples docs reporter troy cc resolution wont or cant fix ts component documentation summary pyroot plays nice with numpy priority normal keywords time milestone owner troy type defect | 1 |
33,175 | 7,046,861,148 | IssuesEvent | 2018-01-02 10:23:22 | line/armeria | https://api.github.com/repos/line/armeria | closed | java.util.NoSuchElementException: No value present when using ZooKeeperUpdatingListener | defect | Hi, I am new to armeria. I build a simple project using armeria 0.5.60 according to [armeria documentation](http://line.github.io/armeria/index.html), everything goes well until using 'Service discovery with ZooKeeper'. I got this exception
```
java.util.NoSuchElementException: No value present
at java.util.Optional.get(Optional.java:135)
at java.util.concurrent.CompletableFuture.reportJoin(CompletableFuture.java:375)
at com.linecorp.armeria.server.zookeeper.ZooKeeperUpdatingListener.serverStarting(ZooKeeperUpdatingListener.java:69)
at java.util.concurrent.CompletableFuture.join(CompletableFuture.java:1934)
at com.linecorp.armeria.server.Server$StateManager.notifyState(Server.java:585)
at com.github.liuzhengyang.server.Bootstrap.main(Bootstrap.java:38)
at com.linecorp.armeria.server.Server$StateManager.enterStarting(Server.java:541)
Caused by: java.lang.IllegalStateException: failed to notify all server listeners
at com.linecorp.armeria.server.Server.start(Server.java:250)
at com.linecorp.armeria.server.Server$StateManager.enterStarting(Server.java:542)
at com.linecorp.armeria.server.Server.start(Server.java:235)
at com.linecorp.armeria.server.Server.start(Server.java:250)
at com.github.liuzhengyang.server.Bootstrap.main(Bootstrap.java:38)
at com.linecorp.armeria.server.Server.start(Server.java:235)
... 1 more
```
And here is my example code.
```
public class Bootstrap {
public static void main(String[] args) {
ServerBuilder sb = new ServerBuilder();
sb.port(8081, "http");
sb.serviceUnder("/hello",
new GrpcServiceBuilder()
.addService(new MyHelloService())
.supportedSerializationFormats(GrpcSerializationFormats.values())
.enableUnframedRequests(true)
.build());
sb.accessLogWriter(AccessLogWriters.common());
sb.serviceUnder("/docs", new DocService());
Server server = sb.build();
ServerListener serverListener = new ZooKeeperUpdatingListener("localhost:2181", "/myProducionEndpoints", 1000);
server.addListener(serverListener);
server.start().join();
}
}
```
If i remove the `ServerListener`, it works well. And i looked into `com.linecorp.armeria.server.Server`, found `primaryActivePort` is set in `ServerPortStartListener`, but `ZooKeeperUpdatingListener.serverStarting` is triggered before `ServerPortStartListener` added to Server start future listeners.
Did I miss something?
I am not sure whether I explained this problem clear or not. Thanks in advance. | 1.0 | java.util.NoSuchElementException: No value present when using ZooKeeperUpdatingListener - Hi, I am new to armeria. I build a simple project using armeria 0.5.60 according to [armeria documentation](http://line.github.io/armeria/index.html), everything goes well until using 'Service discovery with ZooKeeper'. I got this exception
```
java.util.NoSuchElementException: No value present
at java.util.Optional.get(Optional.java:135)
at java.util.concurrent.CompletableFuture.reportJoin(CompletableFuture.java:375)
at com.linecorp.armeria.server.zookeeper.ZooKeeperUpdatingListener.serverStarting(ZooKeeperUpdatingListener.java:69)
at java.util.concurrent.CompletableFuture.join(CompletableFuture.java:1934)
at com.linecorp.armeria.server.Server$StateManager.notifyState(Server.java:585)
at com.github.liuzhengyang.server.Bootstrap.main(Bootstrap.java:38)
at com.linecorp.armeria.server.Server$StateManager.enterStarting(Server.java:541)
Caused by: java.lang.IllegalStateException: failed to notify all server listeners
at com.linecorp.armeria.server.Server.start(Server.java:250)
at com.linecorp.armeria.server.Server$StateManager.enterStarting(Server.java:542)
at com.linecorp.armeria.server.Server.start(Server.java:235)
at com.linecorp.armeria.server.Server.start(Server.java:250)
at com.github.liuzhengyang.server.Bootstrap.main(Bootstrap.java:38)
at com.linecorp.armeria.server.Server.start(Server.java:235)
... 1 more
```
And here is my example code.
```
public class Bootstrap {
public static void main(String[] args) {
ServerBuilder sb = new ServerBuilder();
sb.port(8081, "http");
sb.serviceUnder("/hello",
new GrpcServiceBuilder()
.addService(new MyHelloService())
.supportedSerializationFormats(GrpcSerializationFormats.values())
.enableUnframedRequests(true)
.build());
sb.accessLogWriter(AccessLogWriters.common());
sb.serviceUnder("/docs", new DocService());
Server server = sb.build();
ServerListener serverListener = new ZooKeeperUpdatingListener("localhost:2181", "/myProducionEndpoints", 1000);
server.addListener(serverListener);
server.start().join();
}
}
```
If i remove the `ServerListener`, it works well. And i looked into `com.linecorp.armeria.server.Server`, found `primaryActivePort` is set in `ServerPortStartListener`, but `ZooKeeperUpdatingListener.serverStarting` is triggered before `ServerPortStartListener` added to Server start future listeners.
Did I miss something?
I am not sure whether I explained this problem clear or not. Thanks in advance. | defect | java util nosuchelementexception no value present when using zookeeperupdatinglistener hi i am new to armeria i build a simple project using armeria according to everything goes well until using service discovery with zookeeper i got this exception java util nosuchelementexception no value present at java util optional get optional java at java util concurrent completablefuture reportjoin completablefuture java at com linecorp armeria server zookeeper zookeeperupdatinglistener serverstarting zookeeperupdatinglistener java at java util concurrent completablefuture join completablefuture java at com linecorp armeria server server statemanager notifystate server java at com github liuzhengyang server bootstrap main bootstrap java at com linecorp armeria server server statemanager enterstarting server java caused by java lang illegalstateexception failed to notify all server listeners at com linecorp armeria server server start server java at com linecorp armeria server server statemanager enterstarting server java at com linecorp armeria server server start server java at com linecorp armeria server server start server java at com github liuzhengyang server bootstrap main bootstrap java at com linecorp armeria server server start server java more and here is my example code public class bootstrap public static void main string args serverbuilder sb new serverbuilder sb port http sb serviceunder hello new grpcservicebuilder addservice new myhelloservice supportedserializationformats grpcserializationformats values enableunframedrequests true build sb accesslogwriter accesslogwriters common sb serviceunder docs new docservice server server sb build serverlistener serverlistener new zookeeperupdatinglistener localhost myproducionendpoints server addlistener serverlistener server start join if i remove the serverlistener it works well and i looked into com linecorp armeria server server found primaryactiveport is set in serverportstartlistener but zookeeperupdatinglistener serverstarting is triggered before serverportstartlistener added to server start future listeners did i miss something i am not sure whether i explained this problem clear or not thanks in advance | 1 |
52,835 | 13,225,120,906 | IssuesEvent | 2020-08-17 20:31:58 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | pybinding of TimeResidual requires 5 args (Trac #435) | Migrated from Trac combo core defect | The 2 last arguments of the TimeResidual function of I3Calculator, don't use their default (I3Constants.n_ice_group, n_ice_phase) in python and need to explicitly given which is not the case in C++ land.
only the following works now :
phys_services.I3Calculator.time_residual(track, omgeo[omkey].position, pulse.time,dataclasses.I3Constants.n_ice_group,dataclasses.I3Constants.n_ice_phase)
and not (yet) :
phys_services.I3Calculator.time_residual(track, omgeo[omkey].position, pulse.time)
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/435">https://code.icecube.wisc.edu/projects/icecube/ticket/435</a>, reported by icecubeand owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-07-23T23:33:19",
"_ts": "1343086399000000",
"description": "The 2 last arguments of the TimeResidual function of I3Calculator, don't use their default (I3Constants.n_ice_group, n_ice_phase) in python and need to explicitly given which is not the case in C++ land.\n\nonly the following works now :\nphys_services.I3Calculator.time_residual(track, omgeo[omkey].position, pulse.time,dataclasses.I3Constants.n_ice_group,dataclasses.I3Constants.n_ice_phase)\n\nand not (yet) :\nphys_services.I3Calculator.time_residual(track, omgeo[omkey].position, pulse.time)",
"reporter": "icecube",
"cc": "",
"resolution": "fixed",
"time": "2012-07-23T14:35:48",
"component": "combo core",
"summary": "pybinding of TimeResidual requires 5 args",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
| 1.0 | pybinding of TimeResidual requires 5 args (Trac #435) - The 2 last arguments of the TimeResidual function of I3Calculator, don't use their default (I3Constants.n_ice_group, n_ice_phase) in python and need to explicitly given which is not the case in C++ land.
only the following works now :
phys_services.I3Calculator.time_residual(track, omgeo[omkey].position, pulse.time,dataclasses.I3Constants.n_ice_group,dataclasses.I3Constants.n_ice_phase)
and not (yet) :
phys_services.I3Calculator.time_residual(track, omgeo[omkey].position, pulse.time)
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/435">https://code.icecube.wisc.edu/projects/icecube/ticket/435</a>, reported by icecubeand owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-07-23T23:33:19",
"_ts": "1343086399000000",
"description": "The 2 last arguments of the TimeResidual function of I3Calculator, don't use their default (I3Constants.n_ice_group, n_ice_phase) in python and need to explicitly given which is not the case in C++ land.\n\nonly the following works now :\nphys_services.I3Calculator.time_residual(track, omgeo[omkey].position, pulse.time,dataclasses.I3Constants.n_ice_group,dataclasses.I3Constants.n_ice_phase)\n\nand not (yet) :\nphys_services.I3Calculator.time_residual(track, omgeo[omkey].position, pulse.time)",
"reporter": "icecube",
"cc": "",
"resolution": "fixed",
"time": "2012-07-23T14:35:48",
"component": "combo core",
"summary": "pybinding of TimeResidual requires 5 args",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
| defect | pybinding of timeresidual requires args trac the last arguments of the timeresidual function of don t use their default n ice group n ice phase in python and need to explicitly given which is not the case in c land only the following works now phys services time residual track omgeo position pulse time dataclasses n ice group dataclasses n ice phase and not yet phys services time residual track omgeo position pulse time migrated from json status closed changetime ts description the last arguments of the timeresidual function of don t use their default n ice group n ice phase in python and need to explicitly given which is not the case in c land n nonly the following works now nphys services time residual track omgeo position pulse time dataclasses n ice group dataclasses n ice phase n nand not yet nphys services time residual track omgeo position pulse time reporter icecube cc resolution fixed time component combo core summary pybinding of timeresidual requires args priority normal keywords milestone owner jvansanten type defect | 1 |
65,514 | 19,559,198,629 | IssuesEvent | 2022-01-03 14:03:00 | vector-im/element-android | https://api.github.com/repos/vector-im/element-android | closed | dev tools is blank | T-Defect | ### Steps to reproduce
1. Where are you starting? What can you see?
in any chat room
2. What do you click?
three dots in upper right corner, then dev tools in the menu that pops down.
3. observe a blank screen
### Outcome
#### What did you expect?
some stuff on the screen
#### What happened instead?
it's blank

### Your phone model
Pixel 4a
### Operating system version
Android 11.0
### Application version and app store
1.3.9 [40103090] (F-581c557f), olm 3.2.4 f-droid
### Homeserver
matrix.org
### Will you send logs?
No | 1.0 | dev tools is blank - ### Steps to reproduce
1. Where are you starting? What can you see?
in any chat room
2. What do you click?
three dots in upper right corner, then dev tools in the menu that pops down.
3. observe a blank screen
### Outcome
#### What did you expect?
some stuff on the screen
#### What happened instead?
it's blank

### Your phone model
Pixel 4a
### Operating system version
Android 11.0
### Application version and app store
1.3.9 [40103090] (F-581c557f), olm 3.2.4 f-droid
### Homeserver
matrix.org
### Will you send logs?
No | defect | dev tools is blank steps to reproduce where are you starting what can you see in any chat room what do you click three dots in upper right corner then dev tools in the menu that pops down observe a blank screen outcome what did you expect some stuff on the screen what happened instead it s blank your phone model pixel operating system version android application version and app store f olm f droid homeserver matrix org will you send logs no | 1 |
59,278 | 17,016,797,027 | IssuesEvent | 2021-07-02 13:11:20 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | opened | Shared cycleways both with cars and buses not rendered in opencyclemap ! | Component: opencyclemap Priority: major Type: defect | **[Submitted to the original trac issue database at 2.02am, Monday, 3rd December 2018]**
Hi,
Cycleways that has no lines but only a shared space with roads, like also bus lines shared with bikes, are not rendered in the opencyclemap layer.
Thanks,
Mat | 1.0 | Shared cycleways both with cars and buses not rendered in opencyclemap ! - **[Submitted to the original trac issue database at 2.02am, Monday, 3rd December 2018]**
Hi,
Cycleways that has no lines but only a shared space with roads, like also bus lines shared with bikes, are not rendered in the opencyclemap layer.
Thanks,
Mat | defect | shared cycleways both with cars and buses not rendered in opencyclemap hi cycleways that has no lines but only a shared space with roads like also bus lines shared with bikes are not rendered in the opencyclemap layer thanks mat | 1 |
115,482 | 24,769,992,160 | IssuesEvent | 2022-10-23 02:14:21 | llvm/llvm-project | https://api.github.com/repos/llvm/llvm-project | closed | [RISCV] Backend hangs during `RISCV DAG->DAG Pattern Instruction Selection` when having binary ops after `select` | backend:RISC-V llvm:codegen | ## Description
The backend hangs during `RISCV DAG->DAG Pattern Instruction Selection` pass for the following input IR when targeting `riscv32` / `riscv64`.
The constant operands for `select` must be -1 and 0 (order doesn't matter) to reproduce the problem.
The problem also occurs when `mul` is replaced by `add` / `sub`, but not `udiv` / `sdiv` / `urem` / `srem`.
The problem also occurs when `or` is replaced by `and`, but not `xor`.
The constant used in the 2 binary ops must be greater than or equal to 2048 and the constant operand for `mul` must be equal or slightly greater than the constant operand used for `or` (difference needs to be below 2048) to reproduce the problem.
## Cause
[`DAGCombiner::foldBinOpIntoSelect`](https://github.com/llvm/llvm-project/blob/4309bb28ae77061d528b09dfe1c41335e92bc7a3/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp#L2228) and [`DAGCombiner::foldSelectOfBinops`](https://github.com/llvm/llvm-project/blob/4309bb28ae77061d528b09dfe1c41335e92bc7a3/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp#L24342) are called for infinite number of times and the DAG is alternating between 2 forms as shown below:
```
SelectionDAG has 19 nodes:
t0: ch,glue = EntryToken
t4: i32,ch = CopyFromReg t0, Register:i32 %1
t13: i32 = mul t4, OpaqueConstant:i32<2048>
t6: i32,ch = CopyFromReg t0, Register:i32 %2
t16: ch = store<(store (s32) into %ir.2)> t0, t13, t6, undef:i32
t2: i32,ch = CopyFromReg t0, Register:i32 %0
t7: i1 = truncate t2
t12: i32 = select t7, Constant:i32<-1>, Constant:i32<0>
t14: i32 = or t12, OpaqueConstant:i32<2048>
t18: ch,glue = CopyToReg t16, Register:i32 $x10, t14
t19: ch = RISCVISD::RET_FLAG t18, Register:i32 $x10, t18:1
SelectionDAG has 20 nodes:
t0: ch,glue = EntryToken
t4: i32,ch = CopyFromReg t0, Register:i32 %1
t13: i32 = mul t4, OpaqueConstant:i32<2048>
t6: i32,ch = CopyFromReg t0, Register:i32 %2
t16: ch = store<(store (s32) into %ir.2)> t0, t13, t6, undef:i32
t2: i32,ch = CopyFromReg t0, Register:i32 %0
t7: i1 = truncate t2
t20: i32 = or Constant:i32<-1>, OpaqueConstant:i32<2048>
t21: i32 = or Constant:i32<0>, OpaqueConstant:i32<2048>
t22: i32 = select t7, t20, t21
t18: ch,glue = CopyToReg t16, Register:i32 $x10, t22
t19: ch = RISCVISD::RET_FLAG t18, Register:i32 $x10, t18:1
SelectionDAG has 19 nodes:
t0: ch,glue = EntryToken
t4: i32,ch = CopyFromReg t0, Register:i32 %1
t13: i32 = mul t4, OpaqueConstant:i32<2048>
t6: i32,ch = CopyFromReg t0, Register:i32 %2
t16: ch = store<(store (s32) into %ir.2)> t0, t13, t6, undef:i32
t2: i32,ch = CopyFromReg t0, Register:i32 %0
t7: i1 = truncate t2
t23: i32 = select t7, Constant:i32<-1>, Constant:i32<0>
t24: i32 = or t23, OpaqueConstant:i32<2048>
t18: ch,glue = CopyToReg t16, Register:i32 $x10, t24
t19: ch = RISCVISD::RET_FLAG t18, Register:i32 $x10, t18:1
SelectionDAG has 20 nodes:
t0: ch,glue = EntryToken
t4: i32,ch = CopyFromReg t0, Register:i32 %1
t13: i32 = mul t4, OpaqueConstant:i32<2048>
t6: i32,ch = CopyFromReg t0, Register:i32 %2
t16: ch = store<(store (s32) into %ir.2)> t0, t13, t6, undef:i32
t2: i32,ch = CopyFromReg t0, Register:i32 %0
t7: i1 = truncate t2
t25: i32 = or Constant:i32<-1>, OpaqueConstant:i32<2048>
t26: i32 = or Constant:i32<0>, OpaqueConstant:i32<2048>
t27: i32 = select t7, t25, t26
t18: ch,glue = CopyToReg t16, Register:i32 $x10, t27
t19: ch = RISCVISD::RET_FLAG t18, Register:i32 $x10, t18:1
SelectionDAG has 19 nodes:
t0: ch,glue = EntryToken
t4: i32,ch = CopyFromReg t0, Register:i32 %1
t13: i32 = mul t4, OpaqueConstant:i32<2048>
t6: i32,ch = CopyFromReg t0, Register:i32 %2
t16: ch = store<(store (s32) into %ir.2)> t0, t13, t6, undef:i32
t2: i32,ch = CopyFromReg t0, Register:i32 %0
t7: i1 = truncate t2
t28: i32 = select t7, Constant:i32<-1>, Constant:i32<0>
t29: i32 = or t28, OpaqueConstant:i32<2048>
t18: ch,glue = CopyToReg t16, Register:i32 $x10, t29
t19: ch = RISCVISD::RET_FLAG t18, Register:i32 $x10, t18:1
...
```
### Related Code
https://github.com/llvm/llvm-project/blob/4309bb28ae77061d528b09dfe1c41335e92bc7a3/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp#L24364-L24373
https://github.com/llvm/llvm-project/blob/4309bb28ae77061d528b09dfe1c41335e92bc7a3/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp#L2283-L2304
## Minimal Reproduction
https://godbolt.org/z/5xE3c6xjb
### Code
```ll
define i32 @f(i1 %0, i32 %1, ptr %2) {
BB:
%I = select i1 %0, i32 -1, i32 0
%I1 = mul i32 %1, 2048
%I2 = or i32 2048, %I
store i32 %I1, ptr %2
ret i32 %I2
}
``` | 1.0 | [RISCV] Backend hangs during `RISCV DAG->DAG Pattern Instruction Selection` when having binary ops after `select` - ## Description
The backend hangs during `RISCV DAG->DAG Pattern Instruction Selection` pass for the following input IR when targeting `riscv32` / `riscv64`.
The constant operands for `select` must be -1 and 0 (order doesn't matter) to reproduce the problem.
The problem also occurs when `mul` is replaced by `add` / `sub`, but not `udiv` / `sdiv` / `urem` / `srem`.
The problem also occurs when `or` is replaced by `and`, but not `xor`.
The constant used in the 2 binary ops must be greater than or equal to 2048 and the constant operand for `mul` must be equal or slightly greater than the constant operand used for `or` (difference needs to be below 2048) to reproduce the problem.
## Cause
[`DAGCombiner::foldBinOpIntoSelect`](https://github.com/llvm/llvm-project/blob/4309bb28ae77061d528b09dfe1c41335e92bc7a3/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp#L2228) and [`DAGCombiner::foldSelectOfBinops`](https://github.com/llvm/llvm-project/blob/4309bb28ae77061d528b09dfe1c41335e92bc7a3/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp#L24342) are called for infinite number of times and the DAG is alternating between 2 forms as shown below:
```
SelectionDAG has 19 nodes:
t0: ch,glue = EntryToken
t4: i32,ch = CopyFromReg t0, Register:i32 %1
t13: i32 = mul t4, OpaqueConstant:i32<2048>
t6: i32,ch = CopyFromReg t0, Register:i32 %2
t16: ch = store<(store (s32) into %ir.2)> t0, t13, t6, undef:i32
t2: i32,ch = CopyFromReg t0, Register:i32 %0
t7: i1 = truncate t2
t12: i32 = select t7, Constant:i32<-1>, Constant:i32<0>
t14: i32 = or t12, OpaqueConstant:i32<2048>
t18: ch,glue = CopyToReg t16, Register:i32 $x10, t14
t19: ch = RISCVISD::RET_FLAG t18, Register:i32 $x10, t18:1
SelectionDAG has 20 nodes:
t0: ch,glue = EntryToken
t4: i32,ch = CopyFromReg t0, Register:i32 %1
t13: i32 = mul t4, OpaqueConstant:i32<2048>
t6: i32,ch = CopyFromReg t0, Register:i32 %2
t16: ch = store<(store (s32) into %ir.2)> t0, t13, t6, undef:i32
t2: i32,ch = CopyFromReg t0, Register:i32 %0
t7: i1 = truncate t2
t20: i32 = or Constant:i32<-1>, OpaqueConstant:i32<2048>
t21: i32 = or Constant:i32<0>, OpaqueConstant:i32<2048>
t22: i32 = select t7, t20, t21
t18: ch,glue = CopyToReg t16, Register:i32 $x10, t22
t19: ch = RISCVISD::RET_FLAG t18, Register:i32 $x10, t18:1
SelectionDAG has 19 nodes:
t0: ch,glue = EntryToken
t4: i32,ch = CopyFromReg t0, Register:i32 %1
t13: i32 = mul t4, OpaqueConstant:i32<2048>
t6: i32,ch = CopyFromReg t0, Register:i32 %2
t16: ch = store<(store (s32) into %ir.2)> t0, t13, t6, undef:i32
t2: i32,ch = CopyFromReg t0, Register:i32 %0
t7: i1 = truncate t2
t23: i32 = select t7, Constant:i32<-1>, Constant:i32<0>
t24: i32 = or t23, OpaqueConstant:i32<2048>
t18: ch,glue = CopyToReg t16, Register:i32 $x10, t24
t19: ch = RISCVISD::RET_FLAG t18, Register:i32 $x10, t18:1
SelectionDAG has 20 nodes:
t0: ch,glue = EntryToken
t4: i32,ch = CopyFromReg t0, Register:i32 %1
t13: i32 = mul t4, OpaqueConstant:i32<2048>
t6: i32,ch = CopyFromReg t0, Register:i32 %2
t16: ch = store<(store (s32) into %ir.2)> t0, t13, t6, undef:i32
t2: i32,ch = CopyFromReg t0, Register:i32 %0
t7: i1 = truncate t2
t25: i32 = or Constant:i32<-1>, OpaqueConstant:i32<2048>
t26: i32 = or Constant:i32<0>, OpaqueConstant:i32<2048>
t27: i32 = select t7, t25, t26
t18: ch,glue = CopyToReg t16, Register:i32 $x10, t27
t19: ch = RISCVISD::RET_FLAG t18, Register:i32 $x10, t18:1
SelectionDAG has 19 nodes:
t0: ch,glue = EntryToken
t4: i32,ch = CopyFromReg t0, Register:i32 %1
t13: i32 = mul t4, OpaqueConstant:i32<2048>
t6: i32,ch = CopyFromReg t0, Register:i32 %2
t16: ch = store<(store (s32) into %ir.2)> t0, t13, t6, undef:i32
t2: i32,ch = CopyFromReg t0, Register:i32 %0
t7: i1 = truncate t2
t28: i32 = select t7, Constant:i32<-1>, Constant:i32<0>
t29: i32 = or t28, OpaqueConstant:i32<2048>
t18: ch,glue = CopyToReg t16, Register:i32 $x10, t29
t19: ch = RISCVISD::RET_FLAG t18, Register:i32 $x10, t18:1
...
```
### Related Code
https://github.com/llvm/llvm-project/blob/4309bb28ae77061d528b09dfe1c41335e92bc7a3/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp#L24364-L24373
https://github.com/llvm/llvm-project/blob/4309bb28ae77061d528b09dfe1c41335e92bc7a3/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp#L2283-L2304
## Minimal Reproduction
https://godbolt.org/z/5xE3c6xjb
### Code
```ll
define i32 @f(i1 %0, i32 %1, ptr %2) {
BB:
%I = select i1 %0, i32 -1, i32 0
%I1 = mul i32 %1, 2048
%I2 = or i32 2048, %I
store i32 %I1, ptr %2
ret i32 %I2
}
``` | non_defect | backend hangs during riscv dag dag pattern instruction selection when having binary ops after select description the backend hangs during riscv dag dag pattern instruction selection pass for the following input ir when targeting the constant operands for select must be and order doesn t matter to reproduce the problem the problem also occurs when mul is replaced by add sub but not udiv sdiv urem srem the problem also occurs when or is replaced by and but not xor the constant used in the binary ops must be greater than or equal to and the constant operand for mul must be equal or slightly greater than the constant operand used for or difference needs to be below to reproduce the problem cause and are called for infinite number of times and the dag is alternating between forms as shown below selectiondag has nodes ch glue entrytoken ch copyfromreg register mul opaqueconstant ch copyfromreg register ch store undef ch copyfromreg register truncate select constant constant or opaqueconstant ch glue copytoreg register ch riscvisd ret flag register selectiondag has nodes ch glue entrytoken ch copyfromreg register mul opaqueconstant ch copyfromreg register ch store undef ch copyfromreg register truncate or constant opaqueconstant or constant opaqueconstant select ch glue copytoreg register ch riscvisd ret flag register selectiondag has nodes ch glue entrytoken ch copyfromreg register mul opaqueconstant ch copyfromreg register ch store undef ch copyfromreg register truncate select constant constant or opaqueconstant ch glue copytoreg register ch riscvisd ret flag register selectiondag has nodes ch glue entrytoken ch copyfromreg register mul opaqueconstant ch copyfromreg register ch store undef ch copyfromreg register truncate or constant opaqueconstant or constant opaqueconstant select ch glue copytoreg register ch riscvisd ret flag register selectiondag has nodes ch glue entrytoken ch copyfromreg register mul opaqueconstant ch copyfromreg register ch store undef ch copyfromreg register truncate select constant constant or opaqueconstant ch glue copytoreg register ch riscvisd ret flag register related code minimal reproduction code ll define f ptr bb i select mul or i store ptr ret | 0 |
73,758 | 24,786,756,087 | IssuesEvent | 2022-10-24 10:25:57 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Unable to verify device | T-Defect | ### Steps to reproduce
1. On login, get the popup "You have unverified logins"
2. Click "Review"
3. Click "Verify“ for „This Device"
4. Click "Verify with Security Key"
5. Click "Upload"
6. Select key
7. Click "Continue"
### Outcome
#### What did you expect?
Successful verification
#### What happened instead?
I found myself back at the "Verify this session"-screen from step 4.
Log gives me "Error: Key type master from getCrossSigningKey callback did not match"
Except for the log I see the same on the WebApp and the Element AndroidApp Errors with "Failed to import keys"
### Operating system
Linux
### Application version
Element version: 1.11.8 Olm version: 3.2.12
### How did you install the app?
gentoo portage, package net-im/element-desktop-bin
### Homeserver
synapse 1.69.0
### Will you send logs?
Yes | 1.0 | Unable to verify device - ### Steps to reproduce
1. On login, get the popup "You have unverified logins"
2. Click "Review"
3. Click "Verify“ for „This Device"
4. Click "Verify with Security Key"
5. Click "Upload"
6. Select key
7. Click "Continue"
### Outcome
#### What did you expect?
Successful verification
#### What happened instead?
I found myself back at the "Verify this session"-screen from step 4.
Log gives me "Error: Key type master from getCrossSigningKey callback did not match"
Except for the log I see the same on the WebApp and the Element AndroidApp Errors with "Failed to import keys"
### Operating system
Linux
### Application version
Element version: 1.11.8 Olm version: 3.2.12
### How did you install the app?
gentoo portage, package net-im/element-desktop-bin
### Homeserver
synapse 1.69.0
### Will you send logs?
Yes | defect | unable to verify device steps to reproduce on login get the popup you have unverified logins click review click verify“ for „this device click verify with security key click upload select key click continue outcome what did you expect successful verification what happened instead i found myself back at the verify this session screen from step log gives me error key type master from getcrosssigningkey callback did not match except for the log i see the same on the webapp and the element androidapp errors with failed to import keys operating system linux application version element version olm version how did you install the app gentoo portage package net im element desktop bin homeserver synapse will you send logs yes | 1 |
83,771 | 10,338,475,848 | IssuesEvent | 2019-09-03 17:00:51 | OSOSystem/oso-backend | https://api.github.com/repos/OSOSystem/oso-backend | closed | Obtain CII Best Practices Badge | documentation | A lot of open source projects have obtained this badge. It shows that the project is following certain patterns and rules which are widely adopted by the open source community.
Our current situation is that we are missing some points listed on
https://bestpractices.coreinfrastructure.org/en
This issue should add the missing pieces to meet the requirements.
After this issue is done the badge can be requested. | 1.0 | Obtain CII Best Practices Badge - A lot of open source projects have obtained this badge. It shows that the project is following certain patterns and rules which are widely adopted by the open source community.
Our current situation is that we are missing some points listed on
https://bestpractices.coreinfrastructure.org/en
This issue should add the missing pieces to meet the requirements.
After this issue is done the badge can be requested. | non_defect | obtain cii best practices badge a lot of open source projects have obtained this badge it shows that the project is following certain patterns and rules which are widely adopted by the open source community our current situation is that we are missing some points listed on this issue should add the missing pieces to meet the requirements after this issue is done the badge can be requested | 0 |
454,579 | 13,104,019,407 | IssuesEvent | 2020-08-04 09:32:49 | folkehelseinstituttet/dashboards_sykdomspuls | https://api.github.com/repos/folkehelseinstituttet/dashboards_sykdomspuls | closed | Change obs mail | high priority | - [x] Remove 'Geografisk område'='municip0301' and replace it with Fylke name
- [ ] Two additional columns that show what has happened the 2 previous weeks (e.g. normal/medium/high), maybe colours. Maybe both?
| 1.0 | Change obs mail - - [x] Remove 'Geografisk område'='municip0301' and replace it with Fylke name
- [ ] Two additional columns that show what has happened the 2 previous weeks (e.g. normal/medium/high), maybe colours. Maybe both?
| non_defect | change obs mail remove geografisk område and replace it with fylke name two additional columns that show what has happened the previous weeks e g normal medium high maybe colours maybe both | 0 |
163,537 | 20,363,857,651 | IssuesEvent | 2022-02-21 01:36:59 | benchmarkdebricked/laravel | https://api.github.com/repos/benchmarkdebricked/laravel | opened | CVE-2021-27515 (Medium) detected in url-parse-1.4.7.tgz | security vulnerability | ## CVE-2021-27515 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /laravel/package.json</p>
<p>Path to vulnerable library: /node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- laravel-mix-4.0.16.tgz (Root Library)
- webpack-dev-server-3.7.1.tgz
- sockjs-client-1.3.0.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
url-parse before 1.5.0 mishandles certain uses of backslash such as http:\/ and interprets the URI as a relative path.
<p>Publish Date: 2021-02-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27515>CVE-2021-27515</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515</a></p>
<p>Release Date: 2021-02-22</p>
<p>Fix Resolution (url-parse): 1.5.0</p>
<p>Direct dependency fix Resolution (laravel-mix): 4.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-27515 (Medium) detected in url-parse-1.4.7.tgz - ## CVE-2021-27515 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /laravel/package.json</p>
<p>Path to vulnerable library: /node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- laravel-mix-4.0.16.tgz (Root Library)
- webpack-dev-server-3.7.1.tgz
- sockjs-client-1.3.0.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
url-parse before 1.5.0 mishandles certain uses of backslash such as http:\/ and interprets the URI as a relative path.
<p>Publish Date: 2021-02-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27515>CVE-2021-27515</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515</a></p>
<p>Release Date: 2021-02-22</p>
<p>Fix Resolution (url-parse): 1.5.0</p>
<p>Direct dependency fix Resolution (laravel-mix): 4.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in url parse tgz cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file laravel package json path to vulnerable library node modules url parse package json dependency hierarchy laravel mix tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library vulnerability details url parse before mishandles certain uses of backslash such as http and interprets the uri as a relative path publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse direct dependency fix resolution laravel mix step up your open source security game with whitesource | 0 |
231,556 | 18,778,219,151 | IssuesEvent | 2021-11-08 00:41:37 | backend-br/vagas | https://api.github.com/repos/backend-br/vagas | closed | [REMOTO] Back-end Developer NodeJS Pleno @ Trakto | PJ Pleno NodeJS Express Ruby Remoto Testes automatizados TypeScript Git startup Stale | ## Trakto (https://trakto.io)
<img width="1597" alt="Captura de Tela 2021-09-03 às 09 41 02" src="https://user-images.githubusercontent.com/5194011/132006807-59c0237f-2629-4a91-b353-9f8175abcb46.png">
Somos uma startup uma startup para criação de "design's" digitais e impressos (https://www.youtube.com/watch?v=FLEMUC1peqs). Atuamos no Brasil e outros países, contamos com mais de 30k usuários ativos na plataforma.
## Estudos de caso do Google sobre a Trakto
**Google Cloud** - https://cloud.google.com/customers/trakto
**Google For Startups** - https://startup.google.com/stories/trakto/
**História e valores da Trakto** - https://www.youtube.com/watch?v=Vet268jSUVQ
## Descrição da vaga
Como pessoa desenvolvedora backend na Trakto, você realizará:
- Desenvolver e manter APIs e microsserviços privadas e públicas, como por exemplo Trakto SDK;
- Garantir a entrega automatizada com qualidade das APIs e microserviços em produção;
- Documentar as aplicações de backend desenvolvidas para que outros devs possam utilizar;
- Criar testes automatizados;
- Realizar integrações com aplicações de terceiros e disponibilizar como microsserviços;
- Propor soluções práticas
## Local
100% Remoto
## Requisitos
- Experiência com backend utilizando NodeJS;
- GIT e metodologia de versionamento usando GitFlow;
- Domínio de JS e Typescript;
- Conhecimento e experiência de uso com padrões de projeto (Clean Code, Repository pattern, comunicação ubíqua e outros que possam garantir qualidade);
- Experiência com express ou frameworks derivados, como por exemplo, NestJS;
**Desejáveis:**
- conhecimentos em integração e entrega contínua;
- conhecimento em arquitetura de software e integração com plataformas de terceiros;
**Diferenciais:**
- Uso de stacks mais atuais do mercado;
- Facilidade e liberar para implementar novas ideiais
## Benefícios
- Time off;
- Ajuda de custos com cursos;
## Contratação
PJ
## Como se candidatar
Por favor, acesse o site **https://trakto.io/carreiras para realizar a sua candidatura a vaga/**
## Tempo médio de feedbacks
Costumamos enviar feedbacks em até **15 dias após cada processo**.
E-mail para contato em caso de não haver resposta: marinatorres@trakto.io
## Labels
#### Alocação
- Remoto
#### Regime
- PJ
#### Nível
- Pleno
| 1.0 | [REMOTO] Back-end Developer NodeJS Pleno @ Trakto - ## Trakto (https://trakto.io)
<img width="1597" alt="Captura de Tela 2021-09-03 às 09 41 02" src="https://user-images.githubusercontent.com/5194011/132006807-59c0237f-2629-4a91-b353-9f8175abcb46.png">
Somos uma startup uma startup para criação de "design's" digitais e impressos (https://www.youtube.com/watch?v=FLEMUC1peqs). Atuamos no Brasil e outros países, contamos com mais de 30k usuários ativos na plataforma.
## Estudos de caso do Google sobre a Trakto
**Google Cloud** - https://cloud.google.com/customers/trakto
**Google For Startups** - https://startup.google.com/stories/trakto/
**História e valores da Trakto** - https://www.youtube.com/watch?v=Vet268jSUVQ
## Descrição da vaga
Como pessoa desenvolvedora backend na Trakto, você realizará:
- Desenvolver e manter APIs e microsserviços privadas e públicas, como por exemplo Trakto SDK;
- Garantir a entrega automatizada com qualidade das APIs e microserviços em produção;
- Documentar as aplicações de backend desenvolvidas para que outros devs possam utilizar;
- Criar testes automatizados;
- Realizar integrações com aplicações de terceiros e disponibilizar como microsserviços;
- Propor soluções práticas
## Local
100% Remoto
## Requisitos
- Experiência com backend utilizando NodeJS;
- GIT e metodologia de versionamento usando GitFlow;
- Domínio de JS e Typescript;
- Conhecimento e experiência de uso com padrões de projeto (Clean Code, Repository pattern, comunicação ubíqua e outros que possam garantir qualidade);
- Experiência com express ou frameworks derivados, como por exemplo, NestJS;
**Desejáveis:**
- conhecimentos em integração e entrega contínua;
- conhecimento em arquitetura de software e integração com plataformas de terceiros;
**Diferenciais:**
- Uso de stacks mais atuais do mercado;
- Facilidade e liberar para implementar novas ideiais
## Benefícios
- Time off;
- Ajuda de custos com cursos;
## Contratação
PJ
## Como se candidatar
Por favor, acesse o site **https://trakto.io/carreiras para realizar a sua candidatura a vaga/**
## Tempo médio de feedbacks
Costumamos enviar feedbacks em até **15 dias após cada processo**.
E-mail para contato em caso de não haver resposta: marinatorres@trakto.io
## Labels
#### Alocação
- Remoto
#### Regime
- PJ
#### Nível
- Pleno
| non_defect | back end developer nodejs pleno trakto trakto img width alt captura de tela às src somos uma startup uma startup para criação de design s digitais e impressos atuamos no brasil e outros países contamos com mais de usuários ativos na plataforma estudos de caso do google sobre a trakto google cloud google for startups história e valores da trakto descrição da vaga como pessoa desenvolvedora backend na trakto você realizará desenvolver e manter apis e microsserviços privadas e públicas como por exemplo trakto sdk garantir a entrega automatizada com qualidade das apis e microserviços em produção documentar as aplicações de backend desenvolvidas para que outros devs possam utilizar criar testes automatizados realizar integrações com aplicações de terceiros e disponibilizar como microsserviços propor soluções práticas local remoto requisitos experiência com backend utilizando nodejs git e metodologia de versionamento usando gitflow domínio de js e typescript conhecimento e experiência de uso com padrões de projeto clean code repository pattern comunicação ubíqua e outros que possam garantir qualidade experiência com express ou frameworks derivados como por exemplo nestjs desejáveis conhecimentos em integração e entrega contínua conhecimento em arquitetura de software e integração com plataformas de terceiros diferenciais uso de stacks mais atuais do mercado facilidade e liberar para implementar novas ideiais benefícios time off ajuda de custos com cursos contratação pj como se candidatar por favor acesse o site para realizar a sua candidatura a vaga tempo médio de feedbacks costumamos enviar feedbacks em até dias após cada processo e mail para contato em caso de não haver resposta marinatorres trakto io labels alocação remoto regime pj nível pleno | 0 |
160,612 | 25,198,421,817 | IssuesEvent | 2022-11-12 20:32:11 | gitpod-io/website | https://api.github.com/repos/gitpod-io/website | opened | Reduce font sizes in /cde to allow "Principles of CDEs" to fit in one screen | type: design polish | I have a macbook 13" and the 8 principles of CDEs don't all fit on my screen.
The sections which use `text-p-large` feel a little overwhelming compared to other sections.
`text-p-large` is easy to read, but having to scroll can also detract from the message.
I would suggest using `text-p-medium` (18pt) for the paragraphs, and h4 (28pt) instead of h3 for each heading in the 8 principles.
| 1.0 | Reduce font sizes in /cde to allow "Principles of CDEs" to fit in one screen - I have a macbook 13" and the 8 principles of CDEs don't all fit on my screen.
The sections which use `text-p-large` feel a little overwhelming compared to other sections.
`text-p-large` is easy to read, but having to scroll can also detract from the message.
I would suggest using `text-p-medium` (18pt) for the paragraphs, and h4 (28pt) instead of h3 for each heading in the 8 principles.
| non_defect | reduce font sizes in cde to allow principles of cdes to fit in one screen i have a macbook and the principles of cdes don t all fit on my screen the sections which use text p large feel a little overwhelming compared to other sections text p large is easy to read but having to scroll can also detract from the message i would suggest using text p medium for the paragraphs and instead of for each heading in the principles | 0 |
327,229 | 24,124,022,004 | IssuesEvent | 2022-09-20 21:36:44 | eic/EICrecon | https://api.github.com/repos/eic/EICrecon | closed | Add documentation for ACTS tracking part | documentation | The goal of such documentation is to help people get started with tracking
Relates to [#103](https://github.com/eic/EICrecon/issues/103) | 1.0 | Add documentation for ACTS tracking part - The goal of such documentation is to help people get started with tracking
Relates to [#103](https://github.com/eic/EICrecon/issues/103) | non_defect | add documentation for acts tracking part the goal of such documentation is to help people get started with tracking relates to | 0 |
44,135 | 7,093,617,466 | IssuesEvent | 2018-01-12 21:18:46 | NOAA-PMEL/LAS | https://api.github.com/repos/NOAA-PMEL/LAS | closed | Document how to create a new variable using an F-TDS data set. | Documentation - Instal P: normal S: normal bug | **Reported by @noaaroland on 28 Jan 2008 15:45 UTC**
We're struggling with defining variables in init scripts and I think a better place for that is to define the variable as a new F-TDS data set. Kevin is having some luck doing that by defining some specific averages in the GFDL data collection. I'm assigning this to him so perhaps he can quickly summarize his experiences in the Plone for all to share.
Migrated-From: http://dunkel.pmel.noaa.gov/trac/las/ticket/366 | 1.0 | Document how to create a new variable using an F-TDS data set. - **Reported by @noaaroland on 28 Jan 2008 15:45 UTC**
We're struggling with defining variables in init scripts and I think a better place for that is to define the variable as a new F-TDS data set. Kevin is having some luck doing that by defining some specific averages in the GFDL data collection. I'm assigning this to him so perhaps he can quickly summarize his experiences in the Plone for all to share.
Migrated-From: http://dunkel.pmel.noaa.gov/trac/las/ticket/366 | non_defect | document how to create a new variable using an f tds data set reported by noaaroland on jan utc we re struggling with defining variables in init scripts and i think a better place for that is to define the variable as a new f tds data set kevin is having some luck doing that by defining some specific averages in the gfdl data collection i m assigning this to him so perhaps he can quickly summarize his experiences in the plone for all to share migrated from | 0 |
79,349 | 28,118,873,695 | IssuesEvent | 2023-03-31 12:54:07 | cython/cython | https://api.github.com/repos/cython/cython | opened | [BUG] "pure_cdef_class_dataclass" test fails in Py3.12 | defect | ### Describe the bug
The `_DataclassParams` class apparently changed its constructor.
```
======================================================================
ERROR: runTest (__main__.CythonRunTestCase.runTest)
[2] compiling (c/cy2) and running pure_cdef_class_dataclass
----------------------------------------------------------------------
Traceback (most recent call last):
File "tests/run/pure_cdef_class_dataclass.py", line 10, in init pure_cdef_class_dataclass (pure_cdef_class_dataclass.c:8809)
class MyDataclass:
TypeError: _DataclassParams.__init__() missing 4 required positional arguments: 'match_args', 'kw_only', 'slots', and 'weakref_slot'
```
### Code to reproduce the behaviour:
_No response_
### Expected behaviour
_No response_
### OS
_No response_
### Python version
Py3.12a6
### Cython version
3.0.0b2
### Additional context
_No response_ | 1.0 | [BUG] "pure_cdef_class_dataclass" test fails in Py3.12 - ### Describe the bug
The `_DataclassParams` class apparently changed its constructor.
```
======================================================================
ERROR: runTest (__main__.CythonRunTestCase.runTest)
[2] compiling (c/cy2) and running pure_cdef_class_dataclass
----------------------------------------------------------------------
Traceback (most recent call last):
File "tests/run/pure_cdef_class_dataclass.py", line 10, in init pure_cdef_class_dataclass (pure_cdef_class_dataclass.c:8809)
class MyDataclass:
TypeError: _DataclassParams.__init__() missing 4 required positional arguments: 'match_args', 'kw_only', 'slots', and 'weakref_slot'
```
### Code to reproduce the behaviour:
_No response_
### Expected behaviour
_No response_
### OS
_No response_
### Python version
Py3.12a6
### Cython version
3.0.0b2
### Additional context
_No response_ | defect | pure cdef class dataclass test fails in describe the bug the dataclassparams class apparently changed its constructor error runtest main cythonruntestcase runtest compiling c and running pure cdef class dataclass traceback most recent call last file tests run pure cdef class dataclass py line in init pure cdef class dataclass pure cdef class dataclass c class mydataclass typeerror dataclassparams init missing required positional arguments match args kw only slots and weakref slot code to reproduce the behaviour no response expected behaviour no response os no response python version cython version additional context no response | 1 |
9,592 | 2,615,163,225 | IssuesEvent | 2015-03-01 06:42:47 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | opened | Error 0x04 constantly repeating without success. | auto-migrated Priority-Triage Type-Defect | ```
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
Last SVN (1.4)
1. What operating system are you using (Linux is the only supported OS)?
Linux, Debian
2. Is your wireless card in monitor mode (yes/no)?
yes
3. What is the signal strength of the Access Point you are trying to crack?
-65 (ACKS are coming)
4. What is the manufacturer and model # of the device you are trying to
crack?
It seems to be a Netgear Router
5. What is the entire command line string you are supplying to reaver?
reaver -i wlan0 -vv -b 74:31:70:18:BC:9A -c 11 -w
6. Please describe what you think the issue is.
I have no Idea.
7. Paste the output from Reaver below.
[+] Switching wlan0 to channel 11
[+] Waiting for beacon from 74:31:70:18:BC:9A
[+] Associated with 74:31:70:18:BC:9A (ESSID: glimglim)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M1 message
[+] Received M1 message
[+] Received M1 message
[+] Received WSC NACK
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M1 message
[+] Received M1 message
[+] Received M1 message
[+] Received WSC NACK
[+] Sending WSC NACK
Hello, i have been trying reaver with some netgear router.
I get constantly a 0x04 error, and the M-Packages seems to be sended out of
order.
I tried every possible combination, i never get a success.
I have made a tcpdump, available here:
http://www.xup.in/dl,18777570/broken_reaver.pcap/
Could you please have a look into it and tell me what to do.
Thanks in advance
```
Original issue reported on code.google.com by `inop...@web.de` on 2 Jul 2012 at 10:40 | 1.0 | Error 0x04 constantly repeating without success. - ```
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
Last SVN (1.4)
1. What operating system are you using (Linux is the only supported OS)?
Linux, Debian
2. Is your wireless card in monitor mode (yes/no)?
yes
3. What is the signal strength of the Access Point you are trying to crack?
-65 (ACKS are coming)
4. What is the manufacturer and model # of the device you are trying to
crack?
It seems to be a Netgear Router
5. What is the entire command line string you are supplying to reaver?
reaver -i wlan0 -vv -b 74:31:70:18:BC:9A -c 11 -w
6. Please describe what you think the issue is.
I have no Idea.
7. Paste the output from Reaver below.
[+] Switching wlan0 to channel 11
[+] Waiting for beacon from 74:31:70:18:BC:9A
[+] Associated with 74:31:70:18:BC:9A (ESSID: glimglim)
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M1 message
[+] Received M1 message
[+] Received M1 message
[+] Received WSC NACK
[+] Sending WSC NACK
[!] WPS transaction failed (code: 0x04), re-trying last pin
[+] Trying pin 12345670
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M1 message
[+] Received M1 message
[+] Received M1 message
[+] Received WSC NACK
[+] Sending WSC NACK
Hello, i have been trying reaver with some netgear router.
I get constantly a 0x04 error, and the M-Packages seems to be sended out of
order.
I tried every possible combination, i never get a success.
I have made a tcpdump, available here:
http://www.xup.in/dl,18777570/broken_reaver.pcap/
Could you please have a look into it and tell me what to do.
Thanks in advance
```
Original issue reported on code.google.com by `inop...@web.de` on 2 Jul 2012 at 10:40 | defect | error constantly repeating without success what version of reaver are you using only defects against the latest version will be considered last svn what operating system are you using linux is the only supported os linux debian is your wireless card in monitor mode yes no yes what is the signal strength of the access point you are trying to crack acks are coming what is the manufacturer and model of the device you are trying to crack it seems to be a netgear router what is the entire command line string you are supplying to reaver reaver i vv b bc c w please describe what you think the issue is i have no idea paste the output from reaver below switching to channel waiting for beacon from bc associated with bc essid glimglim trying pin sending eapol start request received identity request sending identity response received identity request sending identity response received message sending message received message received message received message received wsc nack sending wsc nack wps transaction failed code re trying last pin trying pin sending eapol start request received identity request sending identity response received identity request sending identity response received message sending message received message received message received message received wsc nack sending wsc nack hello i have been trying reaver with some netgear router i get constantly a error and the m packages seems to be sended out of order i tried every possible combination i never get a success i have made a tcpdump available here could you please have a look into it and tell me what to do thanks in advance original issue reported on code google com by inop web de on jul at | 1 |
46,388 | 19,121,885,299 | IssuesEvent | 2021-12-01 00:08:08 | dockstore/dockstore | https://api.github.com/repos/dockstore/dockstore | opened | Be able to search one's own entries | enhancement web-service gui | **Is your feature request related to a problem? Please describe.**
It's hard to search through one's own entries on Dockstore. If I vaguely remember some code I wrote but I'm not sure where it is, it's not easy to find in the Dockstore UI.
**Describe the solution you'd like**
The My <Workflows|Tools|Services> page(s) should have search functionality that searches through the source files of all of your entries. Bonus could be advanced options, like only search specific orgs, specific file types, regex, glob, etc.
**Describe alternatives you've considered**
Leave it as is. Search for it in GitHub instead and work your ways backwards. But sometimes you belong to a several orgs in GitHub and you're not sure where to start. And it doesn't handle hosted workflows.
**Additional context**
My actual, real-world use case for this was lame; I was looking for descriptors with http imports to verify a bug, which presumably a "real" bioinformatician wouldn't need to do. But I'm sure there are other things where grep-like functionality is useful.
| 1.0 | Be able to search one's own entries - **Is your feature request related to a problem? Please describe.**
It's hard to search through one's own entries on Dockstore. If I vaguely remember some code I wrote but I'm not sure where it is, it's not easy to find in the Dockstore UI.
**Describe the solution you'd like**
The My <Workflows|Tools|Services> page(s) should have search functionality that searches through the source files of all of your entries. Bonus could be advanced options, like only search specific orgs, specific file types, regex, glob, etc.
**Describe alternatives you've considered**
Leave it as is. Search for it in GitHub instead and work your ways backwards. But sometimes you belong to a several orgs in GitHub and you're not sure where to start. And it doesn't handle hosted workflows.
**Additional context**
My actual, real-world use case for this was lame; I was looking for descriptors with http imports to verify a bug, which presumably a "real" bioinformatician wouldn't need to do. But I'm sure there are other things where grep-like functionality is useful.
| non_defect | be able to search one s own entries is your feature request related to a problem please describe it s hard to search through one s own entries on dockstore if i vaguely remember some code i wrote but i m not sure where it is it s not easy to find in the dockstore ui describe the solution you d like the my page s should have search functionality that searches through the source files of all of your entries bonus could be advanced options like only search specific orgs specific file types regex glob etc describe alternatives you ve considered leave it as is search for it in github instead and work your ways backwards but sometimes you belong to a several orgs in github and you re not sure where to start and it doesn t handle hosted workflows additional context my actual real world use case for this was lame i was looking for descriptors with http imports to verify a bug which presumably a real bioinformatician wouldn t need to do but i m sure there are other things where grep like functionality is useful | 0 |
40,090 | 9,842,248,710 | IssuesEvent | 2019-06-18 08:52:36 | line/centraldogma | https://api.github.com/repos/line/centraldogma | closed | Support proxy protocol | defect | There is no way to specify trusted proxy address in the configuration, which means that the client address won't be logged on access log even if proxy protocol is specified as a supported protocol. It should be able to be configured when building a server with a `ServerBuilder`. | 1.0 | Support proxy protocol - There is no way to specify trusted proxy address in the configuration, which means that the client address won't be logged on access log even if proxy protocol is specified as a supported protocol. It should be able to be configured when building a server with a `ServerBuilder`. | defect | support proxy protocol there is no way to specify trusted proxy address in the configuration which means that the client address won t be logged on access log even if proxy protocol is specified as a supported protocol it should be able to be configured when building a server with a serverbuilder | 1 |
467,558 | 13,450,587,636 | IssuesEvent | 2020-09-08 18:46:58 | center-for-knowledge-communication/mathspring | https://api.github.com/repos/center-for-knowledge-communication/mathspring | opened | Security Breach for emails coming from mathspring@cs.umass.edu | High Priority Super High Priority | Hi folks, there is a security breach. A hacker has found the password for mathspring@cs.umass.edu, and is sending emails asking for further access to CS servers. I also received an email about a potential security breach from github/mathspring. Maybe a hacker found a way to get into github?
The email coming from mathspring@cs.umass.edu is asking for access to swarm2.cs.umass.edu
I have asked system@cs.umass.edu to change the password of the email address until we get this sorted out.
Our code is generating emails for password recovery purposes. I am trying to find out if the password is explicitly written in the java code files. | 2.0 | Security Breach for emails coming from mathspring@cs.umass.edu - Hi folks, there is a security breach. A hacker has found the password for mathspring@cs.umass.edu, and is sending emails asking for further access to CS servers. I also received an email about a potential security breach from github/mathspring. Maybe a hacker found a way to get into github?
The email coming from mathspring@cs.umass.edu is asking for access to swarm2.cs.umass.edu
I have asked system@cs.umass.edu to change the password of the email address until we get this sorted out.
Our code is generating emails for password recovery purposes. I am trying to find out if the password is explicitly written in the java code files. | non_defect | security breach for emails coming from mathspring cs umass edu hi folks there is a security breach a hacker has found the password for mathspring cs umass edu and is sending emails asking for further access to cs servers i also received an email about a potential security breach from github mathspring maybe a hacker found a way to get into github the email coming from mathspring cs umass edu is asking for access to cs umass edu i have asked system cs umass edu to change the password of the email address until we get this sorted out our code is generating emails for password recovery purposes i am trying to find out if the password is explicitly written in the java code files | 0 |
56,647 | 15,256,005,721 | IssuesEvent | 2021-02-20 18:22:15 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | zfs 2.0.2 causes system hangs under load, emits kernel bugs | Status: Triage Needed Type: Defect | <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Debian
Distribution Version | Bullseye (Testing)
Linux Kernel | 5.10.0-3-amd64
Architecture | x86_64
ZFS Version | 2.0.2-1
SPL Version | 2.0.2-1
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
The pool is single raidz2 vdev and was recently upgraded from 0.8.5. Currently no cache device attached, there was one before the upgrade. No dedup. The last scrub a few weeks before the upgrade passed without errors.
### Describe the problem you're observing
The system hangs under load. Sometimes only individual processes accessing the pool hang, sometimes I can't even ssh into the machine.
### Describe how to reproduce the problem
Unknown, the pool is accessible and under a mixed read/write use for minutes to hours until varying kernel errors pop up and it starts becoming partially or entirely unresponsive.
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
dmesg output:
```
[ 1578.930649] ------------[ cut here ]------------
[ 1578.930655] WARNING: CPU: 0 PID: 2660656 at kernel/rcu/tasks.h:1118 exit_tasks_rcu_finish+0x35/0x50
[ 1578.930656] Modules linked in: binfmt_misc bridge stp llc nft_counter xt_CT nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 nft_compat nf_tables nfnetlink intel_rapl_msr intel_rapl_common at24 iTCO_wdt intel_pmc_bxt iTCO_vendor_support watchdog x86_pkg_temp_thermal intel_powerclamp coretemp evdev zfs(POE) zunicode(POE) zzstd(OE) kvm_intel zlua(OE) snd_hda_codec_realtek kvm zavl(POE) snd_hda_codec_generic ledtrig_audio icp(POE) irqbypass rapl intel_cstate intel_uncore zcommon(POE) znvpair(POE) i915 snd_hda_intel snd_intel_dspcfg soundwire_intel soundwire_generic_allocation snd_soc_core mxm_wmi spl(OE) efi_pstore pcspkr snd_compress soundwire_cadence snd_hda_codec i2c_i801 i2c_smbus snd_hda_core snd_hwdep soundwire_bus igb snd_pcm mei_me dca sg e1000e snd_timer mei drm_kms_helper ptp snd pps_core soundcore lpc_ich cec i2c_algo_bit button nls_ascii nls_cp437 vfat fat sch_fq tcp_bbr sunrpc drm fuse configfs ip_tables x_tables sd_mod t10_pi crc_t10dif crct10dif_generic crct10dif_pclmul
[ 1578.930722] crct10dif_common crc32_pclmul ghash_clmulni_intel ahci cryptd libahci libata scsi_mod xhci_pci xhci_hcd ehci_pci ehci_hcd usbcore usb_common wmi video btrfs blake2b_generic libcrc32c crc32c_generic crc32c_intel xor raid6_pq loop msr efivarfs autofs4
[ 1578.930738] CPU: 0 PID: 2660656 Comm: z_rd_int Tainted: P OE 5.10.0-3-amd64 #1 Debian 5.10.13-1
[ 1578.930739] Hardware name: To Be Filled By O.E.M. To Be Filled By O.E.M./Z87 Extreme6, BIOS P2.20 11/25/2013
[ 1578.930741] RIP: 0010:exit_tasks_rcu_finish+0x35/0x50
[ 1578.930743] Code: 7b 01 00 8b 87 78 07 00 00 c6 87 84 07 00 00 01 85 c0 75 16 c7 87 78 07 00 00 00 00 00 00 0f b6 87 81 07 00 00 84 c0 75 05 c3 <0f> 0b eb e6 0f 0b 31 f6 e9 be f1 ff ff cc cc cc cc cc cc cc cc cc
[ 1578.930744] RSP: 0018:ffffa9140b347e98 EFLAGS: 00010206
[ 1578.930746] RAX: 0000000000008900 RBX: ffff8d5e18185a00 RCX: 0000000000073293
[ 1578.930747] RDX: 0000000000073292 RSI: ffffffff9ea8bad3 RDI: ffff8d5e18185a00
[ 1578.930747] RBP: ffffa9140b347ec8 R08: ffff8d5a666da3c8 R09: 0000000000000000
[ 1578.930748] R10: 0000000000000000 R11: ffff8d5dd2124b01 R12: ffffa9140b3475b8
[ 1578.930749] R13: ffffa9140b3475b8 R14: ffffa9140b347ec8 R15: ffffffffa0053bc0
[ 1578.930750] FS: 0000000000000000(0000) GS:ffff8d5fff600000(0000) knlGS:0000000000000000
[ 1578.930751] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 1578.930752] CR2: 00007f27b2fbb280 CR3: 000000046f40a001 CR4: 00000000000706f0
[ 1578.930752] Call Trace:
[ 1578.930756] do_exit+0x716/0xab0
[ 1578.930765] ? taskq_thread_spawn+0x50/0x50 [spl]
[ 1578.930767] kthread+0xe6/0x140
[ 1578.930769] ? __kthread_bind_mask+0x60/0x60
[ 1578.930772] ret_from_fork+0x22/0x30
[ 1578.930773] ---[ end trace 74707b7c78abf030 ]---
[ 1583.119018] list_del corruption. prev->next should be ffff8d5de08f7650, but was ffff8d5de08f7f50
[ 1583.119055] ------------[ cut here ]------------
[ 1583.119056] kernel BUG at lib/list_debug.c:51!
[ 1583.119076] invalid opcode: 0000 [#1] SMP PTI
[ 1583.119094] CPU: 1 PID: 2664199 Comm: z_rd_int Tainted: P W OE 5.10.0-3-amd64 #1 Debian 5.10.13-1
[ 1583.119124] Hardware name: To Be Filled By O.E.M. To Be Filled By O.E.M./Z87 Extreme6, BIOS P2.20 11/25/2013
[ 1583.119158] RIP: 0010:__list_del_entry_valid.cold+0x31/0x47
[ 1583.119178] Code: b4 b1 9f e8 90 16 ff ff 0f 0b 48 c7 c7 48 b5 b1 9f e8 82 16 ff ff 0f 0b 48 89 f2 48 89 fe 48 c7 c7 08 b5 b1 9f e8 6e 16 ff ff <0f> 0b 48 89 fe 4c 89 c2 48 c7 c7 d0 b4 b1 9f e8 5a 16 ff ff 0f 0b
[ 1583.119235] RSP: 0018:ffffa9140f6d3d10 EFLAGS: 00010246
[ 1583.119252] RAX: 0000000000000054 RBX: ffff8d5e18eee220 RCX: 0000000000000000
[ 1583.119273] RDX: 0000000000000000 RSI: ffff8d5fff718a00 RDI: ffff8d5fff718a00
[ 1583.119296] RBP: ffff8d5ddf7b2be0 R08: 0000000000000000 R09: ffffa9140f6d3b38
[ 1583.119316] R10: ffffa9140f6d3b30 R11: ffffffffa00cb368 R12: ffff8d5de08f7630
[ 1583.119335] R13: ffff8d5e18eee5e8 R14: ffff8d5ddf7b2fa8 R15: ffff8d5de08f7650
[ 1583.119360] FS: 0000000000000000(0000) GS:ffff8d5fff700000(0000) knlGS:0000000000000000
[ 1583.119386] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 1583.119405] CR2: 00007f273e944258 CR3: 000000015a926004 CR4: 00000000000706e0
[ 1583.119429] Call Trace:
[ 1583.119500] zio_remove_child+0x5f/0x140 [zfs]
[ 1583.119562] zio_done+0x4c6/0x1150 [zfs]
[ 1583.119618] zio_execute+0x81/0x120 [zfs]
[ 1583.119638] taskq_thread+0x2da/0x520 [spl]
[ 1583.119656] ? wake_up_q+0xa0/0xa0
[ 1583.119711] ? zio_destroy+0xf0/0xf0 [zfs]
[ 1583.119730] ? taskq_thread_spawn+0x50/0x50 [spl]
[ 1583.119746] kthread+0x11b/0x140
[ 1583.119758] ? __kthread_bind_mask+0x60/0x60
[ 1583.119776] ret_from_fork+0x22/0x30
[ 1583.119791] Modules linked in: binfmt_misc bridge stp llc nft_counter xt_CT nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 nft_compat nf_tables nfnetlink intel_rapl_msr intel_rapl_common at24 iTCO_wdt intel_pmc_bxt iTCO_vendor_support watchdog x86_pkg_temp_thermal intel_powerclamp coretemp evdev zfs(POE) zunicode(POE) zzstd(OE) kvm_intel zlua(OE) snd_hda_codec_realtek kvm zavl(POE) snd_hda_codec_generic ledtrig_audio icp(POE) irqbypass rapl intel_cstate intel_uncore zcommon(POE) znvpair(POE) i915 snd_hda_intel snd_intel_dspcfg soundwire_intel soundwire_generic_allocation snd_soc_core mxm_wmi spl(OE) efi_pstore pcspkr snd_compress soundwire_cadence snd_hda_codec i2c_i801 i2c_smbus snd_hda_core snd_hwdep soundwire_bus igb snd_pcm mei_me dca sg e1000e snd_timer mei drm_kms_helper ptp snd pps_core soundcore lpc_ich cec i2c_algo_bit button nls_ascii nls_cp437 vfat fat sch_fq tcp_bbr sunrpc drm fuse configfs ip_tables x_tables sd_mod t10_pi crc_t10dif crct10dif_generic crct10dif_pclmul
[ 1583.119839] crct10dif_common crc32_pclmul ghash_clmulni_intel ahci cryptd libahci libata scsi_mod xhci_pci xhci_hcd ehci_pci ehci_hcd usbcore usb_common wmi video btrfs blake2b_generic libcrc32c crc32c_generic crc32c_intel xor raid6_pq loop msr efivarfs autofs4
[ 1583.120155] ---[ end trace 74707b7c78abf031 ]---
[ 1583.126147] RIP: 0010:__list_del_entry_valid.cold+0x31/0x47
[ 1583.126173] Code: b4 b1 9f e8 90 16 ff ff 0f 0b 48 c7 c7 48 b5 b1 9f e8 82 16 ff ff 0f 0b 48 89 f2 48 89 fe 48 c7 c7 08 b5 b1 9f e8 6e 16 ff ff <0f> 0b 48 89 fe 4c 89 c2 48 c7 c7 d0 b4 b1 9f e8 5a 16 ff ff 0f 0b
[ 1583.126231] RSP: 0018:ffffa9140f6d3d10 EFLAGS: 00010246
[ 1583.127188] RAX: 0000000000000054 RBX: ffff8d5e18eee220 RCX: 0000000000000000
[ 1583.128130] RDX: 0000000000000000 RSI: ffff8d5fff718a00 RDI: ffff8d5fff718a00
[ 1583.129062] RBP: ffff8d5ddf7b2be0 R08: 0000000000000000 R09: ffffa9140f6d3b38
[ 1583.130000] R10: ffffa9140f6d3b30 R11: ffffffffa00cb368 R12: ffff8d5de08f7630
[ 1583.130909] R13: ffff8d5e18eee5e8 R14: ffff8d5ddf7b2fa8 R15: ffff8d5de08f7650
[ 1583.131806] FS: 0000000000000000(0000) GS:ffff8d5fff700000(0000) knlGS:0000000000000000
[ 1583.132714] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 1583.133609] CR2: 00007f273e944258 CR3: 000000015a926004 CR4: 00000000000706e0
client_loop: send disconnect: Broken pipe
```
| 1.0 | zfs 2.0.2 causes system hangs under load, emits kernel bugs - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Debian
Distribution Version | Bullseye (Testing)
Linux Kernel | 5.10.0-3-amd64
Architecture | x86_64
ZFS Version | 2.0.2-1
SPL Version | 2.0.2-1
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
The pool is single raidz2 vdev and was recently upgraded from 0.8.5. Currently no cache device attached, there was one before the upgrade. No dedup. The last scrub a few weeks before the upgrade passed without errors.
### Describe the problem you're observing
The system hangs under load. Sometimes only individual processes accessing the pool hang, sometimes I can't even ssh into the machine.
### Describe how to reproduce the problem
Unknown, the pool is accessible and under a mixed read/write use for minutes to hours until varying kernel errors pop up and it starts becoming partially or entirely unresponsive.
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
dmesg output:
```
[ 1578.930649] ------------[ cut here ]------------
[ 1578.930655] WARNING: CPU: 0 PID: 2660656 at kernel/rcu/tasks.h:1118 exit_tasks_rcu_finish+0x35/0x50
[ 1578.930656] Modules linked in: binfmt_misc bridge stp llc nft_counter xt_CT nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 nft_compat nf_tables nfnetlink intel_rapl_msr intel_rapl_common at24 iTCO_wdt intel_pmc_bxt iTCO_vendor_support watchdog x86_pkg_temp_thermal intel_powerclamp coretemp evdev zfs(POE) zunicode(POE) zzstd(OE) kvm_intel zlua(OE) snd_hda_codec_realtek kvm zavl(POE) snd_hda_codec_generic ledtrig_audio icp(POE) irqbypass rapl intel_cstate intel_uncore zcommon(POE) znvpair(POE) i915 snd_hda_intel snd_intel_dspcfg soundwire_intel soundwire_generic_allocation snd_soc_core mxm_wmi spl(OE) efi_pstore pcspkr snd_compress soundwire_cadence snd_hda_codec i2c_i801 i2c_smbus snd_hda_core snd_hwdep soundwire_bus igb snd_pcm mei_me dca sg e1000e snd_timer mei drm_kms_helper ptp snd pps_core soundcore lpc_ich cec i2c_algo_bit button nls_ascii nls_cp437 vfat fat sch_fq tcp_bbr sunrpc drm fuse configfs ip_tables x_tables sd_mod t10_pi crc_t10dif crct10dif_generic crct10dif_pclmul
[ 1578.930722] crct10dif_common crc32_pclmul ghash_clmulni_intel ahci cryptd libahci libata scsi_mod xhci_pci xhci_hcd ehci_pci ehci_hcd usbcore usb_common wmi video btrfs blake2b_generic libcrc32c crc32c_generic crc32c_intel xor raid6_pq loop msr efivarfs autofs4
[ 1578.930738] CPU: 0 PID: 2660656 Comm: z_rd_int Tainted: P OE 5.10.0-3-amd64 #1 Debian 5.10.13-1
[ 1578.930739] Hardware name: To Be Filled By O.E.M. To Be Filled By O.E.M./Z87 Extreme6, BIOS P2.20 11/25/2013
[ 1578.930741] RIP: 0010:exit_tasks_rcu_finish+0x35/0x50
[ 1578.930743] Code: 7b 01 00 8b 87 78 07 00 00 c6 87 84 07 00 00 01 85 c0 75 16 c7 87 78 07 00 00 00 00 00 00 0f b6 87 81 07 00 00 84 c0 75 05 c3 <0f> 0b eb e6 0f 0b 31 f6 e9 be f1 ff ff cc cc cc cc cc cc cc cc cc
[ 1578.930744] RSP: 0018:ffffa9140b347e98 EFLAGS: 00010206
[ 1578.930746] RAX: 0000000000008900 RBX: ffff8d5e18185a00 RCX: 0000000000073293
[ 1578.930747] RDX: 0000000000073292 RSI: ffffffff9ea8bad3 RDI: ffff8d5e18185a00
[ 1578.930747] RBP: ffffa9140b347ec8 R08: ffff8d5a666da3c8 R09: 0000000000000000
[ 1578.930748] R10: 0000000000000000 R11: ffff8d5dd2124b01 R12: ffffa9140b3475b8
[ 1578.930749] R13: ffffa9140b3475b8 R14: ffffa9140b347ec8 R15: ffffffffa0053bc0
[ 1578.930750] FS: 0000000000000000(0000) GS:ffff8d5fff600000(0000) knlGS:0000000000000000
[ 1578.930751] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 1578.930752] CR2: 00007f27b2fbb280 CR3: 000000046f40a001 CR4: 00000000000706f0
[ 1578.930752] Call Trace:
[ 1578.930756] do_exit+0x716/0xab0
[ 1578.930765] ? taskq_thread_spawn+0x50/0x50 [spl]
[ 1578.930767] kthread+0xe6/0x140
[ 1578.930769] ? __kthread_bind_mask+0x60/0x60
[ 1578.930772] ret_from_fork+0x22/0x30
[ 1578.930773] ---[ end trace 74707b7c78abf030 ]---
[ 1583.119018] list_del corruption. prev->next should be ffff8d5de08f7650, but was ffff8d5de08f7f50
[ 1583.119055] ------------[ cut here ]------------
[ 1583.119056] kernel BUG at lib/list_debug.c:51!
[ 1583.119076] invalid opcode: 0000 [#1] SMP PTI
[ 1583.119094] CPU: 1 PID: 2664199 Comm: z_rd_int Tainted: P W OE 5.10.0-3-amd64 #1 Debian 5.10.13-1
[ 1583.119124] Hardware name: To Be Filled By O.E.M. To Be Filled By O.E.M./Z87 Extreme6, BIOS P2.20 11/25/2013
[ 1583.119158] RIP: 0010:__list_del_entry_valid.cold+0x31/0x47
[ 1583.119178] Code: b4 b1 9f e8 90 16 ff ff 0f 0b 48 c7 c7 48 b5 b1 9f e8 82 16 ff ff 0f 0b 48 89 f2 48 89 fe 48 c7 c7 08 b5 b1 9f e8 6e 16 ff ff <0f> 0b 48 89 fe 4c 89 c2 48 c7 c7 d0 b4 b1 9f e8 5a 16 ff ff 0f 0b
[ 1583.119235] RSP: 0018:ffffa9140f6d3d10 EFLAGS: 00010246
[ 1583.119252] RAX: 0000000000000054 RBX: ffff8d5e18eee220 RCX: 0000000000000000
[ 1583.119273] RDX: 0000000000000000 RSI: ffff8d5fff718a00 RDI: ffff8d5fff718a00
[ 1583.119296] RBP: ffff8d5ddf7b2be0 R08: 0000000000000000 R09: ffffa9140f6d3b38
[ 1583.119316] R10: ffffa9140f6d3b30 R11: ffffffffa00cb368 R12: ffff8d5de08f7630
[ 1583.119335] R13: ffff8d5e18eee5e8 R14: ffff8d5ddf7b2fa8 R15: ffff8d5de08f7650
[ 1583.119360] FS: 0000000000000000(0000) GS:ffff8d5fff700000(0000) knlGS:0000000000000000
[ 1583.119386] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 1583.119405] CR2: 00007f273e944258 CR3: 000000015a926004 CR4: 00000000000706e0
[ 1583.119429] Call Trace:
[ 1583.119500] zio_remove_child+0x5f/0x140 [zfs]
[ 1583.119562] zio_done+0x4c6/0x1150 [zfs]
[ 1583.119618] zio_execute+0x81/0x120 [zfs]
[ 1583.119638] taskq_thread+0x2da/0x520 [spl]
[ 1583.119656] ? wake_up_q+0xa0/0xa0
[ 1583.119711] ? zio_destroy+0xf0/0xf0 [zfs]
[ 1583.119730] ? taskq_thread_spawn+0x50/0x50 [spl]
[ 1583.119746] kthread+0x11b/0x140
[ 1583.119758] ? __kthread_bind_mask+0x60/0x60
[ 1583.119776] ret_from_fork+0x22/0x30
[ 1583.119791] Modules linked in: binfmt_misc bridge stp llc nft_counter xt_CT nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 nft_compat nf_tables nfnetlink intel_rapl_msr intel_rapl_common at24 iTCO_wdt intel_pmc_bxt iTCO_vendor_support watchdog x86_pkg_temp_thermal intel_powerclamp coretemp evdev zfs(POE) zunicode(POE) zzstd(OE) kvm_intel zlua(OE) snd_hda_codec_realtek kvm zavl(POE) snd_hda_codec_generic ledtrig_audio icp(POE) irqbypass rapl intel_cstate intel_uncore zcommon(POE) znvpair(POE) i915 snd_hda_intel snd_intel_dspcfg soundwire_intel soundwire_generic_allocation snd_soc_core mxm_wmi spl(OE) efi_pstore pcspkr snd_compress soundwire_cadence snd_hda_codec i2c_i801 i2c_smbus snd_hda_core snd_hwdep soundwire_bus igb snd_pcm mei_me dca sg e1000e snd_timer mei drm_kms_helper ptp snd pps_core soundcore lpc_ich cec i2c_algo_bit button nls_ascii nls_cp437 vfat fat sch_fq tcp_bbr sunrpc drm fuse configfs ip_tables x_tables sd_mod t10_pi crc_t10dif crct10dif_generic crct10dif_pclmul
[ 1583.119839] crct10dif_common crc32_pclmul ghash_clmulni_intel ahci cryptd libahci libata scsi_mod xhci_pci xhci_hcd ehci_pci ehci_hcd usbcore usb_common wmi video btrfs blake2b_generic libcrc32c crc32c_generic crc32c_intel xor raid6_pq loop msr efivarfs autofs4
[ 1583.120155] ---[ end trace 74707b7c78abf031 ]---
[ 1583.126147] RIP: 0010:__list_del_entry_valid.cold+0x31/0x47
[ 1583.126173] Code: b4 b1 9f e8 90 16 ff ff 0f 0b 48 c7 c7 48 b5 b1 9f e8 82 16 ff ff 0f 0b 48 89 f2 48 89 fe 48 c7 c7 08 b5 b1 9f e8 6e 16 ff ff <0f> 0b 48 89 fe 4c 89 c2 48 c7 c7 d0 b4 b1 9f e8 5a 16 ff ff 0f 0b
[ 1583.126231] RSP: 0018:ffffa9140f6d3d10 EFLAGS: 00010246
[ 1583.127188] RAX: 0000000000000054 RBX: ffff8d5e18eee220 RCX: 0000000000000000
[ 1583.128130] RDX: 0000000000000000 RSI: ffff8d5fff718a00 RDI: ffff8d5fff718a00
[ 1583.129062] RBP: ffff8d5ddf7b2be0 R08: 0000000000000000 R09: ffffa9140f6d3b38
[ 1583.130000] R10: ffffa9140f6d3b30 R11: ffffffffa00cb368 R12: ffff8d5de08f7630
[ 1583.130909] R13: ffff8d5e18eee5e8 R14: ffff8d5ddf7b2fa8 R15: ffff8d5de08f7650
[ 1583.131806] FS: 0000000000000000(0000) GS:ffff8d5fff700000(0000) knlGS:0000000000000000
[ 1583.132714] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 1583.133609] CR2: 00007f273e944258 CR3: 000000015a926004 CR4: 00000000000706e0
client_loop: send disconnect: Broken pipe
```
| defect | zfs causes system hangs under load emits kernel bugs thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name debian distribution version bullseye testing linux kernel architecture zfs version spl version commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version the pool is single vdev and was recently upgraded from currently no cache device attached there was one before the upgrade no dedup the last scrub a few weeks before the upgrade passed without errors describe the problem you re observing the system hangs under load sometimes only individual processes accessing the pool hang sometimes i can t even ssh into the machine describe how to reproduce the problem unknown the pool is accessible and under a mixed read write use for minutes to hours until varying kernel errors pop up and it starts becoming partially or entirely unresponsive include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with dmesg output warning cpu pid at kernel rcu tasks h exit tasks rcu finish modules linked in binfmt misc bridge stp llc nft counter xt ct nf conntrack nf defrag nf defrag nft compat nf tables nfnetlink intel rapl msr intel rapl common itco wdt intel pmc bxt itco vendor support watchdog pkg temp thermal intel powerclamp coretemp evdev zfs poe zunicode poe zzstd oe kvm intel zlua oe snd hda codec realtek kvm zavl poe snd hda codec generic ledtrig audio icp poe irqbypass rapl intel cstate intel uncore zcommon poe znvpair poe snd hda intel snd intel dspcfg soundwire intel soundwire generic allocation snd soc core mxm wmi spl oe efi pstore pcspkr snd compress soundwire cadence snd hda codec smbus snd hda core snd hwdep soundwire bus igb snd pcm mei me dca sg snd timer mei drm kms helper ptp snd pps core soundcore lpc ich cec algo bit button nls ascii nls vfat fat sch fq tcp bbr sunrpc drm fuse configfs ip tables x tables sd mod pi crc generic pclmul common pclmul ghash clmulni intel ahci cryptd libahci libata scsi mod xhci pci xhci hcd ehci pci ehci hcd usbcore usb common wmi video btrfs generic generic intel xor pq loop msr efivarfs cpu pid comm z rd int tainted p oe debian hardware name to be filled by o e m to be filled by o e m bios rip exit tasks rcu finish code eb be ff ff cc cc cc cc cc cc cc cc cc rsp eflags rax rbx rcx rdx rsi rdi rbp fs gs knlgs cs ds es call trace do exit taskq thread spawn kthread kthread bind mask ret from fork list del corruption prev next should be but was kernel bug at lib list debug c invalid opcode smp pti cpu pid comm z rd int tainted p w oe debian hardware name to be filled by o e m to be filled by o e m bios rip list del entry valid cold code ff ff ff ff fe ff ff fe ff ff rsp eflags rax rbx rcx rdx rsi rdi rbp fs gs knlgs cs ds es call trace zio remove child zio done zio execute taskq thread wake up q zio destroy taskq thread spawn kthread kthread bind mask ret from fork modules linked in binfmt misc bridge stp llc nft counter xt ct nf conntrack nf defrag nf defrag nft compat nf tables nfnetlink intel rapl msr intel rapl common itco wdt intel pmc bxt itco vendor support watchdog pkg temp thermal intel powerclamp coretemp evdev zfs poe zunicode poe zzstd oe kvm intel zlua oe snd hda codec realtek kvm zavl poe snd hda codec generic ledtrig audio icp poe irqbypass rapl intel cstate intel uncore zcommon poe znvpair poe snd hda intel snd intel dspcfg soundwire intel soundwire generic allocation snd soc core mxm wmi spl oe efi pstore pcspkr snd compress soundwire cadence snd hda codec smbus snd hda core snd hwdep soundwire bus igb snd pcm mei me dca sg snd timer mei drm kms helper ptp snd pps core soundcore lpc ich cec algo bit button nls ascii nls vfat fat sch fq tcp bbr sunrpc drm fuse configfs ip tables x tables sd mod pi crc generic pclmul common pclmul ghash clmulni intel ahci cryptd libahci libata scsi mod xhci pci xhci hcd ehci pci ehci hcd usbcore usb common wmi video btrfs generic generic intel xor pq loop msr efivarfs rip list del entry valid cold code ff ff ff ff fe ff ff fe ff ff rsp eflags rax rbx rcx rdx rsi rdi rbp fs gs knlgs cs ds es client loop send disconnect broken pipe | 1 |
421,822 | 28,360,514,005 | IssuesEvent | 2023-04-12 10:22:54 | audeering/audb | https://api.github.com/repos/audeering/audb | closed | Be more precise for return types in audb.info | documentation enhancement | For functions like `audb.info.tables()` or `audb.info.schemes()` we only state that they return a `Dict`:

But we could be more precise here stating:
```python
typing.Dict[str, audformat.Table]
```
etc.
See also https://github.com/audeering/audb/pull/276#pullrequestreview-1380760464 | 1.0 | Be more precise for return types in audb.info - For functions like `audb.info.tables()` or `audb.info.schemes()` we only state that they return a `Dict`:

But we could be more precise here stating:
```python
typing.Dict[str, audformat.Table]
```
etc.
See also https://github.com/audeering/audb/pull/276#pullrequestreview-1380760464 | non_defect | be more precise for return types in audb info for functions like audb info tables or audb info schemes we only state that they return a dict but we could be more precise here stating python typing dict etc see also | 0 |
21,305 | 14,517,748,140 | IssuesEvent | 2020-12-13 20:50:35 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | opened | coreclr build failure: "GenerateFileVersionProps" task could not be loaded because installer.tasks.dll not found | area-Infrastructure-coreclr | OS: Windows
Architecture: x86
Configuration: Release
Repro steps (non-deterministic): **build clr+libs+libs.tests -a x86 -c Release**
<pre>
System -> D:\git\runtime\artifacts\bin\manual.System\net6.0-Release\System.dll
System.Data -> D:\git\runtime\artifacts\bin\manual.System.Data\net6.0-Release\System.Data.dll
System.Xml -> D:\git\runtime\artifacts\bin\manual.System.Xml\net6.0-Release\System.Xml.dll
ApiCompat ->
Trimming shared framework assemblies with ILLinker...
externals ->
D:\git\runtime\src\libraries\pretest.proj(55,5): error MSB4062: The "GenerateFileVersionProps" task could not be loaded from the assembly D:\git\runtime\artifacts\bin\installer.tasks\Debug\netstandard2.0\installer.tasks.dll. Could not load file or assembly 'D:\git\runtime\artifacts\bin\installer.tasks\Debug\netstandard2.0\installer.tasks.dll'. The system cannot find the file specified. Confirm that the <UsingTask> declaration is correct, that the assembly and all its dependencies are available, and that the task contains a public class that implements Microsoft.Build.Framework.ITask.
</pre>
| 1.0 | coreclr build failure: "GenerateFileVersionProps" task could not be loaded because installer.tasks.dll not found - OS: Windows
Architecture: x86
Configuration: Release
Repro steps (non-deterministic): **build clr+libs+libs.tests -a x86 -c Release**
<pre>
System -> D:\git\runtime\artifacts\bin\manual.System\net6.0-Release\System.dll
System.Data -> D:\git\runtime\artifacts\bin\manual.System.Data\net6.0-Release\System.Data.dll
System.Xml -> D:\git\runtime\artifacts\bin\manual.System.Xml\net6.0-Release\System.Xml.dll
ApiCompat ->
Trimming shared framework assemblies with ILLinker...
externals ->
D:\git\runtime\src\libraries\pretest.proj(55,5): error MSB4062: The "GenerateFileVersionProps" task could not be loaded from the assembly D:\git\runtime\artifacts\bin\installer.tasks\Debug\netstandard2.0\installer.tasks.dll. Could not load file or assembly 'D:\git\runtime\artifacts\bin\installer.tasks\Debug\netstandard2.0\installer.tasks.dll'. The system cannot find the file specified. Confirm that the <UsingTask> declaration is correct, that the assembly and all its dependencies are available, and that the task contains a public class that implements Microsoft.Build.Framework.ITask.
</pre>
| non_defect | coreclr build failure generatefileversionprops task could not be loaded because installer tasks dll not found os windows architecture configuration release repro steps non deterministic build clr libs libs tests a c release system d git runtime artifacts bin manual system release system dll system data d git runtime artifacts bin manual system data release system data dll system xml d git runtime artifacts bin manual system xml release system xml dll apicompat trimming shared framework assemblies with illinker externals d git runtime src libraries pretest proj error the generatefileversionprops task could not be loaded from the assembly d git runtime artifacts bin installer tasks debug installer tasks dll could not load file or assembly d git runtime artifacts bin installer tasks debug installer tasks dll the system cannot find the file specified confirm that the declaration is correct that the assembly and all its dependencies are available and that the task contains a public class that implements microsoft build framework itask | 0 |
182,340 | 30,833,294,356 | IssuesEvent | 2023-08-02 04:48:46 | pie-sfac/3-14-ketotop | https://api.github.com/repos/pie-sfac/3-14-ketotop | closed | [feat] Component -Tab 컴포넌트 기능개발 | 🔮 Feature 🎨 Design | ## 작업할 내용 간략한 설명
- 상단 Tab 컴포넌트를 지정한 컨벤션에 따라 개발예정입니다.
## 구현 방법 및 예상 동작
- 컴포넌트의 props 로 boolean, instance 를 지정해두면 원하는 UI 를 화면에 렌더를 할수있도록 개발할 예정 입니다.
## 특이사항
- 없습니다.
| 1.0 | [feat] Component -Tab 컴포넌트 기능개발 - ## 작업할 내용 간략한 설명
- 상단 Tab 컴포넌트를 지정한 컨벤션에 따라 개발예정입니다.
## 구현 방법 및 예상 동작
- 컴포넌트의 props 로 boolean, instance 를 지정해두면 원하는 UI 를 화면에 렌더를 할수있도록 개발할 예정 입니다.
## 특이사항
- 없습니다.
| non_defect | component tab 컴포넌트 기능개발 작업할 내용 간략한 설명 상단 tab 컴포넌트를 지정한 컨벤션에 따라 개발예정입니다 구현 방법 및 예상 동작 컴포넌트의 props 로 boolean instance 를 지정해두면 원하는 ui 를 화면에 렌더를 할수있도록 개발할 예정 입니다 특이사항 없습니다 | 0 |
40,936 | 10,232,283,201 | IssuesEvent | 2019-08-18 16:15:33 | swarfer/sketchucam | https://api.github.com/repos/swarfer/sketchucam | closed | Erase tool doesn't remove geometries added by plugin to a model. | Priority-Medium Type-Defect auto-migrated | ```
1. Create a pocket operation
2. Use erase tool to remove operation.
3. Erase tool removes a cut path but doesn't remove geometries added to a model.
```
Original issue reported on code.google.com by `szymansk...@gmail.com` on 27 Feb 2014 at 4:53
Attachments:
- [Screen Shot 2014-02-26 at 8.49.34 PM.png](https://storage.googleapis.com/google-code-attachments/sketchucam/issue-3/comment-0/Screen Shot 2014-02-26 at 8.49.34 PM.png)
| 1.0 | Erase tool doesn't remove geometries added by plugin to a model. - ```
1. Create a pocket operation
2. Use erase tool to remove operation.
3. Erase tool removes a cut path but doesn't remove geometries added to a model.
```
Original issue reported on code.google.com by `szymansk...@gmail.com` on 27 Feb 2014 at 4:53
Attachments:
- [Screen Shot 2014-02-26 at 8.49.34 PM.png](https://storage.googleapis.com/google-code-attachments/sketchucam/issue-3/comment-0/Screen Shot 2014-02-26 at 8.49.34 PM.png)
| defect | erase tool doesn t remove geometries added by plugin to a model create a pocket operation use erase tool to remove operation erase tool removes a cut path but doesn t remove geometries added to a model original issue reported on code google com by szymansk gmail com on feb at attachments shot at pm png | 1 |
2,235 | 2,603,991,510 | IssuesEvent | 2015-02-24 19:06:45 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳治疗男性生殖器疱症 | auto-migrated Priority-Medium Type-Defect | ```
沈阳治疗男性生殖器疱症〓沈陽軍區政治部醫院性病〓TEL:02
4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療�
��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝�
��的歷史悠久、設備精良、技術權威、專家云集,是預防、保
健、醫療、科研康復為一體的綜合性醫院。是國家首批公立��
�等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學�
��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍
空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集��
�二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:32 | 1.0 | 沈阳治疗男性生殖器疱症 - ```
沈阳治疗男性生殖器疱症〓沈陽軍區政治部醫院性病〓TEL:02
4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療�
��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝�
��的歷史悠久、設備精良、技術權威、專家云集,是預防、保
健、醫療、科研康復為一體的綜合性醫院。是國家首批公立��
�等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學�
��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍
空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集��
�二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:32 | defect | 沈阳治疗男性生殖器疱症 沈阳治疗男性生殖器疱症〓沈陽軍區政治部醫院性病〓tel: 〓 , � �� 。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 original issue reported on code google com by gmail com on jun at | 1 |
314,953 | 9,605,079,709 | IssuesEvent | 2019-05-10 22:11:58 | dotkom/onlineweb-frontend | https://api.github.com/repos/dotkom/onlineweb-frontend | opened | Nibble integration in OWF | App: Core Priority: Low Type: Feature | Make it possible to pay in Nibble outside the Online office.
## Use cases
- You do not have your card available when in the office.
- You have forgotten to pay and you are not in the office.
- You really want food, but do not have enough "nibble" cash nor time to refill account and you swear to pay later.
## Possible solutions
- Create a "new" nibble in OWF with the same features - except login (because you are probably logged in with correct user).
- Create a Nibble instance on nibble.online.ntnu.no or something (here you may need login).
- This also adds the possibility to implement a beautiful iframe directly in OWF.
## Potential dangers
Is there any? | 1.0 | Nibble integration in OWF - Make it possible to pay in Nibble outside the Online office.
## Use cases
- You do not have your card available when in the office.
- You have forgotten to pay and you are not in the office.
- You really want food, but do not have enough "nibble" cash nor time to refill account and you swear to pay later.
## Possible solutions
- Create a "new" nibble in OWF with the same features - except login (because you are probably logged in with correct user).
- Create a Nibble instance on nibble.online.ntnu.no or something (here you may need login).
- This also adds the possibility to implement a beautiful iframe directly in OWF.
## Potential dangers
Is there any? | non_defect | nibble integration in owf make it possible to pay in nibble outside the online office use cases you do not have your card available when in the office you have forgotten to pay and you are not in the office you really want food but do not have enough nibble cash nor time to refill account and you swear to pay later possible solutions create a new nibble in owf with the same features except login because you are probably logged in with correct user create a nibble instance on nibble online ntnu no or something here you may need login this also adds the possibility to implement a beautiful iframe directly in owf potential dangers is there any | 0 |
276,413 | 20,982,951,686 | IssuesEvent | 2022-03-28 22:05:37 | danieldevine/bird-elephant | https://api.github.com/repos/danieldevine/bird-elephant | closed | Can a video be uploaded and attached to tweets? | documentation enhancement question | Can a video be uploaded and attached to tweets? Via
`$twitter->tweets()->upload ('./img/video.mp4');`
Or any other ways to achieve this? | 1.0 | Can a video be uploaded and attached to tweets? - Can a video be uploaded and attached to tweets? Via
`$twitter->tweets()->upload ('./img/video.mp4');`
Or any other ways to achieve this? | non_defect | can a video be uploaded and attached to tweets can a video be uploaded and attached to tweets via twitter tweets upload img video or any other ways to achieve this | 0 |
62,623 | 17,096,967,029 | IssuesEvent | 2021-07-09 05:11:58 | SAP/fundamental-ngx | https://api.github.com/repos/SAP/fundamental-ngx | closed | Carousel: Empty State is not aligned with Visual Core | Defect Hunting ariba bug core design_team | **Describe the bug**
Carousel: Empty State is not aligned with Visual Core. Information may not be sufficient (see Ariba UX suggestions below), icon is not proportional in size to the content strings.

> Is this issue related to a specific component?
Carousel empty state
> What version of the Fundamental Library Styles are you using?
v0.25.0
**To Reproduce**
Steps to reproduce the behavior:
1. Go to: https://fundamental-ngx.netlify.app/#/core/carousel
2. Scroll down to: Carousel error message when no item is loaded
3. See error:

**Expected behavior**
Visual Core suggest:

Ariba UX:

**Desktop (please complete the following information):**
- MacOS
- Chrome, Firefox, Safari
- v0.25.0
| 1.0 | Carousel: Empty State is not aligned with Visual Core - **Describe the bug**
Carousel: Empty State is not aligned with Visual Core. Information may not be sufficient (see Ariba UX suggestions below), icon is not proportional in size to the content strings.

> Is this issue related to a specific component?
Carousel empty state
> What version of the Fundamental Library Styles are you using?
v0.25.0
**To Reproduce**
Steps to reproduce the behavior:
1. Go to: https://fundamental-ngx.netlify.app/#/core/carousel
2. Scroll down to: Carousel error message when no item is loaded
3. See error:

**Expected behavior**
Visual Core suggest:

Ariba UX:

**Desktop (please complete the following information):**
- MacOS
- Chrome, Firefox, Safari
- v0.25.0
| defect | carousel empty state is not aligned with visual core describe the bug carousel empty state is not aligned with visual core information may not be sufficient see ariba ux suggestions below icon is not proportional in size to the content strings is this issue related to a specific component carousel empty state what version of the fundamental library styles are you using to reproduce steps to reproduce the behavior go to scroll down to carousel error message when no item is loaded see error expected behavior visual core suggest ariba ux desktop please complete the following information macos chrome firefox safari | 1 |
80,322 | 30,224,714,582 | IssuesEvent | 2023-07-05 22:48:29 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Special symbols in matrix names | T-Defect S-Minor A-Pills O-Uncommon | <!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
### Description
When a user uses special characters in their name it breaks the citing feature. Here is an example of a user that uses two line feed characters in their name which produces the following result in the message input field:

And after posting produces the following message:

The exact JSON with the user details is the following:
```
{
"displayname": "\n\nJonas"
}
```
Interesting note: The API returns the _line feeds_ in the name, but the user claims what they originally inputted was an em-space character. So this is likely the results of an ASCII transformation of the non-ASCII characters (hence his message, intentionally included in the screenshot).
### Steps to reproduce
- Find a user with a hacked name that includes new lines
- Try and cite them
- Be miserable for a second hating Riot and Matrix while they are both really great projects.
- Post the message, then edit it and remove the new lines to produce a nice looking pill.
### Version information
riot: 1.4.0
matrix-react-sdk version: 1.6.0
riot-web version: master
olm version: 3.1.0 | 1.0 | Special symbols in matrix names - <!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
### Description
When a user uses special characters in their name it breaks the citing feature. Here is an example of a user that uses two line feed characters in their name which produces the following result in the message input field:

And after posting produces the following message:

The exact JSON with the user details is the following:
```
{
"displayname": "\n\nJonas"
}
```
Interesting note: The API returns the _line feeds_ in the name, but the user claims what they originally inputted was an em-space character. So this is likely the results of an ASCII transformation of the non-ASCII characters (hence his message, intentionally included in the screenshot).
### Steps to reproduce
- Find a user with a hacked name that includes new lines
- Try and cite them
- Be miserable for a second hating Riot and Matrix while they are both really great projects.
- Post the message, then edit it and remove the new lines to produce a nice looking pill.
### Version information
riot: 1.4.0
matrix-react-sdk version: 1.6.0
riot-web version: master
olm version: 3.1.0 | defect | special symbols in matrix names this is a bug report template by following the instructions below and filling out the sections with your information you will help the us to get all the necessary data to fix your issue you can also preview your report before submitting it you may remove sections that aren t relevant to your particular case text between marks will be invisible in the report description when a user uses special characters in their name it breaks the citing feature here is an example of a user that uses two line feed characters in their name which produces the following result in the message input field and after posting produces the following message the exact json with the user details is the following displayname n njonas interesting note the api returns the line feeds in the name but the user claims what they originally inputted was an em space character so this is likely the results of an ascii transformation of the non ascii characters hence his message intentionally included in the screenshot steps to reproduce find a user with a hacked name that includes new lines try and cite them be miserable for a second hating riot and matrix while they are both really great projects post the message then edit it and remove the new lines to produce a nice looking pill version information riot matrix react sdk version riot web version master olm version | 1 |
46,690 | 13,055,959,953 | IssuesEvent | 2020-07-30 03:14:27 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | [steamshovel] Test failure (Trac #1718) | Incomplete Migration Migrated from Trac combo core defect | Migrated from https://code.icecube.wisc.edu/ticket/1718
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:10",
"description": "{{{\n Start 353: steamshovel::test_shovelscripts.py\n6/6 Test #353: steamshovel::test_shovelscripts.py ...............***Failed 12.77 sec\n...F.\n======================================================================\nFAIL: test_particles_artist_draws_tracks (__main__.TestShovelScripts)\n----------------------------------------------------------------------\nTraceback (most recent call last):\n File \"/home/olivas/icecube/combo/src/steamshovel/resources/test/test_shovelscripts.py\", line 79, in test_particles_artist_draws_tracks\n self.assertImage( name + \".png\", 35 )\n File \"/home/olivas/icecube/combo/src/steamshovel/resources/test/test_shovelscripts.py\", line 46, in assertImage\n self.assertGreater(psnr, threshold)\nAssertionError: 29.780892595095874 not greater than 35\n\n----------------------------------------------------------------------\n}}}",
"reporter": "olivas",
"cc": "david.schultz",
"resolution": "fixed",
"_ts": "1550067190995086",
"component": "combo core",
"summary": "[steamshovel] Test failure",
"priority": "major",
"keywords": "steamshovel X11",
"time": "2016-05-31T17:49:41",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
| 1.0 | [steamshovel] Test failure (Trac #1718) - Migrated from https://code.icecube.wisc.edu/ticket/1718
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:10",
"description": "{{{\n Start 353: steamshovel::test_shovelscripts.py\n6/6 Test #353: steamshovel::test_shovelscripts.py ...............***Failed 12.77 sec\n...F.\n======================================================================\nFAIL: test_particles_artist_draws_tracks (__main__.TestShovelScripts)\n----------------------------------------------------------------------\nTraceback (most recent call last):\n File \"/home/olivas/icecube/combo/src/steamshovel/resources/test/test_shovelscripts.py\", line 79, in test_particles_artist_draws_tracks\n self.assertImage( name + \".png\", 35 )\n File \"/home/olivas/icecube/combo/src/steamshovel/resources/test/test_shovelscripts.py\", line 46, in assertImage\n self.assertGreater(psnr, threshold)\nAssertionError: 29.780892595095874 not greater than 35\n\n----------------------------------------------------------------------\n}}}",
"reporter": "olivas",
"cc": "david.schultz",
"resolution": "fixed",
"_ts": "1550067190995086",
"component": "combo core",
"summary": "[steamshovel] Test failure",
"priority": "major",
"keywords": "steamshovel X11",
"time": "2016-05-31T17:49:41",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
| defect | test failure trac migrated from json status closed changetime description n start steamshovel test shovelscripts py test steamshovel test shovelscripts py failed sec n f n nfail test particles artist draws tracks main testshovelscripts n ntraceback most recent call last n file home olivas icecube combo src steamshovel resources test test shovelscripts py line in test particles artist draws tracks n self assertimage name png n file home olivas icecube combo src steamshovel resources test test shovelscripts py line in assertimage n self assertgreater psnr threshold nassertionerror not greater than n n n reporter olivas cc david schultz resolution fixed ts component combo core summary test failure priority major keywords steamshovel time milestone owner hdembinski type defect | 1 |
285,278 | 21,511,419,140 | IssuesEvent | 2022-04-28 05:13:11 | dusk-network/dusk-blockchain | https://api.github.com/repos/dusk-network/dusk-blockchain | closed | Consolidate all documentation in repository | type:feature type:refactor documentation team:WebTech | **Describe what you want implemented**
Currently documentation is patchy and inconsistently structured. Every package should have a proper, page full (usually) of information that is in README.md. A small script could then find these in each package and based on the folder structure, assemble a single unified document that is easy to update and sites the documentation next to the code. Further, each package's godoc documentation could be derived from this for an auto-generated doc.go.
**Describe "Why" this is needed**
Currently, finding out about everything in Dusk requires an exhaustive patient search. If this was easier, it would be easier to both debug, extend, and, ultimately, translate into Rust.
~~**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.~~
~~**Additional context**
Add any other context or screenshots about the feature request here.~~
| 1.0 | Consolidate all documentation in repository - **Describe what you want implemented**
Currently documentation is patchy and inconsistently structured. Every package should have a proper, page full (usually) of information that is in README.md. A small script could then find these in each package and based on the folder structure, assemble a single unified document that is easy to update and sites the documentation next to the code. Further, each package's godoc documentation could be derived from this for an auto-generated doc.go.
**Describe "Why" this is needed**
Currently, finding out about everything in Dusk requires an exhaustive patient search. If this was easier, it would be easier to both debug, extend, and, ultimately, translate into Rust.
~~**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.~~
~~**Additional context**
Add any other context or screenshots about the feature request here.~~
| non_defect | consolidate all documentation in repository describe what you want implemented currently documentation is patchy and inconsistently structured every package should have a proper page full usually of information that is in readme md a small script could then find these in each package and based on the folder structure assemble a single unified document that is easy to update and sites the documentation next to the code further each package s godoc documentation could be derived from this for an auto generated doc go describe why this is needed currently finding out about everything in dusk requires an exhaustive patient search if this was easier it would be easier to both debug extend and ultimately translate into rust describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here | 0 |
76,255 | 26,333,432,099 | IssuesEvent | 2023-01-10 12:37:45 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | BUG: in version 1.10 No matching signature found when trying to use scipy.interpolate.RegularGridInterpolator | defect | ### Describe your issue.
with scipy1.10 scipy.interpolate.RegularGridInterpolator object gives the following error message when the __call__ method is used:
### Reproducing Code Example
```python
import scipy
import scipy.interpolate
import numpy as np
foo = scipy.interpolate.RegularGridInterpolator(
[np.arange(4), np.arange(5)],
np.zeros([4,5],"f"),
bounds_error=False,
method="linear",
fill_value=None,
)
print( foo( (0,0)))
print( foo( [ (0,0), (0,0) ]))
print( foo(np.zeros([20,21,2],"f")))
```
### Error message
```shell
Traceback (most recent call last):
File "test.py", line 12, in <module>
print( foo( (0,0)))
File "/tmp/pippo/lib/python3.8/site-packages/scipy/interpolate/_rgi.py", line 336, in __call__
result = evaluate_linear_2d(self.values,
File "_rgi_cython.pyx", line 19, in scipy.interpolate._rgi_cython.__pyx_fused_cpdef
TypeError: No matching signature found
```
### SciPy/NumPy/Python version information
1.10.0 1.24.1 sys.version_info(major=3, minor=8, micro=10, releaselevel='final', serial=0) | 1.0 | BUG: in version 1.10 No matching signature found when trying to use scipy.interpolate.RegularGridInterpolator - ### Describe your issue.
with scipy1.10 scipy.interpolate.RegularGridInterpolator object gives the following error message when the __call__ method is used:
### Reproducing Code Example
```python
import scipy
import scipy.interpolate
import numpy as np
foo = scipy.interpolate.RegularGridInterpolator(
[np.arange(4), np.arange(5)],
np.zeros([4,5],"f"),
bounds_error=False,
method="linear",
fill_value=None,
)
print( foo( (0,0)))
print( foo( [ (0,0), (0,0) ]))
print( foo(np.zeros([20,21,2],"f")))
```
### Error message
```shell
Traceback (most recent call last):
File "test.py", line 12, in <module>
print( foo( (0,0)))
File "/tmp/pippo/lib/python3.8/site-packages/scipy/interpolate/_rgi.py", line 336, in __call__
result = evaluate_linear_2d(self.values,
File "_rgi_cython.pyx", line 19, in scipy.interpolate._rgi_cython.__pyx_fused_cpdef
TypeError: No matching signature found
```
### SciPy/NumPy/Python version information
1.10.0 1.24.1 sys.version_info(major=3, minor=8, micro=10, releaselevel='final', serial=0) | defect | bug in version no matching signature found when trying to use scipy interpolate regulargridinterpolator describe your issue with scipy interpolate regulargridinterpolator object gives the following error message when the call method is used reproducing code example python import scipy import scipy interpolate import numpy as np foo scipy interpolate regulargridinterpolator np zeros f bounds error false method linear fill value none print foo print foo print foo np zeros f error message shell traceback most recent call last file test py line in print foo file tmp pippo lib site packages scipy interpolate rgi py line in call result evaluate linear self values file rgi cython pyx line in scipy interpolate rgi cython pyx fused cpdef typeerror no matching signature found scipy numpy python version information sys version info major minor micro releaselevel final serial | 1 |
17,517 | 3,011,118,809 | IssuesEvent | 2015-07-28 16:18:28 | KasaiDot/codejam-commandline | https://api.github.com/repos/KasaiDot/codejam-commandline | closed | Commandline tool gets checker comment | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Make an incorrect submition for problem C-small of yesturday's round
What is the expected output? What do you see instead?
Only word Incorrect. But I got information like "In test 13 6-th can see 9-th
but should see 8-th".
What version of the product are you using? On what operating system?
Windows 7, version dowloadable in downloads, beta4.
But --version says 1.0-beta3.
```
Original issue reported on code.google.com by `Kunyavs...@gmail.com` on 27 May 2012 at 6:12 | 1.0 | Commandline tool gets checker comment - ```
What steps will reproduce the problem?
1. Make an incorrect submition for problem C-small of yesturday's round
What is the expected output? What do you see instead?
Only word Incorrect. But I got information like "In test 13 6-th can see 9-th
but should see 8-th".
What version of the product are you using? On what operating system?
Windows 7, version dowloadable in downloads, beta4.
But --version says 1.0-beta3.
```
Original issue reported on code.google.com by `Kunyavs...@gmail.com` on 27 May 2012 at 6:12 | defect | commandline tool gets checker comment what steps will reproduce the problem make an incorrect submition for problem c small of yesturday s round what is the expected output what do you see instead only word incorrect but i got information like in test th can see th but should see th what version of the product are you using on what operating system windows version dowloadable in downloads but version says original issue reported on code google com by kunyavs gmail com on may at | 1 |
75,669 | 25,987,381,153 | IssuesEvent | 2022-12-20 02:20:01 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | Tree: Selecting and unselecting node programmatically | :lady_beetle: defect workaround | Hello,
I am a newcomer in JSF and Primefaces, so if this is not a bug, sorry for opening the issue. I have written first in the forum, but nobody has answered me.
## 1) Environment
I have downloaded the test project ( https://github.com/primefaces/primefaces-test ) and I have modified it to reproduce error.
I have experienced this error in Firefox, Chrome and Opera, in Windows.
## 2) Expected behavior
I have a tree with checkbox nodes with the next hierarchy:

If I select _2015 -> A -> X_, I want _2016 -> A -> X_ to be selected, too. And viceversa. The same behaviour for all nodes and for the unselect action. To have this behavior, I have onNodeSelect and onNodeUnselect events and I select/unselect programmatically the other nodes.
## 3) Actual behavior
If I only (un)select nodes in one branch (2015 or 2016), the behaviour is as expected. But if I (un)select nodes in one branch, and then I (un)select nodes in the other branchs, some problems happen.
## 4) Steps to reproduce
1. Select _2015->A->X_ node
Then, the _2016->A->X_ node is selected
2. Select _2015->A->Y_ node
Then, the _2016->A->Y_ node is selected
3. Select _2015->A->Z_ node
Then, the _2016->A->Z_ node is selected
Nodes _2015_, _2016_, _2015->A_ and _2016-A_ are selected, too, so all its children are selected.
4. Unselect _2016->A->Y_ node
Then, the _2015->A->Y_ node is unselected, **but the _2016->A->Y_ node is still selected.**
If you debug the code, in this step, at the begining of the "onNodeUnselect" method, the selected property of _2016->A->Y_ is _true_, when _false_ is expected. In fact, by debugging, in this moment in the page the checbox is unselected, as expected, but after running the code, it is selected automatically.
I tried to force to update the "selected" property to false in the onNodeUnselect method (and to true in the onNodeSelect method. Commented code). Then this problem is solved, but if you do next step:
5. Unselect _2016->A->X_ node
Then the _2016->A->X_ node is unselected, **and the _2015->A->Y_ node is selected!**
## 5) Sample XHTML (The whole project is attached, too)
```xml
<!DOCTYPE html>
<html xmlns="http://www.w3.org/1999/xhtml"
xmlns:ui="http://java.sun.com/jsf/facelets"
xmlns:f="http://java.sun.com/jsf/core"
xmlns:p="http://primefaces.org/ui"
xmlns:h="http://java.sun.com/jsf/html">
<h:head>
<title>PrimeFaces Test</title>
</h:head>
<h:body>
<h:form id="tree">
<p:growl id="msgs" showDetail="true" escape="false"/>
<h3>JSF 2.0 Tree Selection Example</h3>
<p:tree value="#{testView.root}" var="doc"
selectionMode="checkbox"
selection="#{testView.selectedNodes}"
dynamic="true">
<p:treeNode>
<h:outputText value="#{doc.name}" />
</p:treeNode>
<p:ajax event="select"
listener="#{testView.onNodeSelect}"
update=":tree"
/>
<p:ajax event="unselect"
listener="#{testView.onNodeUnselect}"
update=":tree"
/>
</p:tree>
<br/>
<p:commandButton value="Show selected" update="msgs" icon="ui-icon-newwin"
actionListener="#{testView.displaySelectedNodes(testView.selectedNodes)}"/>
</h:form>
</h:body>
</html>
```
## 6) Sample bean (The whole project is attached, too)
Bean TestView.java:
```java
package org.primefaces.test;
import java.io.Serializable;
import javax.annotation.PostConstruct;
import javax.faces.application.FacesMessage;
import javax.faces.bean.ManagedBean;
import javax.faces.bean.ViewScoped;
import javax.faces.context.FacesContext;
import org.primefaces.event.NodeSelectEvent;
import org.primefaces.event.NodeUnselectEvent;
import org.primefaces.model.CheckboxTreeNode;
import org.primefaces.model.TreeNode;
@ManagedBean(name = "testView")
@ViewScoped
public class TestView implements Serializable {
private TreeNode root;
private CheckboxTreeNode _2015;
private CheckboxTreeNode a2015;
private CheckboxTreeNode xa2015;
private CheckboxTreeNode ya2015;
private CheckboxTreeNode za2015;
private CheckboxTreeNode _2016;
private CheckboxTreeNode a2016;
private CheckboxTreeNode xa2016;
private CheckboxTreeNode ya2016;
private CheckboxTreeNode za2016;
public TreeNode createCheckboxDocuments() {
root = new CheckboxTreeNode(new Document("root"), null);
root.setExpanded(true);
_2015 = new CheckboxTreeNode(new Document("2015"), root);
_2015.setExpanded(true);
_2016 = new CheckboxTreeNode(new Document("2016"), root);
_2016.setExpanded(true);
a2015 = new CheckboxTreeNode(new Document("A"), _2015);
a2015.setExpanded(true);
a2016 = new CheckboxTreeNode(new Document("A"), _2016);
a2016.setExpanded(true);
xa2015 = new CheckboxTreeNode(new Document("X"), a2015);
xa2015.setExpanded(true);
ya2015 = new CheckboxTreeNode(new Document("Y"), a2015);
ya2015.setExpanded(true);
za2015 = new CheckboxTreeNode(new Document("Z"), a2015);
za2015.setExpanded(true);
xa2016 = new CheckboxTreeNode(new Document("X"), a2016);
xa2016.setExpanded(true);
ya2016 = new CheckboxTreeNode(new Document("Y"), a2016);
ya2016.setExpanded(true);
za2016 = new CheckboxTreeNode(new Document("Z"), a2016);
za2016.setExpanded(true);
return root;
}
private TreeNode[] selectedNodes;
@PostConstruct
public void init() {
root = createCheckboxDocuments();
}
public TreeNode getRoot() {
return root;
}
public TreeNode[] getSelectedNodes() {
return selectedNodes;
}
public void setSelectedNodes(TreeNode[] selectedNodes) {
this.selectedNodes = selectedNodes;
}
public void displaySelectedNodes(TreeNode[] nodes) {
if(nodes != null && nodes.length > 0) {
StringBuilder builder = new StringBuilder();
for(TreeNode node : nodes) {
if (node.isLeaf()) {
while(node.getParent() != null) {
builder.append(node.getData()).append("<-");
node = node.getParent();
}
builder.append("<br />");
}
}
FacesMessage message = new FacesMessage(FacesMessage.SEVERITY_INFO, "Your choices:", builder.toString());
FacesContext.getCurrentInstance().addMessage(null, message);
}
}
public void onNodeUnselect(NodeUnselectEvent event)
{
TreeNode currentTreeNode = event.getTreeNode();
// currentTreeNode.setSelected(false);
TreeNode parent = currentTreeNode;
while (!((Document) parent.getParent().getData()).getName().equals(((Document) root.getData()).getName()))
{
parent = parent.getParent();
//Selects node "2015" or "2016"
}
updateBrother(((Document) currentTreeNode.getData()).getName(), ((Document) parent.getData()).getName(), false);
}
private void updateBrother(String name, String parentName, boolean value)
{
if (parentName.equals("2015"))
{
if (name.equals("2015"))
{
a2015.setSelected(value);
xa2015.setSelected(value);
ya2015.setSelected(value);
za2015.setSelected(value);
_2016.setSelected(value);
a2016.setSelected(value);
xa2016.setSelected(value);
ya2016.setSelected(value);
za2016.setSelected(value);
} else if (name.equals("A"))
{
xa2015.setSelected(value);
ya2015.setSelected(value);
za2015.setSelected(value);
a2016.setSelected(value);
xa2016.setSelected(value);
ya2016.setSelected(value);
za2016.setSelected(value);
} else if (name.equals("X"))
{
xa2016.setSelected(value);
} else if (name.equals("Y"))
{
ya2016.setSelected(value);
}else if (name.equals("Z"))
{
za2016.setSelected(value);
}
} else {
if (name.equals("2016"))
{
a2016.setSelected(value);
xa2016.setSelected(value);
ya2016.setSelected(value);
za2016.setSelected(value);
_2015.setSelected(value);
a2015.setSelected(value);
xa2015.setSelected(value);
ya2015.setSelected(value);
za2015.setSelected(value);
} else if (name.equals("A"))
{
xa2016.setSelected(value);
ya2016.setSelected(value);
za2016.setSelected(value);
a2015.setSelected(value);
xa2015.setSelected(value);
ya2015.setSelected(value);
za2015.setSelected(value);
} else if (name.equals("X"))
{
xa2015.setSelected(value);
} else if (name.equals("Y"))
{
ya2015.setSelected(value);
}else if (name.equals("Z"))
{
za2015.setSelected(value);
}
}
}
public void onNodeSelect(NodeSelectEvent event)
{
TreeNode currentTreeNode = event.getTreeNode();
// currentTreeNode.setSelected(true);
TreeNode parent = currentTreeNode;
while (!((Document) parent.getParent().getData()).getName().equals(((Document) root.getData()).getName()))
{
parent = parent.getParent();
}
updateBrother(((Document) currentTreeNode.getData()).getName(), ((Document) parent.getData()).getName(), true);
}
}
```
Document.java:
```java
package org.primefaces.test;
import java.io.Serializable;
public class Document implements Serializable {
private static final long serialVersionUID = 1L;
private String name;
public Document(String name) {
this.name = name;
}
public String getName() {
return name;
}
public void setName(String name) {
this.name = name;
}
@Override
public String toString() {
return name;
}
}
```
(I wanted to upload a zip, but GitHub did not allow me. I expect tar.gz is ok for you)
[treeNodeTest.tar.gz](https://github.com/primefaces/primefaces/files/925467/treeNodeTest.tar.gz)
| 1.0 | Tree: Selecting and unselecting node programmatically - Hello,
I am a newcomer in JSF and Primefaces, so if this is not a bug, sorry for opening the issue. I have written first in the forum, but nobody has answered me.
## 1) Environment
I have downloaded the test project ( https://github.com/primefaces/primefaces-test ) and I have modified it to reproduce error.
I have experienced this error in Firefox, Chrome and Opera, in Windows.
## 2) Expected behavior
I have a tree with checkbox nodes with the next hierarchy:

If I select _2015 -> A -> X_, I want _2016 -> A -> X_ to be selected, too. And viceversa. The same behaviour for all nodes and for the unselect action. To have this behavior, I have onNodeSelect and onNodeUnselect events and I select/unselect programmatically the other nodes.
## 3) Actual behavior
If I only (un)select nodes in one branch (2015 or 2016), the behaviour is as expected. But if I (un)select nodes in one branch, and then I (un)select nodes in the other branchs, some problems happen.
## 4) Steps to reproduce
1. Select _2015->A->X_ node
Then, the _2016->A->X_ node is selected
2. Select _2015->A->Y_ node
Then, the _2016->A->Y_ node is selected
3. Select _2015->A->Z_ node
Then, the _2016->A->Z_ node is selected
Nodes _2015_, _2016_, _2015->A_ and _2016-A_ are selected, too, so all its children are selected.
4. Unselect _2016->A->Y_ node
Then, the _2015->A->Y_ node is unselected, **but the _2016->A->Y_ node is still selected.**
If you debug the code, in this step, at the begining of the "onNodeUnselect" method, the selected property of _2016->A->Y_ is _true_, when _false_ is expected. In fact, by debugging, in this moment in the page the checbox is unselected, as expected, but after running the code, it is selected automatically.
I tried to force to update the "selected" property to false in the onNodeUnselect method (and to true in the onNodeSelect method. Commented code). Then this problem is solved, but if you do next step:
5. Unselect _2016->A->X_ node
Then the _2016->A->X_ node is unselected, **and the _2015->A->Y_ node is selected!**
## 5) Sample XHTML (The whole project is attached, too)
```xml
<!DOCTYPE html>
<html xmlns="http://www.w3.org/1999/xhtml"
xmlns:ui="http://java.sun.com/jsf/facelets"
xmlns:f="http://java.sun.com/jsf/core"
xmlns:p="http://primefaces.org/ui"
xmlns:h="http://java.sun.com/jsf/html">
<h:head>
<title>PrimeFaces Test</title>
</h:head>
<h:body>
<h:form id="tree">
<p:growl id="msgs" showDetail="true" escape="false"/>
<h3>JSF 2.0 Tree Selection Example</h3>
<p:tree value="#{testView.root}" var="doc"
selectionMode="checkbox"
selection="#{testView.selectedNodes}"
dynamic="true">
<p:treeNode>
<h:outputText value="#{doc.name}" />
</p:treeNode>
<p:ajax event="select"
listener="#{testView.onNodeSelect}"
update=":tree"
/>
<p:ajax event="unselect"
listener="#{testView.onNodeUnselect}"
update=":tree"
/>
</p:tree>
<br/>
<p:commandButton value="Show selected" update="msgs" icon="ui-icon-newwin"
actionListener="#{testView.displaySelectedNodes(testView.selectedNodes)}"/>
</h:form>
</h:body>
</html>
```
## 6) Sample bean (The whole project is attached, too)
Bean TestView.java:
```java
package org.primefaces.test;
import java.io.Serializable;
import javax.annotation.PostConstruct;
import javax.faces.application.FacesMessage;
import javax.faces.bean.ManagedBean;
import javax.faces.bean.ViewScoped;
import javax.faces.context.FacesContext;
import org.primefaces.event.NodeSelectEvent;
import org.primefaces.event.NodeUnselectEvent;
import org.primefaces.model.CheckboxTreeNode;
import org.primefaces.model.TreeNode;
@ManagedBean(name = "testView")
@ViewScoped
public class TestView implements Serializable {
private TreeNode root;
private CheckboxTreeNode _2015;
private CheckboxTreeNode a2015;
private CheckboxTreeNode xa2015;
private CheckboxTreeNode ya2015;
private CheckboxTreeNode za2015;
private CheckboxTreeNode _2016;
private CheckboxTreeNode a2016;
private CheckboxTreeNode xa2016;
private CheckboxTreeNode ya2016;
private CheckboxTreeNode za2016;
public TreeNode createCheckboxDocuments() {
root = new CheckboxTreeNode(new Document("root"), null);
root.setExpanded(true);
_2015 = new CheckboxTreeNode(new Document("2015"), root);
_2015.setExpanded(true);
_2016 = new CheckboxTreeNode(new Document("2016"), root);
_2016.setExpanded(true);
a2015 = new CheckboxTreeNode(new Document("A"), _2015);
a2015.setExpanded(true);
a2016 = new CheckboxTreeNode(new Document("A"), _2016);
a2016.setExpanded(true);
xa2015 = new CheckboxTreeNode(new Document("X"), a2015);
xa2015.setExpanded(true);
ya2015 = new CheckboxTreeNode(new Document("Y"), a2015);
ya2015.setExpanded(true);
za2015 = new CheckboxTreeNode(new Document("Z"), a2015);
za2015.setExpanded(true);
xa2016 = new CheckboxTreeNode(new Document("X"), a2016);
xa2016.setExpanded(true);
ya2016 = new CheckboxTreeNode(new Document("Y"), a2016);
ya2016.setExpanded(true);
za2016 = new CheckboxTreeNode(new Document("Z"), a2016);
za2016.setExpanded(true);
return root;
}
private TreeNode[] selectedNodes;
@PostConstruct
public void init() {
root = createCheckboxDocuments();
}
public TreeNode getRoot() {
return root;
}
public TreeNode[] getSelectedNodes() {
return selectedNodes;
}
public void setSelectedNodes(TreeNode[] selectedNodes) {
this.selectedNodes = selectedNodes;
}
public void displaySelectedNodes(TreeNode[] nodes) {
if(nodes != null && nodes.length > 0) {
StringBuilder builder = new StringBuilder();
for(TreeNode node : nodes) {
if (node.isLeaf()) {
while(node.getParent() != null) {
builder.append(node.getData()).append("<-");
node = node.getParent();
}
builder.append("<br />");
}
}
FacesMessage message = new FacesMessage(FacesMessage.SEVERITY_INFO, "Your choices:", builder.toString());
FacesContext.getCurrentInstance().addMessage(null, message);
}
}
public void onNodeUnselect(NodeUnselectEvent event)
{
TreeNode currentTreeNode = event.getTreeNode();
// currentTreeNode.setSelected(false);
TreeNode parent = currentTreeNode;
while (!((Document) parent.getParent().getData()).getName().equals(((Document) root.getData()).getName()))
{
parent = parent.getParent();
//Selects node "2015" or "2016"
}
updateBrother(((Document) currentTreeNode.getData()).getName(), ((Document) parent.getData()).getName(), false);
}
private void updateBrother(String name, String parentName, boolean value)
{
if (parentName.equals("2015"))
{
if (name.equals("2015"))
{
a2015.setSelected(value);
xa2015.setSelected(value);
ya2015.setSelected(value);
za2015.setSelected(value);
_2016.setSelected(value);
a2016.setSelected(value);
xa2016.setSelected(value);
ya2016.setSelected(value);
za2016.setSelected(value);
} else if (name.equals("A"))
{
xa2015.setSelected(value);
ya2015.setSelected(value);
za2015.setSelected(value);
a2016.setSelected(value);
xa2016.setSelected(value);
ya2016.setSelected(value);
za2016.setSelected(value);
} else if (name.equals("X"))
{
xa2016.setSelected(value);
} else if (name.equals("Y"))
{
ya2016.setSelected(value);
}else if (name.equals("Z"))
{
za2016.setSelected(value);
}
} else {
if (name.equals("2016"))
{
a2016.setSelected(value);
xa2016.setSelected(value);
ya2016.setSelected(value);
za2016.setSelected(value);
_2015.setSelected(value);
a2015.setSelected(value);
xa2015.setSelected(value);
ya2015.setSelected(value);
za2015.setSelected(value);
} else if (name.equals("A"))
{
xa2016.setSelected(value);
ya2016.setSelected(value);
za2016.setSelected(value);
a2015.setSelected(value);
xa2015.setSelected(value);
ya2015.setSelected(value);
za2015.setSelected(value);
} else if (name.equals("X"))
{
xa2015.setSelected(value);
} else if (name.equals("Y"))
{
ya2015.setSelected(value);
}else if (name.equals("Z"))
{
za2015.setSelected(value);
}
}
}
public void onNodeSelect(NodeSelectEvent event)
{
TreeNode currentTreeNode = event.getTreeNode();
// currentTreeNode.setSelected(true);
TreeNode parent = currentTreeNode;
while (!((Document) parent.getParent().getData()).getName().equals(((Document) root.getData()).getName()))
{
parent = parent.getParent();
}
updateBrother(((Document) currentTreeNode.getData()).getName(), ((Document) parent.getData()).getName(), true);
}
}
```
Document.java:
```java
package org.primefaces.test;
import java.io.Serializable;
public class Document implements Serializable {
private static final long serialVersionUID = 1L;
private String name;
public Document(String name) {
this.name = name;
}
public String getName() {
return name;
}
public void setName(String name) {
this.name = name;
}
@Override
public String toString() {
return name;
}
}
```
(I wanted to upload a zip, but GitHub did not allow me. I expect tar.gz is ok for you)
[treeNodeTest.tar.gz](https://github.com/primefaces/primefaces/files/925467/treeNodeTest.tar.gz)
| defect | tree selecting and unselecting node programmatically hello i am a newcomer in jsf and primefaces so if this is not a bug sorry for opening the issue i have written first in the forum but nobody has answered me environment i have downloaded the test project and i have modified it to reproduce error i have experienced this error in firefox chrome and opera in windows expected behavior i have a tree with checkbox nodes with the next hierarchy if i select a x i want a x to be selected too and viceversa the same behaviour for all nodes and for the unselect action to have this behavior i have onnodeselect and onnodeunselect events and i select unselect programmatically the other nodes actual behavior if i only un select nodes in one branch or the behaviour is as expected but if i un select nodes in one branch and then i un select nodes in the other branchs some problems happen steps to reproduce select a x node then the a x node is selected select a y node then the a y node is selected select a z node then the a z node is selected nodes a and a are selected too so all its children are selected unselect a y node then the a y node is unselected but the a y node is still selected if you debug the code in this step at the begining of the onnodeunselect method the selected property of a y is true when false is expected in fact by debugging in this moment in the page the checbox is unselected as expected but after running the code it is selected automatically i tried to force to update the selected property to false in the onnodeunselect method and to true in the onnodeselect method commented code then this problem is solved but if you do next step unselect a x node then the a x node is unselected and the a y node is selected sample xhtml the whole project is attached too xml html xmlns xmlns ui xmlns f xmlns p xmlns h primefaces test jsf tree selection example p tree value testview root var doc selectionmode checkbox selection testview selectednodes dynamic true p ajax event select listener testview onnodeselect update tree p ajax event unselect listener testview onnodeunselect update tree p commandbutton value show selected update msgs icon ui icon newwin actionlistener testview displayselectednodes testview selectednodes sample bean the whole project is attached too bean testview java java package org primefaces test import java io serializable import javax annotation postconstruct import javax faces application facesmessage import javax faces bean managedbean import javax faces bean viewscoped import javax faces context facescontext import org primefaces event nodeselectevent import org primefaces event nodeunselectevent import org primefaces model checkboxtreenode import org primefaces model treenode managedbean name testview viewscoped public class testview implements serializable private treenode root private checkboxtreenode private checkboxtreenode private checkboxtreenode private checkboxtreenode private checkboxtreenode private checkboxtreenode private checkboxtreenode private checkboxtreenode private checkboxtreenode private checkboxtreenode public treenode createcheckboxdocuments root new checkboxtreenode new document root null root setexpanded true new checkboxtreenode new document root setexpanded true new checkboxtreenode new document root setexpanded true new checkboxtreenode new document a setexpanded true new checkboxtreenode new document a setexpanded true new checkboxtreenode new document x setexpanded true new checkboxtreenode new document y setexpanded true new checkboxtreenode new document z setexpanded true new checkboxtreenode new document x setexpanded true new checkboxtreenode new document y setexpanded true new checkboxtreenode new document z setexpanded true return root private treenode selectednodes postconstruct public void init root createcheckboxdocuments public treenode getroot return root public treenode getselectednodes return selectednodes public void setselectednodes treenode selectednodes this selectednodes selectednodes public void displayselectednodes treenode nodes if nodes null nodes length stringbuilder builder new stringbuilder for treenode node nodes if node isleaf while node getparent null builder append node getdata append node node getparent builder append facesmessage message new facesmessage facesmessage severity info your choices builder tostring facescontext getcurrentinstance addmessage null message public void onnodeunselect nodeunselectevent event treenode currenttreenode event gettreenode currenttreenode setselected false treenode parent currenttreenode while document parent getparent getdata getname equals document root getdata getname parent parent getparent selects node or updatebrother document currenttreenode getdata getname document parent getdata getname false private void updatebrother string name string parentname boolean value if parentname equals if name equals setselected value setselected value setselected value setselected value setselected value setselected value setselected value setselected value setselected value else if name equals a setselected value setselected value setselected value setselected value setselected value setselected value setselected value else if name equals x setselected value else if name equals y setselected value else if name equals z setselected value else if name equals setselected value setselected value setselected value setselected value setselected value setselected value setselected value setselected value setselected value else if name equals a setselected value setselected value setselected value setselected value setselected value setselected value setselected value else if name equals x setselected value else if name equals y setselected value else if name equals z setselected value public void onnodeselect nodeselectevent event treenode currenttreenode event gettreenode currenttreenode setselected true treenode parent currenttreenode while document parent getparent getdata getname equals document root getdata getname parent parent getparent updatebrother document currenttreenode getdata getname document parent getdata getname true document java java package org primefaces test import java io serializable public class document implements serializable private static final long serialversionuid private string name public document string name this name name public string getname return name public void setname string name this name name override public string tostring return name i wanted to upload a zip but github did not allow me i expect tar gz is ok for you | 1 |
39,500 | 9,521,861,742 | IssuesEvent | 2019-04-27 02:16:08 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | Postgres Schema: CHAR(36) results in CHAR(1) columns | Defect postgres | This is a:
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.6.7
* Platform and Target: PHP 7.2.17, **PostgreSQL 10.7** (Ubuntu 10.7-0ubuntu0.18.10.1)
### What you did
Fixtures using `$import` on tables that contain a `char(36)` column are created incorrectly as a `char` (which results in `char(1)`).
### What happened
Inserting fixture data generates error:
```
ERROR: value too long for type character(1)
```
### What you expected to happen
I expected a char column with a size of 36 to be created.
We are using postgres UUID columns in the app and that works as expected, but we are using `char(36)` for another non-standard ID.
For now I'll just resize our columns to make it work.
It can be fixed by removing the second part of the conditional [here](https://github.com/cakephp/cakephp/blob/13ab4d788ce553056938b60b3039197b8edd9993/src/Database/Schema/PostgresSchema.php#L402).
```
-- Part of the generated SQL:
CREATE TABLE "dev" (
"id" BIGSERIAL,
"dev_id" CHAR DEFAULT NULL, -- This here should be CHAR(36)
PRIMARY KEY ("id")
) | 1.0 | Postgres Schema: CHAR(36) results in CHAR(1) columns - This is a:
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.6.7
* Platform and Target: PHP 7.2.17, **PostgreSQL 10.7** (Ubuntu 10.7-0ubuntu0.18.10.1)
### What you did
Fixtures using `$import` on tables that contain a `char(36)` column are created incorrectly as a `char` (which results in `char(1)`).
### What happened
Inserting fixture data generates error:
```
ERROR: value too long for type character(1)
```
### What you expected to happen
I expected a char column with a size of 36 to be created.
We are using postgres UUID columns in the app and that works as expected, but we are using `char(36)` for another non-standard ID.
For now I'll just resize our columns to make it work.
It can be fixed by removing the second part of the conditional [here](https://github.com/cakephp/cakephp/blob/13ab4d788ce553056938b60b3039197b8edd9993/src/Database/Schema/PostgresSchema.php#L402).
```
-- Part of the generated SQL:
CREATE TABLE "dev" (
"id" BIGSERIAL,
"dev_id" CHAR DEFAULT NULL, -- This here should be CHAR(36)
PRIMARY KEY ("id")
) | defect | postgres schema char results in char columns this is a bug enhancement feature discussion rfc cakephp version platform and target php postgresql ubuntu what you did fixtures using import on tables that contain a char column are created incorrectly as a char which results in char what happened inserting fixture data generates error error value too long for type character what you expected to happen i expected a char column with a size of to be created we are using postgres uuid columns in the app and that works as expected but we are using char for another non standard id for now i ll just resize our columns to make it work it can be fixed by removing the second part of the conditional part of the generated sql create table dev id bigserial dev id char default null this here should be char primary key id | 1 |
38,859 | 8,974,956,712 | IssuesEvent | 2019-01-30 02:45:03 | zkoss/zkspring | https://api.github.com/repos/zkoss/zkspring | closed | ZK SprinSecurity NullPointerException on SecurityUti.isAccessible method | Priority-Medium Type-Defect auto-migrated | ```
What steps will reproduce the problem?
1. SecurityUti.isAccessible("WRITE", domainObject)
2.
3.
What is the expected output? true or false
What do you see instead? permissionFactory is null, NullPointerException
What version of the product are you using? 3.0
On what operating system? Ubuntu 9.10
Please provide any additional information below.
In the method private static void initializeIfRequired(); there are not
initializations of permissionFactory.
```
Original issue reported on code.google.com by `gesuino....@gmail.com` on 10 Apr 2011 at 5:09
| 1.0 | ZK SprinSecurity NullPointerException on SecurityUti.isAccessible method - ```
What steps will reproduce the problem?
1. SecurityUti.isAccessible("WRITE", domainObject)
2.
3.
What is the expected output? true or false
What do you see instead? permissionFactory is null, NullPointerException
What version of the product are you using? 3.0
On what operating system? Ubuntu 9.10
Please provide any additional information below.
In the method private static void initializeIfRequired(); there are not
initializations of permissionFactory.
```
Original issue reported on code.google.com by `gesuino....@gmail.com` on 10 Apr 2011 at 5:09
| defect | zk sprinsecurity nullpointerexception on securityuti isaccessible method what steps will reproduce the problem securityuti isaccessible write domainobject what is the expected output true or false what do you see instead permissionfactory is null nullpointerexception what version of the product are you using on what operating system ubuntu please provide any additional information below in the method private static void initializeifrequired there are not initializations of permissionfactory original issue reported on code google com by gesuino gmail com on apr at | 1 |
43,262 | 11,585,519,406 | IssuesEvent | 2020-02-23 00:30:27 | idaholab/moose | https://api.github.com/repos/idaholab/moose | opened | A mistake in beam element derivation | T: defect | Hi all,
I think I caught a mistake in the beam element derivation but am not too sure...
In the calculation of the rotational strain increments, kappa_3, the integration seems to be wrong in the sign. It should be int(eps_11 * (-y) * dA); but currently it is int(eps_11 * y * dA).
1. If we draw the diagram on a piece of paper, it is easy to see that a point with positive y coordinate having a stress pointing in +x direction should generate a moment that is clock-wise, i.e. in the -z direction;
2. If we use the current formulation as implemented in MOOSE, a positive increment in theta_z will cause a negative moment_z, which seems incorrect to me as well.
If anyone could kindly confirm, I can submit a pull request and the corresponding test as well.
@sveerara if you have a chance I'd also appreciate it. Thanks in advance Swetha! :)
Shawn
| 1.0 | A mistake in beam element derivation - Hi all,
I think I caught a mistake in the beam element derivation but am not too sure...
In the calculation of the rotational strain increments, kappa_3, the integration seems to be wrong in the sign. It should be int(eps_11 * (-y) * dA); but currently it is int(eps_11 * y * dA).
1. If we draw the diagram on a piece of paper, it is easy to see that a point with positive y coordinate having a stress pointing in +x direction should generate a moment that is clock-wise, i.e. in the -z direction;
2. If we use the current formulation as implemented in MOOSE, a positive increment in theta_z will cause a negative moment_z, which seems incorrect to me as well.
If anyone could kindly confirm, I can submit a pull request and the corresponding test as well.
@sveerara if you have a chance I'd also appreciate it. Thanks in advance Swetha! :)
Shawn
| defect | a mistake in beam element derivation hi all i think i caught a mistake in the beam element derivation but am not too sure in the calculation of the rotational strain increments kappa the integration seems to be wrong in the sign it should be int eps y da but currently it is int eps y da if we draw the diagram on a piece of paper it is easy to see that a point with positive y coordinate having a stress pointing in x direction should generate a moment that is clock wise i e in the z direction if we use the current formulation as implemented in moose a positive increment in theta z will cause a negative moment z which seems incorrect to me as well if anyone could kindly confirm i can submit a pull request and the corresponding test as well sveerara if you have a chance i d also appreciate it thanks in advance swetha shawn | 1 |
258,620 | 27,565,505,797 | IssuesEvent | 2023-03-08 03:12:00 | rvvergara/bookstore_rails_app | https://api.github.com/repos/rvvergara/bookstore_rails_app | closed | CVE-2022-23515 (Medium) detected in loofah-2.2.3.gem - autoclosed | Mend: dependency security vulnerability | ## CVE-2022-23515 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loofah-2.2.3.gem</b></p></summary>
<p>Loofah is a general library for manipulating and transforming HTML/XML
documents and fragments. It's built on top of Nokogiri and libxml2, so
it's fast and has a nice API.
Loofah excels at HTML sanitization (XSS prevention). It includes some
nice HTML sanitizers, which are based on HTML5lib's whitelist, so it
most likely won't make your codes less secure. (These statements have
not been evaluated by Netexperts.)
ActiveRecord extensions for sanitization are available in the
[`loofah-activerecord` gem](https://github.com/flavorjones/loofah-activerecord).</p>
<p>Library home page: <a href="https://rubygems.org/gems/loofah-2.2.3.gem">https://rubygems.org/gems/loofah-2.2.3.gem</a></p>
<p>
Dependency Hierarchy:
- rails-5.2.3.gem (Root Library)
- actionview-5.2.3.gem
- rails-html-sanitizer-1.2.0.gem
- :x: **loofah-2.2.3.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rvvergara/bookstore_rails_app/commit/0cc6845b68c9648d6cc64293d4276d9e63d3123e">0cc6845b68c9648d6cc64293d4276d9e63d3123e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Loofah is a general library for manipulating and transforming HTML/XML documents and fragments, built on top of Nokogiri. Loofah >= 2.1.0, < 2.19.1 is vulnerable to cross-site scripting via the image/svg+xml media type in data URIs. This issue is patched in version 2.19.1.
<p>Publish Date: 2022-12-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23515>CVE-2022-23515</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-228g-948r-83gx">https://github.com/advisories/GHSA-228g-948r-83gx</a></p>
<p>Release Date: 2022-12-14</p>
<p>Fix Resolution: loofah - 2.19.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-23515 (Medium) detected in loofah-2.2.3.gem - autoclosed - ## CVE-2022-23515 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loofah-2.2.3.gem</b></p></summary>
<p>Loofah is a general library for manipulating and transforming HTML/XML
documents and fragments. It's built on top of Nokogiri and libxml2, so
it's fast and has a nice API.
Loofah excels at HTML sanitization (XSS prevention). It includes some
nice HTML sanitizers, which are based on HTML5lib's whitelist, so it
most likely won't make your codes less secure. (These statements have
not been evaluated by Netexperts.)
ActiveRecord extensions for sanitization are available in the
[`loofah-activerecord` gem](https://github.com/flavorjones/loofah-activerecord).</p>
<p>Library home page: <a href="https://rubygems.org/gems/loofah-2.2.3.gem">https://rubygems.org/gems/loofah-2.2.3.gem</a></p>
<p>
Dependency Hierarchy:
- rails-5.2.3.gem (Root Library)
- actionview-5.2.3.gem
- rails-html-sanitizer-1.2.0.gem
- :x: **loofah-2.2.3.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rvvergara/bookstore_rails_app/commit/0cc6845b68c9648d6cc64293d4276d9e63d3123e">0cc6845b68c9648d6cc64293d4276d9e63d3123e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Loofah is a general library for manipulating and transforming HTML/XML documents and fragments, built on top of Nokogiri. Loofah >= 2.1.0, < 2.19.1 is vulnerable to cross-site scripting via the image/svg+xml media type in data URIs. This issue is patched in version 2.19.1.
<p>Publish Date: 2022-12-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23515>CVE-2022-23515</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-228g-948r-83gx">https://github.com/advisories/GHSA-228g-948r-83gx</a></p>
<p>Release Date: 2022-12-14</p>
<p>Fix Resolution: loofah - 2.19.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in loofah gem autoclosed cve medium severity vulnerability vulnerable library loofah gem loofah is a general library for manipulating and transforming html xml documents and fragments it s built on top of nokogiri and so it s fast and has a nice api loofah excels at html sanitization xss prevention it includes some nice html sanitizers which are based on s whitelist so it most likely won t make your codes less secure these statements have not been evaluated by netexperts activerecord extensions for sanitization are available in the library home page a href dependency hierarchy rails gem root library actionview gem rails html sanitizer gem x loofah gem vulnerable library found in head commit a href found in base branch master vulnerability details loofah is a general library for manipulating and transforming html xml documents and fragments built on top of nokogiri loofah is vulnerable to cross site scripting via the image svg xml media type in data uris this issue is patched in version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution loofah step up your open source security game with mend | 0 |
33,393 | 27,425,887,462 | IssuesEvent | 2023-03-01 20:15:33 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | Issues with installation process that connects an existing DB | type: bug work: infrastructure status: draft restricted: maintainers | - [x] Tester Marius reports (server credentials in Upwork)
- [ ] It seems that even if you select existing database, it still tries to start a docker container for the database, creating a conflict?
- [x] Tester Mohammad reports an error as well ([details here](https://docs.google.com/document/d/15m9eZFocAsU1V9inLKxC6i_KQxMdu28snRrBPOrf5Hk/edit)) | 1.0 | Issues with installation process that connects an existing DB - - [x] Tester Marius reports (server credentials in Upwork)
- [ ] It seems that even if you select existing database, it still tries to start a docker container for the database, creating a conflict?
- [x] Tester Mohammad reports an error as well ([details here](https://docs.google.com/document/d/15m9eZFocAsU1V9inLKxC6i_KQxMdu28snRrBPOrf5Hk/edit)) | non_defect | issues with installation process that connects an existing db tester marius reports server credentials in upwork it seems that even if you select existing database it still tries to start a docker container for the database creating a conflict tester mohammad reports an error as well | 0 |
18,531 | 3,069,562,866 | IssuesEvent | 2015-08-18 21:06:40 | kaoecoito/sintegra-ie | https://api.github.com/repos/kaoecoito/sintegra-ie | closed | IE ISENTO | auto-migrated Priority-Medium Type-Defect | ```
Pelo que verifiquei a sua biblioteca não valida a IE "isento".
```
Original issue reported on code.google.com by `aldeir....@gmail.com` on 10 Jan 2014 at 2:39 | 1.0 | IE ISENTO - ```
Pelo que verifiquei a sua biblioteca não valida a IE "isento".
```
Original issue reported on code.google.com by `aldeir....@gmail.com` on 10 Jan 2014 at 2:39 | defect | ie isento pelo que verifiquei a sua biblioteca não valida a ie isento original issue reported on code google com by aldeir gmail com on jan at | 1 |
80,911 | 30,594,534,109 | IssuesEvent | 2023-07-21 20:24:14 | dotCMS/core | https://api.github.com/repos/dotCMS/core | opened | push-publish docker compose examples are broken | Type : Defect Triage | ### Parent Issue
_No response_
### Problem Statement
These two files are meant to facilitate push-publishing testing
- docker/docker-compose-examples/push-publish/docker-compose-receiver.yml
- docker/docker-compose-examples/push-publish/docker-compose-sender.yml
They should be used to start up two individual nodes. Typically one would end up running in port 8081 and the second in port 8082. Therefore they can be accessed from the browser typically by typing
`http://localhost:8081/c` and `http://localhost:8082/c`
As the user consuming this pp-environment one is supposed to set up a remote endpoint so the sender node can talk to it by entering the IP address of the remote node.
Typically a remote configuration should look like this:

Now Let's suppose you introduce a pp conflict between both nodes (sender and receiver) following the instructions outlined here
https://github.com/dotCMS/core/issues/25229#issuecomment-1640982865
When you hit the `check integrity` button within this setup one would expect the report of the conflict. But in reality, nothing happens. No conflicts are shown here.
And the reason is that the docker container doesn't resolve localhost As our localhost. It thinks that localhost is the same container itself. Regardless of the port number. It sees itself as localhost. Therefore these two instances can NOT see each other.
In reality, when you hit `check integrity` the dotCMS sender instance connects to itself. And the CSV file that gets downloaded contains the exact same image of the sender's database. Therefore no conflict arises when comparing.
Because basically, it is comparing the instance against itself.
I was able to corroborate this. Setting up a remote debugger against the sender node comparing that every time I added a new file or folder. The changes would make it into the file coming from the "remote" receiver.
On the other hand, if the same exercise is carried out against a true remote node with an "absolute" address that can be resolved by the container to an external IP. e.g. using **demo. dotcms.com** The remote endpoint this time shows conflicts as expected. Because the container internally was capable of resolving the address.
in short, localhost is ambiguous for the docker container. And we should try to use an absolute "real" IP address here for testing.
Potential solution:
We could include both nodes (sender and receiver) in the same docker-compose and include in the networks an alias pointing to each one of the instances. And use that instance when configuring the PP Environments.
### Steps to Reproduce
start up the two containers
These two files are meant to facilitate push-publishing testing
- docker/docker-compose-examples/push-publish/docker-compose-receiver.yml
- docker/docker-compose-examples/push-publish/docker-compose-sender.yml
configure sender and make it point http://localhost:8081
now create a conflict following these instructions
https://github.com/dotCMS/core/issues/25229#issuecomment-1640982865
Hit check integrity. You'll see no conflicts are found
Now repeat the same steps but this time using another endpoint pointing to **demo.dotcms.com**
Now see how this time we have conflicts as we expected.
### Acceptance Criteria
We should be able to use these files
- docker/docker-compose-examples/push-publish/docker-compose-receiver.yml
- docker/docker-compose-examples/push-publish/docker-compose-sender.yml
To simulate PP and they should be able to see each other.
### dotCMS Version
current master
### Proposed Objective
Quality Assurance
### Proposed Priority
Priority 3 - Average
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
_No response_
### Assumptions & Initiation Needs
_No response_
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | 1.0 | push-publish docker compose examples are broken - ### Parent Issue
_No response_
### Problem Statement
These two files are meant to facilitate push-publishing testing
- docker/docker-compose-examples/push-publish/docker-compose-receiver.yml
- docker/docker-compose-examples/push-publish/docker-compose-sender.yml
They should be used to start up two individual nodes. Typically one would end up running in port 8081 and the second in port 8082. Therefore they can be accessed from the browser typically by typing
`http://localhost:8081/c` and `http://localhost:8082/c`
As the user consuming this pp-environment one is supposed to set up a remote endpoint so the sender node can talk to it by entering the IP address of the remote node.
Typically a remote configuration should look like this:

Now Let's suppose you introduce a pp conflict between both nodes (sender and receiver) following the instructions outlined here
https://github.com/dotCMS/core/issues/25229#issuecomment-1640982865
When you hit the `check integrity` button within this setup one would expect the report of the conflict. But in reality, nothing happens. No conflicts are shown here.
And the reason is that the docker container doesn't resolve localhost As our localhost. It thinks that localhost is the same container itself. Regardless of the port number. It sees itself as localhost. Therefore these two instances can NOT see each other.
In reality, when you hit `check integrity` the dotCMS sender instance connects to itself. And the CSV file that gets downloaded contains the exact same image of the sender's database. Therefore no conflict arises when comparing.
Because basically, it is comparing the instance against itself.
I was able to corroborate this. Setting up a remote debugger against the sender node comparing that every time I added a new file or folder. The changes would make it into the file coming from the "remote" receiver.
On the other hand, if the same exercise is carried out against a true remote node with an "absolute" address that can be resolved by the container to an external IP. e.g. using **demo. dotcms.com** The remote endpoint this time shows conflicts as expected. Because the container internally was capable of resolving the address.
in short, localhost is ambiguous for the docker container. And we should try to use an absolute "real" IP address here for testing.
Potential solution:
We could include both nodes (sender and receiver) in the same docker-compose and include in the networks an alias pointing to each one of the instances. And use that instance when configuring the PP Environments.
### Steps to Reproduce
start up the two containers
These two files are meant to facilitate push-publishing testing
- docker/docker-compose-examples/push-publish/docker-compose-receiver.yml
- docker/docker-compose-examples/push-publish/docker-compose-sender.yml
configure sender and make it point http://localhost:8081
now create a conflict following these instructions
https://github.com/dotCMS/core/issues/25229#issuecomment-1640982865
Hit check integrity. You'll see no conflicts are found
Now repeat the same steps but this time using another endpoint pointing to **demo.dotcms.com**
Now see how this time we have conflicts as we expected.
### Acceptance Criteria
We should be able to use these files
- docker/docker-compose-examples/push-publish/docker-compose-receiver.yml
- docker/docker-compose-examples/push-publish/docker-compose-sender.yml
To simulate PP and they should be able to see each other.
### dotCMS Version
current master
### Proposed Objective
Quality Assurance
### Proposed Priority
Priority 3 - Average
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
_No response_
### Assumptions & Initiation Needs
_No response_
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | defect | push publish docker compose examples are broken parent issue no response problem statement these two files are meant to facilitate push publishing testing docker docker compose examples push publish docker compose receiver yml docker docker compose examples push publish docker compose sender yml they should be used to start up two individual nodes typically one would end up running in port and the second in port therefore they can be accessed from the browser typically by typing and as the user consuming this pp environment one is supposed to set up a remote endpoint so the sender node can talk to it by entering the ip address of the remote node typically a remote configuration should look like this now let s suppose you introduce a pp conflict between both nodes sender and receiver following the instructions outlined here when you hit the check integrity button within this setup one would expect the report of the conflict but in reality nothing happens no conflicts are shown here and the reason is that the docker container doesn t resolve localhost as our localhost it thinks that localhost is the same container itself regardless of the port number it sees itself as localhost therefore these two instances can not see each other in reality when you hit check integrity the dotcms sender instance connects to itself and the csv file that gets downloaded contains the exact same image of the sender s database therefore no conflict arises when comparing because basically it is comparing the instance against itself i was able to corroborate this setting up a remote debugger against the sender node comparing that every time i added a new file or folder the changes would make it into the file coming from the remote receiver on the other hand if the same exercise is carried out against a true remote node with an absolute address that can be resolved by the container to an external ip e g using demo dotcms com the remote endpoint this time shows conflicts as expected because the container internally was capable of resolving the address in short localhost is ambiguous for the docker container and we should try to use an absolute real ip address here for testing potential solution we could include both nodes sender and receiver in the same docker compose and include in the networks an alias pointing to each one of the instances and use that instance when configuring the pp environments steps to reproduce start up the two containers these two files are meant to facilitate push publishing testing docker docker compose examples push publish docker compose receiver yml docker docker compose examples push publish docker compose sender yml configure sender and make it point now create a conflict following these instructions hit check integrity you ll see no conflicts are found now repeat the same steps but this time using another endpoint pointing to demo dotcms com now see how this time we have conflicts as we expected acceptance criteria we should be able to use these files docker docker compose examples push publish docker compose receiver yml docker docker compose examples push publish docker compose sender yml to simulate pp and they should be able to see each other dotcms version current master proposed objective quality assurance proposed priority priority average external links slack conversations support tickets figma designs etc no response assumptions initiation needs no response quality assurance notes workarounds no response sub tasks estimates no response | 1 |
260,857 | 27,784,776,815 | IssuesEvent | 2023-03-17 01:36:33 | raindigi/reaction | https://api.github.com/repos/raindigi/reaction | opened | CVE-2023-28155 (Medium) detected in request-2.88.2.tgz | Mend: dependency security vulnerability | ## CVE-2023-28155 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>request-2.88.2.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@reactioncommerce/api-plugin-authentication/node_modules/request/package.json,/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- api-plugin-authentication-2.2.3.tgz (Root Library)
- logger-1.1.3.tgz
- node-loggly-bulk-2.2.5.tgz
- :x: **request-2.88.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/raindigi/reaction/commit/c3f5e6b9d647cd1f977b184ae9c079f1ae297353">c3f5e6b9d647cd1f977b184ae9c079f1ae297353</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-28155 (Medium) detected in request-2.88.2.tgz - ## CVE-2023-28155 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>request-2.88.2.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@reactioncommerce/api-plugin-authentication/node_modules/request/package.json,/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- api-plugin-authentication-2.2.3.tgz (Root Library)
- logger-1.1.3.tgz
- node-loggly-bulk-2.2.5.tgz
- :x: **request-2.88.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/raindigi/reaction/commit/c3f5e6b9d647cd1f977b184ae9c079f1ae297353">c3f5e6b9d647cd1f977b184ae9c079f1ae297353</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in request tgz cve medium severity vulnerability vulnerable library request tgz simplified http request client library home page a href path to dependency file package json path to vulnerable library node modules reactioncommerce api plugin authentication node modules request package json node modules request package json dependency hierarchy api plugin authentication tgz root library logger tgz node loggly bulk tgz x request tgz vulnerable library found in head commit a href found in base branch master vulnerability details unsupported when assigned the request package through for node js allows a bypass of ssrf mitigations via an attacker controller server that does a cross protocol redirect http to https or https to http note this vulnerability only affects products that are no longer supported by the maintainer publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend | 0 |
411,263 | 12,016,152,378 | IssuesEvent | 2020-04-10 15:29:52 | scality/metalk8s | https://api.github.com/repos/scality/metalk8s | closed | Documentation does not indicate default admin credentials | complexity:easy priority:high topic:authentication topic:docs topic:operations | **Component**:
'documentation'
**What happened**:
After deploying the bootstrap node, if we follow the documentation to access the MetalK8s Admin UI, we cannot login: https://metal-k8s.readthedocs.io/en/development-2.6/installation/services.html#metalk8s-gui
Actually the doc is still referring to admin/admin default creds but this has changed with the introduction of dex. It is now: admin@metalk8s.invalid/password
Also, I guess the login page screenshot in the doc is not the right one.
**What was expected**:
Doc should provide the right information and also provide a link to user management page if Platform Administrator wants to provision new admin user: https://metal-k8s.readthedocs.io/en/development-2.6/operation/cluster_and_service_configuration.html
Change the login screenshot.
| 1.0 | Documentation does not indicate default admin credentials - **Component**:
'documentation'
**What happened**:
After deploying the bootstrap node, if we follow the documentation to access the MetalK8s Admin UI, we cannot login: https://metal-k8s.readthedocs.io/en/development-2.6/installation/services.html#metalk8s-gui
Actually the doc is still referring to admin/admin default creds but this has changed with the introduction of dex. It is now: admin@metalk8s.invalid/password
Also, I guess the login page screenshot in the doc is not the right one.
**What was expected**:
Doc should provide the right information and also provide a link to user management page if Platform Administrator wants to provision new admin user: https://metal-k8s.readthedocs.io/en/development-2.6/operation/cluster_and_service_configuration.html
Change the login screenshot.
| non_defect | documentation does not indicate default admin credentials component documentation what happened after deploying the bootstrap node if we follow the documentation to access the admin ui we cannot login actually the doc is still referring to admin admin default creds but this has changed with the introduction of dex it is now admin invalid password also i guess the login page screenshot in the doc is not the right one what was expected doc should provide the right information and also provide a link to user management page if platform administrator wants to provision new admin user change the login screenshot | 0 |
45,935 | 13,055,825,202 | IssuesEvent | 2020-07-30 02:50:52 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | Tests should dump output to per project directories (Trac #292) | Incomplete Migration Migrated from Trac combo reconstruction defect | Migrated from https://code.icecube.wisc.edu/ticket/292
```json
{
"status": "closed",
"changetime": "2015-07-07T22:32:33",
"description": "Duplicate of #291",
"reporter": "kislat",
"cc": "",
"resolution": "duplicate",
"_ts": "1436308353324715",
"component": "combo reconstruction",
"summary": "Tests should dump output to per project directories",
"priority": "normal",
"keywords": "",
"time": "2011-07-19T07:02:52",
"milestone": "",
"owner": "kislat",
"type": "defect"
}
```
| 1.0 | Tests should dump output to per project directories (Trac #292) - Migrated from https://code.icecube.wisc.edu/ticket/292
```json
{
"status": "closed",
"changetime": "2015-07-07T22:32:33",
"description": "Duplicate of #291",
"reporter": "kislat",
"cc": "",
"resolution": "duplicate",
"_ts": "1436308353324715",
"component": "combo reconstruction",
"summary": "Tests should dump output to per project directories",
"priority": "normal",
"keywords": "",
"time": "2011-07-19T07:02:52",
"milestone": "",
"owner": "kislat",
"type": "defect"
}
```
| defect | tests should dump output to per project directories trac migrated from json status closed changetime description duplicate of reporter kislat cc resolution duplicate ts component combo reconstruction summary tests should dump output to per project directories priority normal keywords time milestone owner kislat type defect | 1 |
202,321 | 15,280,734,790 | IssuesEvent | 2021-02-23 06:55:35 | projectcontour/contour | https://api.github.com/repos/projectcontour/contour | closed | Add promlint analyzer for metrics static analysis. | area/testing good first issue help wanted release-note | **Please describe the problem you have**
Add [promlint](https://godoc.org/github.com/prometheus/client_golang/prometheus/testutil/promlint) static analysis to make sure we are using Prometheus metrics conventions consistently. | 1.0 | Add promlint analyzer for metrics static analysis. - **Please describe the problem you have**
Add [promlint](https://godoc.org/github.com/prometheus/client_golang/prometheus/testutil/promlint) static analysis to make sure we are using Prometheus metrics conventions consistently. | non_defect | add promlint analyzer for metrics static analysis please describe the problem you have add static analysis to make sure we are using prometheus metrics conventions consistently | 0 |
61,389 | 17,023,682,128 | IssuesEvent | 2021-07-03 03:16:41 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Add rendering of tourism=chalet | Component: mapnik Priority: minor Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 2.00pm, Tuesday, 15th February 2011]**
tourism=chalet isn't rendered. Maybe it could done like the newly added tourism=motel. | 1.0 | Add rendering of tourism=chalet - **[Submitted to the original trac issue database at 2.00pm, Tuesday, 15th February 2011]**
tourism=chalet isn't rendered. Maybe it could done like the newly added tourism=motel. | defect | add rendering of tourism chalet tourism chalet isn t rendered maybe it could done like the newly added tourism motel | 1 |
125,901 | 16,855,886,481 | IssuesEvent | 2021-06-21 06:35:49 | npocccties/chibichilo | https://api.github.com/repos/npocccties/chibichilo | opened | [ブック視聴画面]モバイル(1カラム)表示かつ動画の切り替わり時に動画が画面内に表示されることを保証する | design enhancement | ### どのような問題を解決する機能追加でしょうか?
モバイル表示時に手動でトピックを切り替えるには、動画が表示される領域より下までスクロールしてトピック・セクションの一覧をクリックする必要がある。
クリックしてトピックを切り替えられたとしても、切り替わった動画を表示するには、上にスクロールして動画のスクロール追従される領域に戻る必要がある。
### あなたが望む解決方法について書いてください
トピックが切り替わった際、動画が画面内に表示されていなければ、動画が画面内に表示されるように自動的にスクロールする
### あなたが考えた別解について書いてください
問題について深く考えると、複数の案を検討しているはずです。その取捨選択した別解について、明解かつ簡潔に書いてください。
### 本 issue を close する条件
本件を、終了(close)とする条件について下記から選ぶか、追記をお願いします。
- [x] 議論の結論がでたか? (一定期間結論がでないものは、close 対象となりうる)
- [x] コードに実装された。
- [ ] ドキュメントに反映され、周知された。
- [ ] 議論が outdate になってしまった。(コメントで特記し、close)
### 追記・備考
追記すべきことや、機能追加についての資料(スクリーンショットなど)などを追記してください。
| 1.0 | [ブック視聴画面]モバイル(1カラム)表示かつ動画の切り替わり時に動画が画面内に表示されることを保証する - ### どのような問題を解決する機能追加でしょうか?
モバイル表示時に手動でトピックを切り替えるには、動画が表示される領域より下までスクロールしてトピック・セクションの一覧をクリックする必要がある。
クリックしてトピックを切り替えられたとしても、切り替わった動画を表示するには、上にスクロールして動画のスクロール追従される領域に戻る必要がある。
### あなたが望む解決方法について書いてください
トピックが切り替わった際、動画が画面内に表示されていなければ、動画が画面内に表示されるように自動的にスクロールする
### あなたが考えた別解について書いてください
問題について深く考えると、複数の案を検討しているはずです。その取捨選択した別解について、明解かつ簡潔に書いてください。
### 本 issue を close する条件
本件を、終了(close)とする条件について下記から選ぶか、追記をお願いします。
- [x] 議論の結論がでたか? (一定期間結論がでないものは、close 対象となりうる)
- [x] コードに実装された。
- [ ] ドキュメントに反映され、周知された。
- [ ] 議論が outdate になってしまった。(コメントで特記し、close)
### 追記・備考
追記すべきことや、機能追加についての資料(スクリーンショットなど)などを追記してください。
| non_defect | モバイル 表示かつ動画の切り替わり時に動画が画面内に表示されることを保証する どのような問題を解決する機能追加でしょうか モバイル表示時に手動でトピックを切り替えるには、動画が表示される領域より下までスクロールしてトピック・セクションの一覧をクリックする必要がある。 クリックしてトピックを切り替えられたとしても、切り替わった動画を表示するには、上にスクロールして動画のスクロール追従される領域に戻る必要がある。 あなたが望む解決方法について書いてください トピックが切り替わった際、動画が画面内に表示されていなければ、動画が画面内に表示されるように自動的にスクロールする あなたが考えた別解について書いてください 問題について深く考えると、複数の案を検討しているはずです。その取捨選択した別解について、明解かつ簡潔に書いてください。 本 issue を close する条件 本件を、終了 close とする条件について下記から選ぶか、追記をお願いします。 議論の結論がでたか 一定期間結論がでないものは、close 対象となりうる コードに実装された。 ドキュメントに反映され、周知された。 議論が outdate になってしまった。 コメントで特記し、close 追記・備考 追記すべきことや、機能追加についての資料 スクリーンショットなど などを追記してください。 | 0 |
818 | 2,594,128,406 | IssuesEvent | 2015-02-20 00:00:29 | BALL-Project/ball | https://api.github.com/repos/BALL-Project/ball | closed | SMARTS matcher does strange things... | C: BALL Core P: major R: invalid T: defect | **Reported by anhi on 22 Jun 38777177 06:40 UTC**
With the current implementation of the SMARTS matcher, it is not possible to match hydrogen atoms.
If I understand the SMARTS specification right, [should match each hydrogen atom. Applying [H](H]) to
H2O, e.g., does not match anything, though. [O] works as expected. I am not sure which other
elements are affected, but I am pretty sure it is due to the grammar we employ. | 1.0 | SMARTS matcher does strange things... - **Reported by anhi on 22 Jun 38777177 06:40 UTC**
With the current implementation of the SMARTS matcher, it is not possible to match hydrogen atoms.
If I understand the SMARTS specification right, [should match each hydrogen atom. Applying [H](H]) to
H2O, e.g., does not match anything, though. [O] works as expected. I am not sure which other
elements are affected, but I am pretty sure it is due to the grammar we employ. | defect | smarts matcher does strange things reported by anhi on jun utc with the current implementation of the smarts matcher it is not possible to match hydrogen atoms if i understand the smarts specification right h to e g does not match anything though works as expected i am not sure which other elements are affected but i am pretty sure it is due to the grammar we employ | 1 |
143,837 | 19,256,448,851 | IssuesEvent | 2021-12-09 11:50:05 | tildabio/composable | https://api.github.com/repos/tildabio/composable | opened | CVE-2021-27918 (High) detected in multiple libraries | security vulnerability | ## CVE-2021-27918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>github.com/golang/net-ca1201d0de80cfde86cb01aea620983605dfe99b</b>, <b>github.com/kubernetes-retired/testing_frameworks-v0.1.1</b>, <b>github.com/golang/net-65e2d4e15006aab9813ff8769e768bbf4bb667a0</b></p></summary>
<p>
<details><summary><b>github.com/golang/net-ca1201d0de80cfde86cb01aea620983605dfe99b</b></p></summary>
<p>[mirror] Go supplementary network libraries</p>
<p>
Dependency Hierarchy:
- github.com/onsi/gomega-v1.5.0 (Root Library)
- :x: **github.com/golang/net-ca1201d0de80cfde86cb01aea620983605dfe99b** (Vulnerable Library)
</details>
<details><summary><b>github.com/kubernetes-retired/testing_frameworks-v0.1.1</b></p></summary>
<p>[EOL] test frameworks for testing kubernetes</p>
<p>
Dependency Hierarchy:
- github.com/kubernetes-sigs/controller-runtime-v0.2.0 (Root Library)
- :x: **github.com/kubernetes-retired/testing_frameworks-v0.1.1** (Vulnerable Library)
</details>
<details><summary><b>github.com/golang/net-65e2d4e15006aab9813ff8769e768bbf4bb667a0</b></p></summary>
<p>[mirror] Go supplementary network libraries</p>
<p>
Dependency Hierarchy:
- github.com/onsi/gomega-v1.4.2 (Root Library)
- :x: **github.com/golang/net-65e2d4e15006aab9813ff8769e768bbf4bb667a0** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/tildabio/composable/commit/af8aa41dc3cfebd35daec7382d85fd4b238fe08c">af8aa41dc3cfebd35daec7382d85fd4b238fe08c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
encoding/xml in Go before 1.15.9 and 1.16.x before 1.16.1 has an infinite loop if a custom TokenReader (for xml.NewTokenDecoder) returns EOF in the middle of an element. This can occur in the Decode, DecodeElement, or Skip method.
<p>Publish Date: 2021-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27918>CVE-2021-27918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://groups.google.com/g/golang-announce/c/MfiLYjG-RAw">https://groups.google.com/g/golang-announce/c/MfiLYjG-RAw</a></p>
<p>Release Date: 2021-03-11</p>
<p>Fix Resolution: 1.15.9, 1.16.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-27918 (High) detected in multiple libraries - ## CVE-2021-27918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>github.com/golang/net-ca1201d0de80cfde86cb01aea620983605dfe99b</b>, <b>github.com/kubernetes-retired/testing_frameworks-v0.1.1</b>, <b>github.com/golang/net-65e2d4e15006aab9813ff8769e768bbf4bb667a0</b></p></summary>
<p>
<details><summary><b>github.com/golang/net-ca1201d0de80cfde86cb01aea620983605dfe99b</b></p></summary>
<p>[mirror] Go supplementary network libraries</p>
<p>
Dependency Hierarchy:
- github.com/onsi/gomega-v1.5.0 (Root Library)
- :x: **github.com/golang/net-ca1201d0de80cfde86cb01aea620983605dfe99b** (Vulnerable Library)
</details>
<details><summary><b>github.com/kubernetes-retired/testing_frameworks-v0.1.1</b></p></summary>
<p>[EOL] test frameworks for testing kubernetes</p>
<p>
Dependency Hierarchy:
- github.com/kubernetes-sigs/controller-runtime-v0.2.0 (Root Library)
- :x: **github.com/kubernetes-retired/testing_frameworks-v0.1.1** (Vulnerable Library)
</details>
<details><summary><b>github.com/golang/net-65e2d4e15006aab9813ff8769e768bbf4bb667a0</b></p></summary>
<p>[mirror] Go supplementary network libraries</p>
<p>
Dependency Hierarchy:
- github.com/onsi/gomega-v1.4.2 (Root Library)
- :x: **github.com/golang/net-65e2d4e15006aab9813ff8769e768bbf4bb667a0** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/tildabio/composable/commit/af8aa41dc3cfebd35daec7382d85fd4b238fe08c">af8aa41dc3cfebd35daec7382d85fd4b238fe08c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
encoding/xml in Go before 1.15.9 and 1.16.x before 1.16.1 has an infinite loop if a custom TokenReader (for xml.NewTokenDecoder) returns EOF in the middle of an element. This can occur in the Decode, DecodeElement, or Skip method.
<p>Publish Date: 2021-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27918>CVE-2021-27918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://groups.google.com/g/golang-announce/c/MfiLYjG-RAw">https://groups.google.com/g/golang-announce/c/MfiLYjG-RAw</a></p>
<p>Release Date: 2021-03-11</p>
<p>Fix Resolution: 1.15.9, 1.16.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries github com golang net github com kubernetes retired testing frameworks github com golang net github com golang net go supplementary network libraries dependency hierarchy github com onsi gomega root library x github com golang net vulnerable library github com kubernetes retired testing frameworks test frameworks for testing kubernetes dependency hierarchy github com kubernetes sigs controller runtime root library x github com kubernetes retired testing frameworks vulnerable library github com golang net go supplementary network libraries dependency hierarchy github com onsi gomega root library x github com golang net vulnerable library found in head commit a href found in base branch main vulnerability details encoding xml in go before and x before has an infinite loop if a custom tokenreader for xml newtokendecoder returns eof in the middle of an element this can occur in the decode decodeelement or skip method publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
53,266 | 13,261,315,945 | IssuesEvent | 2020-08-20 19:40:20 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | I3MCTree missing in hdf5 and root files (Trac #1082) | Migrated from Trac combo reconstruction defect | When I write i3-files into hdf5 and/or root files I3MCTree is missing. The files are only copied, there are no cuts or other processing involved and everything is booked (BookEverything = True). This issue doesn't occur with icerec release V04-06-00, but with V04-11-02, so it changed somewhere in between.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1082">https://code.icecube.wisc.edu/projects/icecube/ticket/1082</a>, reported by lisaand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-08-10T20:31:17",
"_ts": "1439238677802127",
"description": "When I write i3-files into hdf5 and/or root files I3MCTree is missing. The files are only copied, there are no cuts or other processing involved and everything is booked (BookEverything = True). This issue doesn't occur with icerec release V04-06-00, but with V04-11-02, so it changed somewhere in between.",
"reporter": "lisa",
"cc": "",
"resolution": "fixed",
"time": "2015-08-03T12:29:20",
"component": "combo reconstruction",
"summary": "I3MCTree missing in hdf5 and root files",
"priority": "blocker",
"keywords": "I3MCTree, I3TableWriter",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| 1.0 | I3MCTree missing in hdf5 and root files (Trac #1082) - When I write i3-files into hdf5 and/or root files I3MCTree is missing. The files are only copied, there are no cuts or other processing involved and everything is booked (BookEverything = True). This issue doesn't occur with icerec release V04-06-00, but with V04-11-02, so it changed somewhere in between.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1082">https://code.icecube.wisc.edu/projects/icecube/ticket/1082</a>, reported by lisaand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-08-10T20:31:17",
"_ts": "1439238677802127",
"description": "When I write i3-files into hdf5 and/or root files I3MCTree is missing. The files are only copied, there are no cuts or other processing involved and everything is booked (BookEverything = True). This issue doesn't occur with icerec release V04-06-00, but with V04-11-02, so it changed somewhere in between.",
"reporter": "lisa",
"cc": "",
"resolution": "fixed",
"time": "2015-08-03T12:29:20",
"component": "combo reconstruction",
"summary": "I3MCTree missing in hdf5 and root files",
"priority": "blocker",
"keywords": "I3MCTree, I3TableWriter",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
| defect | missing in and root files trac when i write files into and or root files is missing the files are only copied there are no cuts or other processing involved and everything is booked bookeverything true this issue doesn t occur with icerec release but with so it changed somewhere in between migrated from json status closed changetime ts description when i write files into and or root files is missing the files are only copied there are no cuts or other processing involved and everything is booked bookeverything true this issue doesn t occur with icerec release but with so it changed somewhere in between reporter lisa cc resolution fixed time component combo reconstruction summary missing in and root files priority blocker keywords milestone owner olivas type defect | 1 |
66,767 | 12,824,857,265 | IssuesEvent | 2020-07-06 14:07:13 | MeAmAnUsername/pie | https://api.github.com/repos/MeAmAnUsername/pie | opened | remove try-p2j-ast-exp | Component: code base Priority: low Status: specified Type: enhancement | It turns errors at stratego compile time into errors at Java compile time / Java runtime, which is bad. | 1.0 | remove try-p2j-ast-exp - It turns errors at stratego compile time into errors at Java compile time / Java runtime, which is bad. | non_defect | remove try ast exp it turns errors at stratego compile time into errors at java compile time java runtime which is bad | 0 |
29,732 | 5,846,034,465 | IssuesEvent | 2017-05-10 15:22:54 | idaholab/raven | https://api.github.com/repos/idaholab/raven | closed | Image tester cause crash when images are not the same size | defect priority_normal | --------
Issue Description
--------
##### What did you expect to see happen?
The image tester to return a failure.
##### What did you see instead?
The image tester failed:
```
Traceback (most recent call last):
File "./backend_run_tests", line 128, in <module>
TestHarness.buildAndRun(sys.argv, app_name, MOOSE_DIR)
File "/home/jjc/sub/raven/moose/python/TestHarness/TestHarness.py", line 34, in buildAndRun
harness.findAndRunTests()
File "/home/jjc/sub/raven/moose/python/TestHarness/TestHarness.py", line 192, in findAndRunTests
self.launchTest(find_only, file, dirpath, testers)
File "/home/jjc/sub/raven/moose/python/TestHarness/TestHarness.py", line 274, in launchTest
self.runner.run(tester, command)
File "/home/jjc/sub/raven/moose/python/TestHarness/RunParallel.py", line 107, in run
self.spinwait()
File "/home/jjc/sub/raven/moose/python/TestHarness/RunParallel.py", line 282, in spinwait
self.returnToTestHarness(job_index)
File "/home/jjc/sub/raven/moose/python/TestHarness/RunParallel.py", line 240, in returnToTestHarness
output = tester.processResults(tester.specs['moose_dir'], p.returncode, self.options, output)
File "scripts/TestHarness/testers/RavenFramework.py", line 184, in processResults
return self.rawProcessResults(moose_dir, retcode, options, output)
File "scripts/TestHarness/testers/RavenFramework.py", line 262, in rawProcessResults
(imgSame,imgMessages) = imgDiff.diff()
File "scripts/TestHarness/testers/RAVENImageDiff.py", line 86, in diff
self.__messages += 'Gold and test image are not the same shape: '+str(goldImage.shape)+', '+testImage.shape
TypeError: coercing to Unicode: need string or buffer, tuple found
```
##### Do you have a suggested fix for the development team?
##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or improvement?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. __because this is a defect in the error reporting of the test harness, it cannot result in incorrect results for users__
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [x] 1. If the issue is a defect, is the defect fixed?
- [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) __there is not currently a standard method for testing the regression test system__
- [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? __will not impact users directly or frequently__
- [x] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [x] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
| 1.0 | Image tester cause crash when images are not the same size - --------
Issue Description
--------
##### What did you expect to see happen?
The image tester to return a failure.
##### What did you see instead?
The image tester failed:
```
Traceback (most recent call last):
File "./backend_run_tests", line 128, in <module>
TestHarness.buildAndRun(sys.argv, app_name, MOOSE_DIR)
File "/home/jjc/sub/raven/moose/python/TestHarness/TestHarness.py", line 34, in buildAndRun
harness.findAndRunTests()
File "/home/jjc/sub/raven/moose/python/TestHarness/TestHarness.py", line 192, in findAndRunTests
self.launchTest(find_only, file, dirpath, testers)
File "/home/jjc/sub/raven/moose/python/TestHarness/TestHarness.py", line 274, in launchTest
self.runner.run(tester, command)
File "/home/jjc/sub/raven/moose/python/TestHarness/RunParallel.py", line 107, in run
self.spinwait()
File "/home/jjc/sub/raven/moose/python/TestHarness/RunParallel.py", line 282, in spinwait
self.returnToTestHarness(job_index)
File "/home/jjc/sub/raven/moose/python/TestHarness/RunParallel.py", line 240, in returnToTestHarness
output = tester.processResults(tester.specs['moose_dir'], p.returncode, self.options, output)
File "scripts/TestHarness/testers/RavenFramework.py", line 184, in processResults
return self.rawProcessResults(moose_dir, retcode, options, output)
File "scripts/TestHarness/testers/RavenFramework.py", line 262, in rawProcessResults
(imgSame,imgMessages) = imgDiff.diff()
File "scripts/TestHarness/testers/RAVENImageDiff.py", line 86, in diff
self.__messages += 'Gold and test image are not the same shape: '+str(goldImage.shape)+', '+testImage.shape
TypeError: coercing to Unicode: need string or buffer, tuple found
```
##### Do you have a suggested fix for the development team?
##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or improvement?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. __because this is a defect in the error reporting of the test harness, it cannot result in incorrect results for users__
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [x] 1. If the issue is a defect, is the defect fixed?
- [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) __there is not currently a standard method for testing the regression test system__
- [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? __will not impact users directly or frequently__
- [x] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [x] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
| defect | image tester cause crash when images are not the same size issue description what did you expect to see happen the image tester to return a failure what did you see instead the image tester failed traceback most recent call last file backend run tests line in testharness buildandrun sys argv app name moose dir file home jjc sub raven moose python testharness testharness py line in buildandrun harness findandruntests file home jjc sub raven moose python testharness testharness py line in findandruntests self launchtest find only file dirpath testers file home jjc sub raven moose python testharness testharness py line in launchtest self runner run tester command file home jjc sub raven moose python testharness runparallel py line in run self spinwait file home jjc sub raven moose python testharness runparallel py line in spinwait self returntotestharness job index file home jjc sub raven moose python testharness runparallel py line in returntotestharness output tester processresults tester specs p returncode self options output file scripts testharness testers ravenframework py line in processresults return self rawprocessresults moose dir retcode options output file scripts testharness testers ravenframework py line in rawprocessresults imgsame imgmessages imgdiff diff file scripts testharness testers ravenimagediff py line in diff self messages gold and test image are not the same shape str goldimage shape testimage shape typeerror coercing to unicode need string or buffer tuple found do you have a suggested fix for the development team please attach the input file s that generate this error the simpler the input the faster we can find the issue for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users because this is a defect in the error reporting of the test harness it cannot result in incorrect results for users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not there is not currently a standard method for testing the regression test system if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master will not impact users directly or frequently if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided | 1 |
390,849 | 11,564,709,373 | IssuesEvent | 2020-02-20 09:11:12 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | opened | How to customize redirect url on 404? | Priority/Normal Type/Question | ### Description:
I want on 404 error redirect to a custom page and prevent redirecting to `/carbon/admin/login.jsp`.
### Steps to reproduce:
### Affected Product Version:
I am using AM v3.0.0.
### Environment details (with versions):
- OS: Linux
- Client: Ubuntu 18.04
- Env (Docker/K8s): none
---
### Optional Fields
#### Related Issues:
#### Suggested Labels:
#### Suggested Assignees: | 1.0 | How to customize redirect url on 404? - ### Description:
I want on 404 error redirect to a custom page and prevent redirecting to `/carbon/admin/login.jsp`.
### Steps to reproduce:
### Affected Product Version:
I am using AM v3.0.0.
### Environment details (with versions):
- OS: Linux
- Client: Ubuntu 18.04
- Env (Docker/K8s): none
---
### Optional Fields
#### Related Issues:
#### Suggested Labels:
#### Suggested Assignees: | non_defect | how to customize redirect url on description i want on error redirect to a custom page and prevent redirecting to carbon admin login jsp steps to reproduce affected product version i am using am environment details with versions os linux client ubuntu env docker none optional fields related issues suggested labels suggested assignees | 0 |
71,662 | 15,207,828,663 | IssuesEvent | 2021-02-17 01:06:51 | billmcchesney1/foxtrot | https://api.github.com/repos/billmcchesney1/foxtrot | closed | CVE-2019-17571 (High) detected in log4j-1.2.17.jar - autoclosed | security vulnerability | ## CVE-2019-17571 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary>
<p>Apache Log4j 1.2</p>
<p>Path to dependency file: foxtrot/foxtrot-translator/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p>
<p>
Dependency Hierarchy:
- hbase-server-1.2.1.jar (Root Library)
- :x: **log4j-1.2.17.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17571">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17571</a></p>
<p>Release Date: 2019-12-20</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","packageFilePaths":["/foxtrot-translator/pom.xml","/foxtrot-server/pom.xml","/foxtrot-core/pom.xml","/foxtrot-sql/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.hbase:hbase-server:1.2.1;log4j:log4j:1.2.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.logging.log4j:log4j-core:2.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-17571","vulnerabilityDetails":"Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-17571 (High) detected in log4j-1.2.17.jar - autoclosed - ## CVE-2019-17571 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary>
<p>Apache Log4j 1.2</p>
<p>Path to dependency file: foxtrot/foxtrot-translator/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p>
<p>
Dependency Hierarchy:
- hbase-server-1.2.1.jar (Root Library)
- :x: **log4j-1.2.17.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17571">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17571</a></p>
<p>Release Date: 2019-12-20</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","packageFilePaths":["/foxtrot-translator/pom.xml","/foxtrot-server/pom.xml","/foxtrot-core/pom.xml","/foxtrot-sql/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.hbase:hbase-server:1.2.1;log4j:log4j:1.2.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.logging.log4j:log4j-core:2.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-17571","vulnerabilityDetails":"Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_defect | cve high detected in jar autoclosed cve high severity vulnerability vulnerable library jar apache path to dependency file foxtrot foxtrot translator pom xml path to vulnerable library home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar home wss scanner repository jar dependency hierarchy hbase server jar root library x jar vulnerable library found in base branch master vulnerability details included in is a socketserver class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data this affects versions up to up to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache hbase hbase server isminimumfixversionavailable true minimumfixversion org apache logging core basebranches vulnerabilityidentifier cve vulnerabilitydetails included in is a socketserver class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data this affects versions up to up to vulnerabilityurl | 0 |
24,996 | 4,165,060,911 | IssuesEvent | 2016-06-19 07:51:59 | nanopb/nanopb | https://api.github.com/repos/nanopb/nanopb | closed | Message size not calculated if a submessage includes bytes | Component-Generator FixedInGit Priority-Medium Type-Defect | **Steps to reproduce the issue**
Compile the following .proto file:
```
message TestMessage {
required uint32 id = 1;
required bytes payload = 2;
}
message EncapsulatedMessage {
required uint32 id = 1;
required TestMessage test = 2;
}
```
(On Windows)
```
> nanopb-0.3.5-windows-x86\generator-bin\protoc.exe --nanopb_out=. test.proto
```
The header file contains a calculated message size that references the undefined size of TestMessage:
```
/* Maximum encoded size of messages (where known) */
#define EncapsulatedMessage_size (12 + TestMessage_size)
```
**What happens?**
The header file fails to compile.
```
test.pb.h:50:56: error: 'TestMessage_size' undeclared here (not in a function)
```
This may be related to #189 and #172 but I'm not using options to specify a size. I'm using the callback interface. | 1.0 | Message size not calculated if a submessage includes bytes - **Steps to reproduce the issue**
Compile the following .proto file:
```
message TestMessage {
required uint32 id = 1;
required bytes payload = 2;
}
message EncapsulatedMessage {
required uint32 id = 1;
required TestMessage test = 2;
}
```
(On Windows)
```
> nanopb-0.3.5-windows-x86\generator-bin\protoc.exe --nanopb_out=. test.proto
```
The header file contains a calculated message size that references the undefined size of TestMessage:
```
/* Maximum encoded size of messages (where known) */
#define EncapsulatedMessage_size (12 + TestMessage_size)
```
**What happens?**
The header file fails to compile.
```
test.pb.h:50:56: error: 'TestMessage_size' undeclared here (not in a function)
```
This may be related to #189 and #172 but I'm not using options to specify a size. I'm using the callback interface. | defect | message size not calculated if a submessage includes bytes steps to reproduce the issue compile the following proto file message testmessage required id required bytes payload message encapsulatedmessage required id required testmessage test on windows nanopb windows generator bin protoc exe nanopb out test proto the header file contains a calculated message size that references the undefined size of testmessage maximum encoded size of messages where known define encapsulatedmessage size testmessage size what happens the header file fails to compile test pb h error testmessage size undeclared here not in a function this may be related to and but i m not using options to specify a size i m using the callback interface | 1 |
1,556 | 2,603,967,556 | IssuesEvent | 2015-02-24 18:59:24 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳疱疹治疗 | auto-migrated Priority-Medium Type-Defect | ```
沈阳疱疹治疗〓沈陽軍區政治部醫院性病〓TEL:024-31023308〓��
�立于1946年,68年專注于性傳播疾病的研究和治療。位于沈陽�
��沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史悠�
��、設備精良、技術權威、專家云集,是預防、保健、醫療、
科研康復為一體的綜合性醫院。是國家首批公立甲等部隊醫��
�、全國首批醫療規范定點單位,是第四軍醫大學、東南大學�
��知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤部
衛生部評為衛生工作先進單位,先后兩次榮立集體二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:11 | 1.0 | 沈阳疱疹治疗 - ```
沈阳疱疹治疗〓沈陽軍區政治部醫院性病〓TEL:024-31023308〓��
�立于1946年,68年專注于性傳播疾病的研究和治療。位于沈陽�
��沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史悠�
��、設備精良、技術權威、專家云集,是預防、保健、醫療、
科研康復為一體的綜合性醫院。是國家首批公立甲等部隊醫��
�、全國首批醫療規范定點單位,是第四軍醫大學、東南大學�
��知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤部
衛生部評為衛生工作先進單位,先后兩次榮立集體二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:11 | defect | 沈阳疱疹治疗 沈阳疱疹治疗〓沈陽軍區政治部醫院性病〓tel: 〓�� � , 。位于沈陽� �� 。是一所與新中國同建立共輝煌的歷史悠� ��、設備精良、技術權威、專家云集,是預防、保健、醫療、 科研康復為一體的綜合性醫院。是國家首批公立甲等部隊醫�� �、全國首批醫療規范定點單位,是第四軍醫大學、東南大學� ��知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤部 衛生部評為衛生工作先進單位,先后兩次榮立集體二等功。 original issue reported on code google com by gmail com on jun at | 1 |
116,444 | 9,852,719,318 | IssuesEvent | 2019-06-19 13:27:25 | BEXIS2/Core | https://api.github.com/repos/BEXIS2/Core | closed | Main Menu - 2 or more line header overlay content | Priority: Low Status: Testing Required Type: Bug | once you have several additional modules in the header it is very likely the header span over two or more lines. Currently this is a overlay on the main content and makes the first lines not accessible.
| 1.0 | Main Menu - 2 or more line header overlay content - once you have several additional modules in the header it is very likely the header span over two or more lines. Currently this is a overlay on the main content and makes the first lines not accessible.
| non_defect | main menu or more line header overlay content once you have several additional modules in the header it is very likely the header span over two or more lines currently this is a overlay on the main content and makes the first lines not accessible | 0 |
596,827 | 18,144,668,704 | IssuesEvent | 2021-09-25 07:46:36 | joonas-yoon/boj-extended | https://api.github.com/repos/joonas-yoon/boj-extended | closed | 기본 크기 옵션 선택시 여백이 없어집니다. | 👍 enhancement priority: 4 | <!-- 👋 안녕하세요, 제보해주셔서 감사합니다! -->
## 내용
<!-- 어떤 일이 있었는지 간단하게 적어주세요 -->
기본 크기 옵션일 때 창 너비를 좁게 만들면 양쪽 여백이 없습니다. 근데 넓게 보기 옵션일 때는 정상적으로 여백이 있습니다.
## 환경
<!--- Windows 버전, 사용하시는 브라우저 버전 등 -->
Chrome 93.0.4577.82 (arm64)
## 오류 메시지 또는 스크린샷
<!-- 개발자 도구(F12) - 콘솔(Console)에 나온 오류 메시지를 적어주시면, 더욱 정확하게 해결할 수 있습니다. -->
* 기본 크기

* 넓은 크기일 때 ( or 확장프로그램 비활성화 시)

| 1.0 | 기본 크기 옵션 선택시 여백이 없어집니다. - <!-- 👋 안녕하세요, 제보해주셔서 감사합니다! -->
## 내용
<!-- 어떤 일이 있었는지 간단하게 적어주세요 -->
기본 크기 옵션일 때 창 너비를 좁게 만들면 양쪽 여백이 없습니다. 근데 넓게 보기 옵션일 때는 정상적으로 여백이 있습니다.
## 환경
<!--- Windows 버전, 사용하시는 브라우저 버전 등 -->
Chrome 93.0.4577.82 (arm64)
## 오류 메시지 또는 스크린샷
<!-- 개발자 도구(F12) - 콘솔(Console)에 나온 오류 메시지를 적어주시면, 더욱 정확하게 해결할 수 있습니다. -->
* 기본 크기

* 넓은 크기일 때 ( or 확장프로그램 비활성화 시)

| non_defect | 기본 크기 옵션 선택시 여백이 없어집니다 내용 기본 크기 옵션일 때 창 너비를 좁게 만들면 양쪽 여백이 없습니다 근데 넓게 보기 옵션일 때는 정상적으로 여백이 있습니다 환경 chrome 오류 메시지 또는 스크린샷 기본 크기 넓은 크기일 때 or 확장프로그램 비활성화 시 | 0 |
160,117 | 6,076,848,751 | IssuesEvent | 2017-06-16 00:59:56 | Supadog/DB_iti | https://api.github.com/repos/Supadog/DB_iti | closed | Transcripts: Option to remove grade and honours | Medium priority | - Add option in variables to choose if grade and honours should be displayed on transcript
- E.g. SG students should not have that on their transcripts. | 1.0 | Transcripts: Option to remove grade and honours - - Add option in variables to choose if grade and honours should be displayed on transcript
- E.g. SG students should not have that on their transcripts. | non_defect | transcripts option to remove grade and honours add option in variables to choose if grade and honours should be displayed on transcript e g sg students should not have that on their transcripts | 0 |
267,480 | 28,509,055,038 | IssuesEvent | 2023-04-19 01:31:24 | dpteam/RK3188_TABLET | https://api.github.com/repos/dpteam/RK3188_TABLET | closed | CVE-2021-32078 (High) detected in linuxv3.0 - autoclosed | Mend: dependency security vulnerability | ## CVE-2021-32078 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/arm/mach-footbridge/personal-pci.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/arm/mach-footbridge/personal-pci.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An Out-of-Bounds Read was discovered in arch/arm/mach-footbridge/personal-pci.c in the Linux kernel through 5.12.11 because of the lack of a check for a value that shouldn't be negative, e.g., access to element -2 of an array, aka CID-298a58e165e4.
<p>Publish Date: 2021-06-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-32078>CVE-2021-32078</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-32078">https://www.linuxkernelcves.com/cves/CVE-2021-32078</a></p>
<p>Release Date: 2021-06-17</p>
<p>Fix Resolution: v5.13-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-32078 (High) detected in linuxv3.0 - autoclosed - ## CVE-2021-32078 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/arm/mach-footbridge/personal-pci.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/arm/mach-footbridge/personal-pci.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An Out-of-Bounds Read was discovered in arch/arm/mach-footbridge/personal-pci.c in the Linux kernel through 5.12.11 because of the lack of a check for a value that shouldn't be negative, e.g., access to element -2 of an array, aka CID-298a58e165e4.
<p>Publish Date: 2021-06-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-32078>CVE-2021-32078</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-32078">https://www.linuxkernelcves.com/cves/CVE-2021-32078</a></p>
<p>Release Date: 2021-06-17</p>
<p>Fix Resolution: v5.13-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in autoclosed cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch master vulnerable source files arch arm mach footbridge personal pci c arch arm mach footbridge personal pci c vulnerability details an out of bounds read was discovered in arch arm mach footbridge personal pci c in the linux kernel through because of the lack of a check for a value that shouldn t be negative e g access to element of an array aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
521,400 | 15,109,087,276 | IssuesEvent | 2021-02-08 17:24:01 | kubernetes-sigs/cluster-api | https://api.github.com/repos/kubernetes-sigs/cluster-api | closed | MachineHealthCheck deletes machines without NodeRef doesn't delete the node | kind/bug kind/support lifecycle/stale priority/awaiting-more-evidence | **What steps did you take and what happened:**
[A clear and concise description on how to REPRODUCE the bug.]
With MHC enabled, the noderef is not set in machine, but the machine is deleted by MHC, the node created by the machine is not deleted.
**What did you expect to happen:**
The node created by the machine should be deleted.
**Anything else you would like to add:**
[Miscellaneous information that will assist in solving the issue.]
**Environment:**
- Cluster-api version: 0.3.10
- Minikube/KIND version: NA
- Kubernetes version: (use `kubectl version`):
- OS (e.g. from `/etc/os-release`):
/kind bug
[One or more /area label. See https://github.com/kubernetes-sigs/cluster-api/labels?q=area for the list of labels]
| 1.0 | MachineHealthCheck deletes machines without NodeRef doesn't delete the node - **What steps did you take and what happened:**
[A clear and concise description on how to REPRODUCE the bug.]
With MHC enabled, the noderef is not set in machine, but the machine is deleted by MHC, the node created by the machine is not deleted.
**What did you expect to happen:**
The node created by the machine should be deleted.
**Anything else you would like to add:**
[Miscellaneous information that will assist in solving the issue.]
**Environment:**
- Cluster-api version: 0.3.10
- Minikube/KIND version: NA
- Kubernetes version: (use `kubectl version`):
- OS (e.g. from `/etc/os-release`):
/kind bug
[One or more /area label. See https://github.com/kubernetes-sigs/cluster-api/labels?q=area for the list of labels]
| non_defect | machinehealthcheck deletes machines without noderef doesn t delete the node what steps did you take and what happened with mhc enabled the noderef is not set in machine but the machine is deleted by mhc the node created by the machine is not deleted what did you expect to happen the node created by the machine should be deleted anything else you would like to add environment cluster api version minikube kind version na kubernetes version use kubectl version os e g from etc os release kind bug | 0 |
111,720 | 4,481,843,505 | IssuesEvent | 2016-08-29 01:29:30 | blakeohare/crayon | https://api.github.com/repos/blakeohare/crayon | closed | FileIO implementation needs to be moved to library supplemental code | feature request High Priority Preview Release | Additionally, UserData needs to be moved into its own library as well to ensure clear permission boundaries. | 1.0 | FileIO implementation needs to be moved to library supplemental code - Additionally, UserData needs to be moved into its own library as well to ensure clear permission boundaries. | non_defect | fileio implementation needs to be moved to library supplemental code additionally userdata needs to be moved into its own library as well to ensure clear permission boundaries | 0 |
6,132 | 2,610,221,764 | IssuesEvent | 2015-02-26 19:10:21 | chrsmith/somefinders | https://api.github.com/repos/chrsmith/somefinders | opened | audio dll для 3d инструктор | auto-migrated Priority-Medium Type-Defect | ```
'''Горислав Ильин'''
День добрый никак не могу найти .audio dll для 3d
инструктор. как то выкладывали уже
'''Виталий Орлов'''
Качай тут http://bit.ly/1ai7xLZ
'''Гвидон Павлов'''
Спасибо вроде то но просит телефон вводить
'''Вартан Анисимов'''
Не это не влияет на баланс
'''Анатолий Ершов'''
Неа все ок у меня ничего не списало
Информация о файле: audio dll для 3d инструктор
Загружен: В этом месяце
Скачан раз: 317
Рейтинг: 949
Средняя скорость скачивания: 1122
Похожих файлов: 15
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 1:12 | 1.0 | audio dll для 3d инструктор - ```
'''Горислав Ильин'''
День добрый никак не могу найти .audio dll для 3d
инструктор. как то выкладывали уже
'''Виталий Орлов'''
Качай тут http://bit.ly/1ai7xLZ
'''Гвидон Павлов'''
Спасибо вроде то но просит телефон вводить
'''Вартан Анисимов'''
Не это не влияет на баланс
'''Анатолий Ершов'''
Неа все ок у меня ничего не списало
Информация о файле: audio dll для 3d инструктор
Загружен: В этом месяце
Скачан раз: 317
Рейтинг: 949
Средняя скорость скачивания: 1122
Похожих файлов: 15
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 1:12 | defect | audio dll для инструктор горислав ильин день добрый никак не могу найти audio dll для инструктор как то выкладывали уже виталий орлов качай тут гвидон павлов спасибо вроде то но просит телефон вводить вартан анисимов не это не влияет на баланс анатолий ершов неа все ок у меня ничего не списало информация о файле audio dll для инструктор загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at | 1 |
81,648 | 31,168,359,101 | IssuesEvent | 2023-08-16 21:50:41 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Terribly frustrating :( All encrypted chats showing unable to decrypt after sign-in | T-Defect | ### Steps to reproduce
1. Where are you starting? What can you see?
I signed-in after being "forced signed-out" when app complained browser deleted files due to low on space, and when I signed back-in and gave security key, all the "encrypted chats" are full of "unable to decrypt message", unencrypted chats work.
### Outcome
#### What did you expect?
I would expect that the chat files would NOT be "deleted" and be forced signed out.
#### What happened instead?
But if that can't be avoided, I would expect app to work properly after I signed back in and verified with security key etc, yet encrypted chats say "unable to decrypt..."
UPDATE:
I did a right click on element and quit it but when I try to re-open it now, it doesn't work, won't open or maybe crashes trying to open, nothing comes up etc. Another install that uses a 'portable' element desktop and connect to different account and different server (connects to main matrix.org server) does work fine still and was not affected at all by the low space issue it seems. So currently I'm stuck and can't open/start my main element install, i may need to try and re-install/upgrade etc.
### Operating system
Windows 10
### Application version
Element version: 1.11.38 Olm version: 3.2.14
### How did you install the app?
_No response_
### Homeserver
Dendrite server, maybe give a good easy way to see version?
### Will you send logs?
No | 1.0 | Terribly frustrating :( All encrypted chats showing unable to decrypt after sign-in - ### Steps to reproduce
1. Where are you starting? What can you see?
I signed-in after being "forced signed-out" when app complained browser deleted files due to low on space, and when I signed back-in and gave security key, all the "encrypted chats" are full of "unable to decrypt message", unencrypted chats work.
### Outcome
#### What did you expect?
I would expect that the chat files would NOT be "deleted" and be forced signed out.
#### What happened instead?
But if that can't be avoided, I would expect app to work properly after I signed back in and verified with security key etc, yet encrypted chats say "unable to decrypt..."
UPDATE:
I did a right click on element and quit it but when I try to re-open it now, it doesn't work, won't open or maybe crashes trying to open, nothing comes up etc. Another install that uses a 'portable' element desktop and connect to different account and different server (connects to main matrix.org server) does work fine still and was not affected at all by the low space issue it seems. So currently I'm stuck and can't open/start my main element install, i may need to try and re-install/upgrade etc.
### Operating system
Windows 10
### Application version
Element version: 1.11.38 Olm version: 3.2.14
### How did you install the app?
_No response_
### Homeserver
Dendrite server, maybe give a good easy way to see version?
### Will you send logs?
No | defect | terribly frustrating all encrypted chats showing unable to decrypt after sign in steps to reproduce where are you starting what can you see i signed in after being forced signed out when app complained browser deleted files due to low on space and when i signed back in and gave security key all the encrypted chats are full of unable to decrypt message unencrypted chats work outcome what did you expect i would expect that the chat files would not be deleted and be forced signed out what happened instead but if that can t be avoided i would expect app to work properly after i signed back in and verified with security key etc yet encrypted chats say unable to decrypt update i did a right click on element and quit it but when i try to re open it now it doesn t work won t open or maybe crashes trying to open nothing comes up etc another install that uses a portable element desktop and connect to different account and different server connects to main matrix org server does work fine still and was not affected at all by the low space issue it seems so currently i m stuck and can t open start my main element install i may need to try and re install upgrade etc operating system windows application version element version olm version how did you install the app no response homeserver dendrite server maybe give a good easy way to see version will you send logs no | 1 |
27,195 | 2,690,941,704 | IssuesEvent | 2015-03-31 18:45:05 | ngds/ckanext-ngds | https://api.github.com/repos/ngds/ckanext-ngds | closed | need to be able to see full metadata record in XML | NGDS-v1.0 priority.Major sys.Details sys.Metadata sys.UX | maybe JSON too? Need button from Details page to see full, raw metadata. | 1.0 | need to be able to see full metadata record in XML - maybe JSON too? Need button from Details page to see full, raw metadata. | non_defect | need to be able to see full metadata record in xml maybe json too need button from details page to see full raw metadata | 0 |
220,888 | 16,988,751,533 | IssuesEvent | 2021-06-30 17:27:42 | dkeparute/3-Mergaite | https://api.github.com/repos/dkeparute/3-Mergaite | closed | Atlikti dizaino analizę | documentation | Įvertinti į kiek dalių, blokų suskirstytas bus puslapis , t.y. header'is , turinys ir kita. | 1.0 | Atlikti dizaino analizę - Įvertinti į kiek dalių, blokų suskirstytas bus puslapis , t.y. header'is , turinys ir kita. | non_defect | atlikti dizaino analizę įvertinti į kiek dalių blokų suskirstytas bus puslapis t y header is turinys ir kita | 0 |
36,872 | 8,176,703,104 | IssuesEvent | 2018-08-28 08:26:10 | hazelcast/hazelcast-cpp-client | https://api.github.com/repos/hazelcast/hazelcast-cpp-client | opened | Client Connection cannot detect idle connection to send ping and server shutdown connection | Type: Defect | Client connection should detect that a connection is reading but not sending any data to server to detect heartbeat idle case.
Current implementation checks only last read time which is not enough for reading only connections like just simple listener use case.
Issue from java client: https://github.com/hazelcast/hazelcast/issues/13576
Fix for java client : https://github.com/hazelcast/hazelcast/pull/13577 | 1.0 | Client Connection cannot detect idle connection to send ping and server shutdown connection - Client connection should detect that a connection is reading but not sending any data to server to detect heartbeat idle case.
Current implementation checks only last read time which is not enough for reading only connections like just simple listener use case.
Issue from java client: https://github.com/hazelcast/hazelcast/issues/13576
Fix for java client : https://github.com/hazelcast/hazelcast/pull/13577 | defect | client connection cannot detect idle connection to send ping and server shutdown connection client connection should detect that a connection is reading but not sending any data to server to detect heartbeat idle case current implementation checks only last read time which is not enough for reading only connections like just simple listener use case issue from java client fix for java client | 1 |
21,602 | 3,525,826,571 | IssuesEvent | 2016-01-14 00:17:53 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | opened | Isolate.spawn is broken in snapshots when run near a foreign .packages file | area-vm priority-critical Type-Defect vm-regression | This breakage appears under the following conditions:
* The script contains a `package:` import.
* The script uses `Isolate.spawn()`.
* The script is run from a snapshot.
* The directory containing the snapshot, or a parent of that directory, contains a `.packages` file that's different from the one used to create the snapshot.
To reproduce this, create a package with the following files:
```dart
// bin/bin.dart
import 'dart:isolate';
import 'package:app/app.dart';
main() async {
await Isolate.spawn(entrypoint, null);
}
void entrypoint(_) {
}
```
```dart
// lib/app.dart
// This file can be empty.
```
```yaml
// pubspec.yaml
name: app
```
Run:
```
$ pub get
$ dart --snapshot=app.dart.snapshot bin/app.dart
$ rm .packages
$ touch .packages
$ dart bin.dart.snapshot
```
You should see an error like the following:
```
Unhandled exception:
IsolateSpawnException: Unable to spawn isolate: Unhandled exception:
Load Error for "package:app/app.dart": No mapping for 'app' package when resolving 'package:app/app.dart'.
#0 _asyncLoadErrorCallback (dart:_builtin:155)
#1 _asyncLoadError (dart:_builtin:566)
#2 _loadPackage (dart:_builtin:605)
#3 _loadData (dart:_builtin:637)
#4 _loadDataAsync (dart:_builtin:657)
#5 _loadScriptCallback (dart:_builtin:153)
#6 _handleLoaderReply (dart:_builtin:370)
#7 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:148)
'file:///tmp/app/bin/bin.dart': error: line 3 pos 1: library handler failed
import 'package:app/app.dart';
^
#0 Isolate.spawn.<spawn_async_body> (dart:isolate-patch/isolate_patch.dart)
#1 _asyncErrorWrapperHelper.<anonymous closure> (dart:async-patch/async_patch.dart:34)
#2 _RootZone.runBinary (dart:async/zone.dart:1154)
#3 _Future._propagateToListeners.handleError (dart:async/future_impl.dart:579)
#4 _Future._propagateToListeners (dart:async/future_impl.dart:641)
#5 _Future._completeError (dart:async/future_impl.dart:432)
#6 _SyncCompleter._completeError (dart:async/future_impl.dart:56)
#7 _Completer.completeError (dart:async/future_impl.dart:27)
#8 Isolate._spawnCommon.<anonymous closure> (dart:isolate-patch/isolate_patch.dart:413)
#9 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:148)
```
This is a regression. Running `git bisect` indicates that this error appeared as of dart-lang/sdk@6d066c7e53f82b70c3a5bfc4916d8a0c9a26a6f4. It's likely related to dart-lang/pub#1379. | 1.0 | Isolate.spawn is broken in snapshots when run near a foreign .packages file - This breakage appears under the following conditions:
* The script contains a `package:` import.
* The script uses `Isolate.spawn()`.
* The script is run from a snapshot.
* The directory containing the snapshot, or a parent of that directory, contains a `.packages` file that's different from the one used to create the snapshot.
To reproduce this, create a package with the following files:
```dart
// bin/bin.dart
import 'dart:isolate';
import 'package:app/app.dart';
main() async {
await Isolate.spawn(entrypoint, null);
}
void entrypoint(_) {
}
```
```dart
// lib/app.dart
// This file can be empty.
```
```yaml
// pubspec.yaml
name: app
```
Run:
```
$ pub get
$ dart --snapshot=app.dart.snapshot bin/app.dart
$ rm .packages
$ touch .packages
$ dart bin.dart.snapshot
```
You should see an error like the following:
```
Unhandled exception:
IsolateSpawnException: Unable to spawn isolate: Unhandled exception:
Load Error for "package:app/app.dart": No mapping for 'app' package when resolving 'package:app/app.dart'.
#0 _asyncLoadErrorCallback (dart:_builtin:155)
#1 _asyncLoadError (dart:_builtin:566)
#2 _loadPackage (dart:_builtin:605)
#3 _loadData (dart:_builtin:637)
#4 _loadDataAsync (dart:_builtin:657)
#5 _loadScriptCallback (dart:_builtin:153)
#6 _handleLoaderReply (dart:_builtin:370)
#7 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:148)
'file:///tmp/app/bin/bin.dart': error: line 3 pos 1: library handler failed
import 'package:app/app.dart';
^
#0 Isolate.spawn.<spawn_async_body> (dart:isolate-patch/isolate_patch.dart)
#1 _asyncErrorWrapperHelper.<anonymous closure> (dart:async-patch/async_patch.dart:34)
#2 _RootZone.runBinary (dart:async/zone.dart:1154)
#3 _Future._propagateToListeners.handleError (dart:async/future_impl.dart:579)
#4 _Future._propagateToListeners (dart:async/future_impl.dart:641)
#5 _Future._completeError (dart:async/future_impl.dart:432)
#6 _SyncCompleter._completeError (dart:async/future_impl.dart:56)
#7 _Completer.completeError (dart:async/future_impl.dart:27)
#8 Isolate._spawnCommon.<anonymous closure> (dart:isolate-patch/isolate_patch.dart:413)
#9 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:148)
```
This is a regression. Running `git bisect` indicates that this error appeared as of dart-lang/sdk@6d066c7e53f82b70c3a5bfc4916d8a0c9a26a6f4. It's likely related to dart-lang/pub#1379. | defect | isolate spawn is broken in snapshots when run near a foreign packages file this breakage appears under the following conditions the script contains a package import the script uses isolate spawn the script is run from a snapshot the directory containing the snapshot or a parent of that directory contains a packages file that s different from the one used to create the snapshot to reproduce this create a package with the following files dart bin bin dart import dart isolate import package app app dart main async await isolate spawn entrypoint null void entrypoint dart lib app dart this file can be empty yaml pubspec yaml name app run pub get dart snapshot app dart snapshot bin app dart rm packages touch packages dart bin dart snapshot you should see an error like the following unhandled exception isolatespawnexception unable to spawn isolate unhandled exception load error for package app app dart no mapping for app package when resolving package app app dart asyncloaderrorcallback dart builtin asyncloaderror dart builtin loadpackage dart builtin loaddata dart builtin loaddataasync dart builtin loadscriptcallback dart builtin handleloaderreply dart builtin rawreceiveportimpl handlemessage dart isolate patch isolate patch dart file tmp app bin bin dart error line pos library handler failed import package app app dart isolate spawn dart isolate patch isolate patch dart asyncerrorwrapperhelper dart async patch async patch dart rootzone runbinary dart async zone dart future propagatetolisteners handleerror dart async future impl dart future propagatetolisteners dart async future impl dart future completeerror dart async future impl dart synccompleter completeerror dart async future impl dart completer completeerror dart async future impl dart isolate spawncommon dart isolate patch isolate patch dart rawreceiveportimpl handlemessage dart isolate patch isolate patch dart this is a regression running git bisect indicates that this error appeared as of dart lang sdk it s likely related to dart lang pub | 1 |
24,888 | 12,422,001,746 | IssuesEvent | 2020-05-23 19:42:41 | erikgrinaker/toydb | https://api.github.com/repos/erikgrinaker/toydb | closed | Improve SQL key encoding | correctness performance | All storage keys (even individual rows and index entries) currently use full identifiers for tables and columns, they should use integer identifiers instead. They must also escape separators. | True | Improve SQL key encoding - All storage keys (even individual rows and index entries) currently use full identifiers for tables and columns, they should use integer identifiers instead. They must also escape separators. | non_defect | improve sql key encoding all storage keys even individual rows and index entries currently use full identifiers for tables and columns they should use integer identifiers instead they must also escape separators | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.