Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
22,180 | 3,609,970,086 | IssuesEvent | 2016-02-05 01:40:40 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | opened | dartanalyzer reports the same error multiple times when given multiple entry points | area-analyzer Type-Defect | To reproduce, create three files:
entry1.dart
```
import "target.dart";
main() {
print(x);
}
```
entry2.dart
```
import "target.dart";
main() {
print(x);
}
```
target.dart
```
var x = 3;
asdf
```
Then run dartanalyzer on both entry files:
```
$ dartanalyzer entry1.dart entry2.dart
```
It will analyze them sequentially and print the same error twice. If given a lot of files (for example using a glob) then you get lots of duplicates. | 1.0 | dartanalyzer reports the same error multiple times when given multiple entry points - To reproduce, create three files:
entry1.dart
```
import "target.dart";
main() {
print(x);
}
```
entry2.dart
```
import "target.dart";
main() {
print(x);
}
```
target.dart
```
var x = 3;
asdf
```
Then run dartanalyzer on both entry files:
```
$ dartanalyzer entry1.dart entry2.dart
```
It will analyze them sequentially and print the same error twice. If given a lot of files (for example using a glob) then you get lots of duplicates. | defect | dartanalyzer reports the same error multiple times when given multiple entry points to reproduce create three files dart import target dart main print x dart import target dart main print x target dart var x asdf then run dartanalyzer on both entry files dartanalyzer dart dart it will analyze them sequentially and print the same error twice if given a lot of files for example using a glob then you get lots of duplicates | 1 |
7,497 | 6,919,225,546 | IssuesEvent | 2017-11-29 14:52:17 | nubisproject/nubis-nat | https://api.github.com/repos/nubisproject/nubis-nat | closed | Updating nubis-proxy puppet module | enhancement security | Once we have everything in nubisproject/nubis-nat#11 done we need to start using the upstream puppet module thias/puppet-squid3
| True | Updating nubis-proxy puppet module - Once we have everything in nubisproject/nubis-nat#11 done we need to start using the upstream puppet module thias/puppet-squid3
| non_defect | updating nubis proxy puppet module once we have everything in nubisproject nubis nat done we need to start using the upstream puppet module thias puppet | 0 |
180,208 | 30,467,694,296 | IssuesEvent | 2023-07-17 11:35:50 | jdi-testing/jdn-ai | https://api.github.com/repos/jdi-testing/jdn-ai | closed | [US-6-9] Visual updates of the filter icon | enhancement Design needed Filter Feature | - сейчас её плохо видно (в фигме сделал ярче до 85%)
- увеличить иконку до 16px, сейчас svg 14px
- неудобное расположение если плагин открыт в ширину, иконка отдельно остается справа
- добавить активное состояние | 1.0 | [US-6-9] Visual updates of the filter icon - - сейчас её плохо видно (в фигме сделал ярче до 85%)
- увеличить иконку до 16px, сейчас svg 14px
- неудобное расположение если плагин открыт в ширину, иконка отдельно остается справа
- добавить активное состояние | non_defect | visual updates of the filter icon сейчас её плохо видно в фигме сделал ярче до увеличить иконку до сейчас svg неудобное расположение если плагин открыт в ширину иконка отдельно остается справа добавить активное состояние | 0 |
28,219 | 5,221,382,505 | IssuesEvent | 2017-01-27 01:16:46 | elTiempoVuela/https-finder | https://api.github.com/repos/elTiempoVuela/https-finder | closed | Enhancement | auto-migrated Priority-Medium Type-Defect | ```
A little button for temporarely disable the fonction will be nice.
Thanks.
```
Original issue reported on code.google.com by `lcasinc...@gmail.com` on 1 Mar 2013 at 6:09
| 1.0 | Enhancement - ```
A little button for temporarely disable the fonction will be nice.
Thanks.
```
Original issue reported on code.google.com by `lcasinc...@gmail.com` on 1 Mar 2013 at 6:09
| defect | enhancement a little button for temporarely disable the fonction will be nice thanks original issue reported on code google com by lcasinc gmail com on mar at | 1 |
81,161 | 30,735,070,358 | IssuesEvent | 2023-07-28 06:55:35 | vector-im/element-x-android | https://api.github.com/repos/vector-im/element-x-android | opened | Able to send empty and non-trimmed text messages | T-Defect | ### Steps to reproduce
This may be a business decision but common sense is to not allow sending empty messages and trim messages before sending.
### Outcome

### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
0.1.2-nightly
### Homeserver
_No response_
### Will you send logs?
No
### Are you willing to provide a PR?
No | 1.0 | Able to send empty and non-trimmed text messages - ### Steps to reproduce
This may be a business decision but common sense is to not allow sending empty messages and trim messages before sending.
### Outcome

### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
0.1.2-nightly
### Homeserver
_No response_
### Will you send logs?
No
### Are you willing to provide a PR?
No | defect | able to send empty and non trimmed text messages steps to reproduce this may be a business decision but common sense is to not allow sending empty messages and trim messages before sending outcome your phone model no response operating system version no response application version and app store nightly homeserver no response will you send logs no are you willing to provide a pr no | 1 |
15,180 | 2,850,246,461 | IssuesEvent | 2015-05-31 12:05:16 | damonkohler/sl4a | https://api.github.com/repos/damonkohler/sl4a | opened | smsSend replaces "¡" (inverted exclamation mark) with "@" | auto-migrated Priority-Medium Type-Defect | _From @GoogleCodeExporter on May 31, 2015 11:31_
```
What device(s) are you experiencing the problem on?
HTC One S
What firmware version are you running on the device?
Android 4.1.1
What steps will reproduce the problem?
droid.smsSend(myNumber, "¡")
droid.smsSend(myNumber, "Any string containing ¡")
What is the expected output? What do you see instead?
I expect to get SMS containing "¡", instead I get new SMS with string
containing "@" where "¡" is supposed to be.
What version of the product are you using? On what operating system?
Using sl4a r6 with Python3 for Android
Tested only with Python 3.2.2
```
Original issue reported on code.google.com by `petrmach...@gmail.com` on 21 Jun 2013 at 11:03
_Copied from original issue: damonkohler/android-scripting#689_ | 1.0 | smsSend replaces "¡" (inverted exclamation mark) with "@" - _From @GoogleCodeExporter on May 31, 2015 11:31_
```
What device(s) are you experiencing the problem on?
HTC One S
What firmware version are you running on the device?
Android 4.1.1
What steps will reproduce the problem?
droid.smsSend(myNumber, "¡")
droid.smsSend(myNumber, "Any string containing ¡")
What is the expected output? What do you see instead?
I expect to get SMS containing "¡", instead I get new SMS with string
containing "@" where "¡" is supposed to be.
What version of the product are you using? On what operating system?
Using sl4a r6 with Python3 for Android
Tested only with Python 3.2.2
```
Original issue reported on code.google.com by `petrmach...@gmail.com` on 21 Jun 2013 at 11:03
_Copied from original issue: damonkohler/android-scripting#689_ | defect | smssend replaces ¡ inverted exclamation mark with from googlecodeexporter on may what device s are you experiencing the problem on htc one s what firmware version are you running on the device android what steps will reproduce the problem droid smssend mynumber ¡ droid smssend mynumber any string containing ¡ what is the expected output what do you see instead i expect to get sms containing ¡ instead i get new sms with string containing where ¡ is supposed to be what version of the product are you using on what operating system using with for android tested only with python original issue reported on code google com by petrmach gmail com on jun at copied from original issue damonkohler android scripting | 1 |
20,962 | 3,441,578,754 | IssuesEvent | 2015-12-14 19:02:51 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | Bug with Timers and the System Clock | area-vm Type-Defect | If I set a timer like this:
```dart
import "dart:async";
main() async {
var i = 1;
new Timer.periodic(const Duration(seconds: 1), (_) {
print("[${new DateTime.now().millisecondsSinceEpoch}] Hello #${i}");
i++;
});
}
```
and then spring my system clock forward say 2 or 3 hours, the timer will execute approximately the amount of times it would have executed if I kept it running from the time right before I set the clock to the time I set the clock to.
If I set my clock backwards, the timer will seem to stop.
Setting Clock Forwards: http://cl.ly/3I0a1W15040i
Setting Clock Backwards: http://cl.ly/1Y251F0W2P2P
This is a pretty critical issue, as system time changes can happen for many reasons, one of which includes Out of Box experiences, where the system time can change as a result of a user setting their timezone on a device.
Reproduced with Dart v1.12.2 on ARM
| 1.0 | Bug with Timers and the System Clock - If I set a timer like this:
```dart
import "dart:async";
main() async {
var i = 1;
new Timer.periodic(const Duration(seconds: 1), (_) {
print("[${new DateTime.now().millisecondsSinceEpoch}] Hello #${i}");
i++;
});
}
```
and then spring my system clock forward say 2 or 3 hours, the timer will execute approximately the amount of times it would have executed if I kept it running from the time right before I set the clock to the time I set the clock to.
If I set my clock backwards, the timer will seem to stop.
Setting Clock Forwards: http://cl.ly/3I0a1W15040i
Setting Clock Backwards: http://cl.ly/1Y251F0W2P2P
This is a pretty critical issue, as system time changes can happen for many reasons, one of which includes Out of Box experiences, where the system time can change as a result of a user setting their timezone on a device.
Reproduced with Dart v1.12.2 on ARM
| defect | bug with timers and the system clock if i set a timer like this dart import dart async main async var i new timer periodic const duration seconds print hello i i and then spring my system clock forward say or hours the timer will execute approximately the amount of times it would have executed if i kept it running from the time right before i set the clock to the time i set the clock to if i set my clock backwards the timer will seem to stop setting clock forwards setting clock backwards this is a pretty critical issue as system time changes can happen for many reasons one of which includes out of box experiences where the system time can change as a result of a user setting their timezone on a device reproduced with dart on arm | 1 |
5,386 | 2,610,186,660 | IssuesEvent | 2015-02-26 18:59:13 | chrsmith/quchuseban | https://api.github.com/repos/chrsmith/quchuseban | opened | 介绍色斑怎么才能去掉 | auto-migrated Priority-Medium Type-Defect | ```
《摘要》
30年前,一个年轻人离开故乡,开始创造自己的前途。他动身
的第一站,是去拜访本族的族长,请求指点。老族长正在练��
�,他听说本族有位后辈开始踏上人生的旅途,就写了3个字:
不要怕。然后抬起头来,望着年轻人说:“孩子,人生的秘��
�只有6个字,今天先告诉你3个,供你半生受用。”30年后,这
个从前的年轻人已是人到中年,有了一些成就,也添了很多��
�心事。归程漫漫,到了家乡,他又去拜访那位族长。他到了�
��长家里,才知道老人家几年前已经去世,家人取出一个密封
的信封对他说:“这是族长生前留给你的,他说有一天你会��
�来。”还乡的游子这才想起来,30年前他在这里听到人生的��
�半秘诀,拆开信封,里面赫然又是3个大字:不要悔。色斑怎
么才能去掉,
《客户案例》
相信很多姐妹都会有黄褐斑的烦恼吧,下面给大家说说��
�的祛斑经历吧。<br>
我的皮肤一直比较干,结婚前也没什么大的问题,结婚��
�,有了孩子,再加上工作上的事情一多,每天感觉都很累,�
��上还要做家务照顾孩子,慢慢的皮肤就变坏了,后来脸上居
然开始长了一块一块黄褐色的斑,我这才傻眼了。<br>
自从长斑后,我就变的敏感起来,出门逛街总是觉得别��
�盯着我看,看中的衣服穿在自己身上总觉得配着这张脸怎么�
��怎么别扭,到商场的化妆柜区,售货员马上就冲上来直奔主
题,上来就说我的斑怎样怎样,用什么可以祛斑,都快烦死��
�。<br>
自己也想祛斑,用过祛斑霜,还想去做激光祛斑,可想��
�他的副作用又打了退堂鼓,身边的朋友也是东一个办法西一�
��偏方的。用的最郁闷的方法就是,说是把一个鸡蛋泡在醋里
,等鸡蛋壳一个月后融化在醋里,再兑水喝,现在想想,这��
�办法真不知道是真的有效还是有人在恶搞,反正是郁闷的很�
��<br>
后来还是我最好的朋友,帮我四处打听,那天她到我家��
�给我带了两个个大包装盒,她叮嘱我每天按时使用,有什么�
��题打这个电话问人家,又给我写了个电话号码,我打开一看
是「黛芙薇尔精华液」,看产品介绍是祛斑的,原来那个电��
�使他们的专家热线,当时感觉挺无奈的,这丫头也太单纯了�
��,这也信,但这毕竟是她的心意,就试着用了,没想到用了
两个多月,脸上的斑就真的淡化了,等我把这个好消息告诉��
�的时候,这小丫头还挺得意,这下我可算了结了一块心病。
阅读了色斑怎么才能去掉,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
色斑怎么才能去掉,同时为您分享祛斑小方法
1、将带根的香菜洗净,加水煎煮,用菜汤洗脸,坚持使用可�
��令面部的色斑逐渐消除。
2、桃花、杏花各10克以水浸泡,过滤浸汁,用于洗脸。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:58 | 1.0 | 介绍色斑怎么才能去掉 - ```
《摘要》
30年前,一个年轻人离开故乡,开始创造自己的前途。他动身
的第一站,是去拜访本族的族长,请求指点。老族长正在练��
�,他听说本族有位后辈开始踏上人生的旅途,就写了3个字:
不要怕。然后抬起头来,望着年轻人说:“孩子,人生的秘��
�只有6个字,今天先告诉你3个,供你半生受用。”30年后,这
个从前的年轻人已是人到中年,有了一些成就,也添了很多��
�心事。归程漫漫,到了家乡,他又去拜访那位族长。他到了�
��长家里,才知道老人家几年前已经去世,家人取出一个密封
的信封对他说:“这是族长生前留给你的,他说有一天你会��
�来。”还乡的游子这才想起来,30年前他在这里听到人生的��
�半秘诀,拆开信封,里面赫然又是3个大字:不要悔。色斑怎
么才能去掉,
《客户案例》
相信很多姐妹都会有黄褐斑的烦恼吧,下面给大家说说��
�的祛斑经历吧。<br>
我的皮肤一直比较干,结婚前也没什么大的问题,结婚��
�,有了孩子,再加上工作上的事情一多,每天感觉都很累,�
��上还要做家务照顾孩子,慢慢的皮肤就变坏了,后来脸上居
然开始长了一块一块黄褐色的斑,我这才傻眼了。<br>
自从长斑后,我就变的敏感起来,出门逛街总是觉得别��
�盯着我看,看中的衣服穿在自己身上总觉得配着这张脸怎么�
��怎么别扭,到商场的化妆柜区,售货员马上就冲上来直奔主
题,上来就说我的斑怎样怎样,用什么可以祛斑,都快烦死��
�。<br>
自己也想祛斑,用过祛斑霜,还想去做激光祛斑,可想��
�他的副作用又打了退堂鼓,身边的朋友也是东一个办法西一�
��偏方的。用的最郁闷的方法就是,说是把一个鸡蛋泡在醋里
,等鸡蛋壳一个月后融化在醋里,再兑水喝,现在想想,这��
�办法真不知道是真的有效还是有人在恶搞,反正是郁闷的很�
��<br>
后来还是我最好的朋友,帮我四处打听,那天她到我家��
�给我带了两个个大包装盒,她叮嘱我每天按时使用,有什么�
��题打这个电话问人家,又给我写了个电话号码,我打开一看
是「黛芙薇尔精华液」,看产品介绍是祛斑的,原来那个电��
�使他们的专家热线,当时感觉挺无奈的,这丫头也太单纯了�
��,这也信,但这毕竟是她的心意,就试着用了,没想到用了
两个多月,脸上的斑就真的淡化了,等我把这个好消息告诉��
�的时候,这小丫头还挺得意,这下我可算了结了一块心病。
阅读了色斑怎么才能去掉,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
色斑怎么才能去掉,同时为您分享祛斑小方法
1、将带根的香菜洗净,加水煎煮,用菜汤洗脸,坚持使用可�
��令面部的色斑逐渐消除。
2、桃花、杏花各10克以水浸泡,过滤浸汁,用于洗脸。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:58 | defect | 介绍色斑怎么才能去掉 《摘要》 ,一个年轻人离开故乡,开始创造自己的前途。他动身 的第一站,是去拜访本族的族长,请求指点。老族长正在练�� �,他听说本族有位后辈开始踏上人生的旅途, : 不要怕。然后抬起头来,望着年轻人说:“孩子,人生的秘�� � , ,供你半生受用。” ,这 个从前的年轻人已是人到中年,有了一些成就,也添了很多�� �心事。归程漫漫,到了家乡,他又去拜访那位族长。他到了� ��长家里,才知道老人家几年前已经去世,家人取出一个密封 的信封对他说:“这是族长生前留给你的,他说有一天你会�� �来。”还乡的游子这才想起来, �� �半秘诀,拆开信封, :不要悔。色斑怎 么才能去掉, 《客户案例》 相信很多姐妹都会有黄褐斑的烦恼吧,下面给大家说说�� �的祛斑经历吧。 我的皮肤一直比较干,结婚前也没什么大的问题,结婚�� �,有了孩子,再加上工作上的事情一多,每天感觉都很累,� ��上还要做家务照顾孩子,慢慢的皮肤就变坏了,后来脸上居 然开始长了一块一块黄褐色的斑,我这才傻眼了。 自从长斑后,我就变的敏感起来,出门逛街总是觉得别�� �盯着我看,看中的衣服穿在自己身上总觉得配着这张脸怎么� ��怎么别扭,到商场的化妆柜区,售货员马上就冲上来直奔主 题,上来就说我的斑怎样怎样,用什么可以祛斑,都快烦死�� �。 自己也想祛斑,用过祛斑霜,还想去做激光祛斑,可想�� �他的副作用又打了退堂鼓,身边的朋友也是东一个办法西一� ��偏方的。用的最郁闷的方法就是,说是把一个鸡蛋泡在醋里 ,等鸡蛋壳一个月后融化在醋里,再兑水喝,现在想想,这�� �办法真不知道是真的有效还是有人在恶搞,反正是郁闷的很� �� 后来还是我最好的朋友,帮我四处打听,那天她到我家�� �给我带了两个个大包装盒,她叮嘱我每天按时使用,有什么� ��题打这个电话问人家,又给我写了个电话号码,我打开一看 是「黛芙薇尔精华液」,看产品介绍是祛斑的,原来那个电�� �使他们的专家热线,当时感觉挺无奈的,这丫头也太单纯了� ��,这也信,但这毕竟是她的心意,就试着用了,没想到用了 两个多月,脸上的斑就真的淡化了,等我把这个好消息告诉�� �的时候,这小丫头还挺得意,这下我可算了结了一块心病。 阅读了色斑怎么才能去掉,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 色斑怎么才能去掉,同时为您分享祛斑小方法 、将带根的香菜洗净,加水煎煮,用菜汤洗脸,坚持使用可� ��令面部的色斑逐渐消除。 、桃花、 ,过滤浸汁,用于洗脸。 original issue reported on code google com by additive gmail com on jul at | 1 |
72,093 | 3,371,857,591 | IssuesEvent | 2015-11-23 20:57:02 | INN/Largo | https://api.github.com/repos/INN/Largo | closed | default widget styles are still terrible | priority: high type: improvement | in looking at http://raleighpublicrecord-org.largoproject.staging.wpengine.com/ a couple of things
- [ ] we need some padding in widgets to open up the layout a bit. It was previous 20px or so and got removed in 0.4 but we need to add it back.
- [ ] the gray bottom border on widgets is probably not necessary and it causes issues in the homepage bottom widget area (when that layout is selected), probably just remove it and make sure the default bottom margin on widgets is consistent and generous enough to make them feel distinct
- [ ] widget title styles. something. just anything.
throwing this to @kaeti first for input but we need to look at this ASAP. | 1.0 | default widget styles are still terrible - in looking at http://raleighpublicrecord-org.largoproject.staging.wpengine.com/ a couple of things
- [ ] we need some padding in widgets to open up the layout a bit. It was previous 20px or so and got removed in 0.4 but we need to add it back.
- [ ] the gray bottom border on widgets is probably not necessary and it causes issues in the homepage bottom widget area (when that layout is selected), probably just remove it and make sure the default bottom margin on widgets is consistent and generous enough to make them feel distinct
- [ ] widget title styles. something. just anything.
throwing this to @kaeti first for input but we need to look at this ASAP. | non_defect | default widget styles are still terrible in looking at a couple of things we need some padding in widgets to open up the layout a bit it was previous or so and got removed in but we need to add it back the gray bottom border on widgets is probably not necessary and it causes issues in the homepage bottom widget area when that layout is selected probably just remove it and make sure the default bottom margin on widgets is consistent and generous enough to make them feel distinct widget title styles something just anything throwing this to kaeti first for input but we need to look at this asap | 0 |
53,407 | 6,314,509,773 | IssuesEvent | 2017-07-24 11:03:02 | dwyl/best-evidence | https://api.github.com/repos/dwyl/best-evidence | closed | Reduce space around 'Start Searching' button to avoid need to scroll | enhancement please-test T25m | **As a** user
**I want to** avoid scrolling to view page content
**So that** I can access the information I need effortlessly and quickly.
Currently the 'Start Searching' button on the login page for existing users is not fully visible on an iPhone 5. To remove the need for scrolling for users, we can reduce the spacing around the button.
Here's how it will now look on iPhone 5:
<img src="https://user-images.githubusercontent.com/16775804/28466198-8f486a34-6e23-11e7-9715-c2224909668c.png" width=250px />
| 1.0 | Reduce space around 'Start Searching' button to avoid need to scroll - **As a** user
**I want to** avoid scrolling to view page content
**So that** I can access the information I need effortlessly and quickly.
Currently the 'Start Searching' button on the login page for existing users is not fully visible on an iPhone 5. To remove the need for scrolling for users, we can reduce the spacing around the button.
Here's how it will now look on iPhone 5:
<img src="https://user-images.githubusercontent.com/16775804/28466198-8f486a34-6e23-11e7-9715-c2224909668c.png" width=250px />
| non_defect | reduce space around start searching button to avoid need to scroll as a user i want to avoid scrolling to view page content so that i can access the information i need effortlessly and quickly currently the start searching button on the login page for existing users is not fully visible on an iphone to remove the need for scrolling for users we can reduce the spacing around the button here s how it will now look on iphone | 0 |
17,481 | 10,708,268,136 | IssuesEvent | 2019-10-24 19:18:34 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Regional VNET Integration General Availability | Pri1 app-service/svc cxp product-question triaged | Dear, Any ETA available on the GA date for this feature?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a7a98803-1438-b1b5-f543-7dd88bc4294e
* Version Independent ID: 37ff1d0f-ed8e-5e4d-1f4c-1b9f6cffb938
* Content: [Integrate app with Azure Virtual Network - Azure App Service](https://docs.microsoft.com/en-us/azure/app-service/web-sites-integrate-with-vnet?fireglass_rsn=true#feedback)
* Content Source: [articles/app-service/web-sites-integrate-with-vnet.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/web-sites-integrate-with-vnet.md)
* Service: **app-service**
* GitHub Login: @ccompy
* Microsoft Alias: **ccompy** | 1.0 | Regional VNET Integration General Availability - Dear, Any ETA available on the GA date for this feature?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a7a98803-1438-b1b5-f543-7dd88bc4294e
* Version Independent ID: 37ff1d0f-ed8e-5e4d-1f4c-1b9f6cffb938
* Content: [Integrate app with Azure Virtual Network - Azure App Service](https://docs.microsoft.com/en-us/azure/app-service/web-sites-integrate-with-vnet?fireglass_rsn=true#feedback)
* Content Source: [articles/app-service/web-sites-integrate-with-vnet.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/web-sites-integrate-with-vnet.md)
* Service: **app-service**
* GitHub Login: @ccompy
* Microsoft Alias: **ccompy** | non_defect | regional vnet integration general availability dear any eta available on the ga date for this feature document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service github login ccompy microsoft alias ccompy | 0 |
12,261 | 2,685,576,771 | IssuesEvent | 2015-03-30 02:58:13 | IssueMigrationTest/Test5 | https://api.github.com/repos/IssueMigrationTest/Test5 | closed | Error printout for module TKinter | auto-migrated Priority-Medium Type-Defect | **Issue by gunnar.f...@gmail.com**
_7 Jul 2013 at 1:09 GMT_
_Originally opened on Google Code_
----
```
running command
shedskin libMain.py
That is I am trying to convert my python script libMain.py to C++
------------------------------------
Recieving the following Error printout:
*** SHED SKIN Python-to-C++ Compiler 0.9.1 ***
Copyright 2005-2011 Mark Dufour; License GNU GPL version 3 (See LICENSE)
*ERROR* libMain.py:11: cannot locate module: Tkinter
-------------------
My code looks like this
#!/usr/bin/python
# -*- coding: utf-8 -*-
title = 'Pmw.advancedmenuBar demonstration'
# Import Pmw from this directory tree.
import sys
sys.path[:0] = ['../../..']
import Tkinter <-------------THIS is where i fails
import Pmw
import libSearch
#from FileHandler import *
import fileHandler
etc
------------------------------
Is this because the tool can't handle Tkinter or hav I done something wrong???
OS is UBUNTU 12.04.
I used the following command to install the tool
sudo apt-get install shedskin
I hope I get the latiest version by using it
Gunnar
```
| 1.0 | Error printout for module TKinter - **Issue by gunnar.f...@gmail.com**
_7 Jul 2013 at 1:09 GMT_
_Originally opened on Google Code_
----
```
running command
shedskin libMain.py
That is I am trying to convert my python script libMain.py to C++
------------------------------------
Recieving the following Error printout:
*** SHED SKIN Python-to-C++ Compiler 0.9.1 ***
Copyright 2005-2011 Mark Dufour; License GNU GPL version 3 (See LICENSE)
*ERROR* libMain.py:11: cannot locate module: Tkinter
-------------------
My code looks like this
#!/usr/bin/python
# -*- coding: utf-8 -*-
title = 'Pmw.advancedmenuBar demonstration'
# Import Pmw from this directory tree.
import sys
sys.path[:0] = ['../../..']
import Tkinter <-------------THIS is where i fails
import Pmw
import libSearch
#from FileHandler import *
import fileHandler
etc
------------------------------
Is this because the tool can't handle Tkinter or hav I done something wrong???
OS is UBUNTU 12.04.
I used the following command to install the tool
sudo apt-get install shedskin
I hope I get the latiest version by using it
Gunnar
```
| defect | error printout for module tkinter issue by gunnar f gmail com jul at gmt originally opened on google code running command shedskin libmain py that is i am trying to convert my python script libmain py to c recieving the following error printout shed skin python to c compiler copyright mark dufour license gnu gpl version see license error libmain py cannot locate module tkinter my code looks like this usr bin python coding utf title pmw advancedmenubar demonstration import pmw from this directory tree import sys sys path import tkinter this is where i fails import pmw import libsearch from filehandler import import filehandler etc is this because the tool can t handle tkinter or hav i done something wrong os is ubuntu i used the following command to install the tool sudo apt get install shedskin i hope i get the latiest version by using it gunnar | 1 |
19,939 | 3,281,874,678 | IssuesEvent | 2015-10-28 01:03:19 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | Fixture Indexes' name problem. | console Defect On hold | Hello,
We have stumbled upon the following warning message while running our test cases on our Cakephp 3.1.2 project:
```
Warning Error: Fixture creation for "threads" failed "SQLSTATE[HY000]: General error: 1 index user_id already exists" in [/var/www/project/vendor/cakephp/cakephp/src/TestSuite/Fixture/TestFixture.php, line 240]
```
Our tests are currently using two tables that have `belongsTo` associations with our users table, thus when we bake our fixtures, both tables fixtures define the index name for the user_id fields as `user_id`:
```php
'_indexes' =>
[
'user_id' => ['type' => 'index', 'columns' => ['user_id'], 'length' => []],
],
```
After some research, we have found that tables' indexes names should be unique, so we believe this is the reason of the warning.
We are wondering if we might be making a mistake in our fixture baking?
```
bin/cake bake fixture Threads -n 2
```
or if it perhaps this is a normal behavior of the fixture baking process, and that by default the names of the index for fixtures is the name of the field itself?
I thank you in advance for your response; should you need more information, please let me know.
Thanks
| 1.0 | Fixture Indexes' name problem. - Hello,
We have stumbled upon the following warning message while running our test cases on our Cakephp 3.1.2 project:
```
Warning Error: Fixture creation for "threads" failed "SQLSTATE[HY000]: General error: 1 index user_id already exists" in [/var/www/project/vendor/cakephp/cakephp/src/TestSuite/Fixture/TestFixture.php, line 240]
```
Our tests are currently using two tables that have `belongsTo` associations with our users table, thus when we bake our fixtures, both tables fixtures define the index name for the user_id fields as `user_id`:
```php
'_indexes' =>
[
'user_id' => ['type' => 'index', 'columns' => ['user_id'], 'length' => []],
],
```
After some research, we have found that tables' indexes names should be unique, so we believe this is the reason of the warning.
We are wondering if we might be making a mistake in our fixture baking?
```
bin/cake bake fixture Threads -n 2
```
or if it perhaps this is a normal behavior of the fixture baking process, and that by default the names of the index for fixtures is the name of the field itself?
I thank you in advance for your response; should you need more information, please let me know.
Thanks
| defect | fixture indexes name problem hello we have stumbled upon the following warning message while running our test cases on our cakephp project warning error fixture creation for threads failed sqlstate general error index user id already exists in our tests are currently using two tables that have belongsto associations with our users table thus when we bake our fixtures both tables fixtures define the index name for the user id fields as user id php indexes user id length after some research we have found that tables indexes names should be unique so we believe this is the reason of the warning we are wondering if we might be making a mistake in our fixture baking bin cake bake fixture threads n or if it perhaps this is a normal behavior of the fixture baking process and that by default the names of the index for fixtures is the name of the field itself i thank you in advance for your response should you need more information please let me know thanks | 1 |
255,774 | 8,126,361,687 | IssuesEvent | 2018-08-17 01:42:38 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | VisIt 2.7.3 no longer runs on the RZ | Bug Likelihood: 3 - Occasional Priority: Normal Severity: 3 - Major Irritation Support Group: DOE/ASC | VisIt 2.7.3 no longer runs on the RZ. There is a library that it was linked with that no longer exists. Here is the e-mail exchange with the person from LC on it. We should rebuild 2.7.3 to fix this problem since there are still people using old clients out there.
Hi Eric,
It looks like you own the VisIT builds in /usr/gapps/visit on LC machines. Let me know if this should be going somewhere else though. I took a call from a user yesterday who has an older version of VisIT (2.7.3) on his desktop. He was trying to start a compute engine on rzzeus from VisIT on his desktop, but VisIT would hang when he tried to start it. There was an error on the console:
/usr/gapps/visit/2.7.3/linux-x86_64/bin/viewer: error while loading shared libraries: libMagick++.so.2: cannot open shared object file: No such file or directory
I can reproduce the behavior by running /usr/gapps/visit/bin/legacylauncher -v 2.7.3 on rzzeus. I asked the user to move to a newer version of VisIT on his desktop, but I thought I'd also check in with you about whether the older versions of VisIT could be relinked against more recent versions of libMagick++.
Thanks,
Ryan Day
â
Livermore Computing Hotline
Hours: M-F:8A-12P,1-4:45P
Email: lc-hotline@llnl.gov
Phone: 925-422-4531
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2485
Status: Rejected
Project: VisIt
Tracker: Bug
Priority: Normal
Subject: VisIt 2.7.3 no longer runs on the RZ
Assigned to:
Category:
Target version:
Author: Eric Brugger
Start: 12/17/2015
Due date:
% Done: 0
Estimated time:
Created: 12/17/2015 07:10 pm
Updated: 01/12/2016 06:57 pm
Likelihood: 3 - Occasional
Severity: 3 - Major Irritation
Found in version: 2.7.3
Impact:
Expected Use:
OS: Linux
Support Group: DOE/ASC
Description:
VisIt 2.7.3 no longer runs on the RZ. There is a library that it was linked with that no longer exists. Here is the e-mail exchange with the person from LC on it. We should rebuild 2.7.3 to fix this problem since there are still people using old clients out there.
Hi Eric,
It looks like you own the VisIT builds in /usr/gapps/visit on LC machines. Let me know if this should be going somewhere else though. I took a call from a user yesterday who has an older version of VisIT (2.7.3) on his desktop. He was trying to start a compute engine on rzzeus from VisIT on his desktop, but VisIT would hang when he tried to start it. There was an error on the console:
/usr/gapps/visit/2.7.3/linux-x86_64/bin/viewer: error while loading shared libraries: libMagick++.so.2: cannot open shared object file: No such file or directory
I can reproduce the behavior by running /usr/gapps/visit/bin/legacylauncher -v 2.7.3 on rzzeus. I asked the user to move to a newer version of VisIT on his desktop, but I thought I'd also check in with you about whether the older versions of VisIT could be relinked against more recent versions of libMagick++.
Thanks,
Ryan Day
â
Livermore Computing Hotline
Hours: M-F:8A-12P,1-4:45P
Email: lc-hotline@llnl.gov
Phone: 925-422-4531
Comments:
We will have the user upgrade to the latest version of VisIt.
| 1.0 | VisIt 2.7.3 no longer runs on the RZ - VisIt 2.7.3 no longer runs on the RZ. There is a library that it was linked with that no longer exists. Here is the e-mail exchange with the person from LC on it. We should rebuild 2.7.3 to fix this problem since there are still people using old clients out there.
Hi Eric,
It looks like you own the VisIT builds in /usr/gapps/visit on LC machines. Let me know if this should be going somewhere else though. I took a call from a user yesterday who has an older version of VisIT (2.7.3) on his desktop. He was trying to start a compute engine on rzzeus from VisIT on his desktop, but VisIT would hang when he tried to start it. There was an error on the console:
/usr/gapps/visit/2.7.3/linux-x86_64/bin/viewer: error while loading shared libraries: libMagick++.so.2: cannot open shared object file: No such file or directory
I can reproduce the behavior by running /usr/gapps/visit/bin/legacylauncher -v 2.7.3 on rzzeus. I asked the user to move to a newer version of VisIT on his desktop, but I thought I'd also check in with you about whether the older versions of VisIT could be relinked against more recent versions of libMagick++.
Thanks,
Ryan Day
â
Livermore Computing Hotline
Hours: M-F:8A-12P,1-4:45P
Email: lc-hotline@llnl.gov
Phone: 925-422-4531
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2485
Status: Rejected
Project: VisIt
Tracker: Bug
Priority: Normal
Subject: VisIt 2.7.3 no longer runs on the RZ
Assigned to:
Category:
Target version:
Author: Eric Brugger
Start: 12/17/2015
Due date:
% Done: 0
Estimated time:
Created: 12/17/2015 07:10 pm
Updated: 01/12/2016 06:57 pm
Likelihood: 3 - Occasional
Severity: 3 - Major Irritation
Found in version: 2.7.3
Impact:
Expected Use:
OS: Linux
Support Group: DOE/ASC
Description:
VisIt 2.7.3 no longer runs on the RZ. There is a library that it was linked with that no longer exists. Here is the e-mail exchange with the person from LC on it. We should rebuild 2.7.3 to fix this problem since there are still people using old clients out there.
Hi Eric,
It looks like you own the VisIT builds in /usr/gapps/visit on LC machines. Let me know if this should be going somewhere else though. I took a call from a user yesterday who has an older version of VisIT (2.7.3) on his desktop. He was trying to start a compute engine on rzzeus from VisIT on his desktop, but VisIT would hang when he tried to start it. There was an error on the console:
/usr/gapps/visit/2.7.3/linux-x86_64/bin/viewer: error while loading shared libraries: libMagick++.so.2: cannot open shared object file: No such file or directory
I can reproduce the behavior by running /usr/gapps/visit/bin/legacylauncher -v 2.7.3 on rzzeus. I asked the user to move to a newer version of VisIT on his desktop, but I thought I'd also check in with you about whether the older versions of VisIT could be relinked against more recent versions of libMagick++.
Thanks,
Ryan Day
â
Livermore Computing Hotline
Hours: M-F:8A-12P,1-4:45P
Email: lc-hotline@llnl.gov
Phone: 925-422-4531
Comments:
We will have the user upgrade to the latest version of VisIt.
| non_defect | visit no longer runs on the rz visit no longer runs on the rz there is a library that it was linked with that no longer exists here is the e mail exchange with the person from lc on it we should rebuild to fix this problem since there are still people using old clients out there hi eric it looks like you own the visit builds in usr gapps visit on lc machines let me know if this should be going somewhere else though i took a call from a user yesterday who has an older version of visit on his desktop he was trying to start a compute engine on rzzeus from visit on his desktop but visit would hang when he tried to start it there was an error on the console usr gapps visit linux bin viewer error while loading shared libraries libmagick so cannot open shared object file no such file or directory i can reproduce the behavior by running usr gapps visit bin legacylauncher v on rzzeus i asked the user to move to a newer version of visit on his desktop but i thought i d also check in with you about whether the older versions of visit could be relinked against more recent versions of libmagick thanks ryan day â livermore computing hotline hours m f email lc hotline llnl gov phone redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status rejected project visit tracker bug priority normal subject visit no longer runs on the rz assigned to category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity major irritation found in version impact expected use os linux support group doe asc description visit no longer runs on the rz there is a library that it was linked with that no longer exists here is the e mail exchange with the person from lc on it we should rebuild to fix this problem since there are still people using old clients out there hi eric it looks like you own the visit builds in usr gapps visit on lc machines let me know if this should be going somewhere else though i took a call from a user yesterday who has an older version of visit on his desktop he was trying to start a compute engine on rzzeus from visit on his desktop but visit would hang when he tried to start it there was an error on the console usr gapps visit linux bin viewer error while loading shared libraries libmagick so cannot open shared object file no such file or directory i can reproduce the behavior by running usr gapps visit bin legacylauncher v on rzzeus i asked the user to move to a newer version of visit on his desktop but i thought i d also check in with you about whether the older versions of visit could be relinked against more recent versions of libmagick thanks ryan day â livermore computing hotline hours m f email lc hotline llnl gov phone comments we will have the user upgrade to the latest version of visit | 0 |
22,974 | 3,735,368,577 | IssuesEvent | 2016-03-08 11:46:23 | janjarfalk/jquery-elastic | https://api.github.com/repos/janjarfalk/jquery-elastic | closed | Textareas don't auto-expand if you don't press the ENTER key | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Enable the elastic plugin on a textarea
2. Type a long paragraph in the textarea that exceeds the viewable area
(but with no carriage returns, ie. never press ENTER)
What is the expected output? What do you see instead?
I would expect that the textarea would auto-grow. However, it doesn't. It
appears that the textarea only auto-expands if you press ENTER after each
line of text.
What version of the product are you using? On what operating system?
v1.6.1, Windows XP SP3 - tested on Firefox 3.5.2 and IE 8
Please provide any additional information below.
```
Original issue reported on code.google.com by `michael....@gmail.com` on 4 Sep 2009 at 7:31 | 1.0 | Textareas don't auto-expand if you don't press the ENTER key - ```
What steps will reproduce the problem?
1. Enable the elastic plugin on a textarea
2. Type a long paragraph in the textarea that exceeds the viewable area
(but with no carriage returns, ie. never press ENTER)
What is the expected output? What do you see instead?
I would expect that the textarea would auto-grow. However, it doesn't. It
appears that the textarea only auto-expands if you press ENTER after each
line of text.
What version of the product are you using? On what operating system?
v1.6.1, Windows XP SP3 - tested on Firefox 3.5.2 and IE 8
Please provide any additional information below.
```
Original issue reported on code.google.com by `michael....@gmail.com` on 4 Sep 2009 at 7:31 | defect | textareas don t auto expand if you don t press the enter key what steps will reproduce the problem enable the elastic plugin on a textarea type a long paragraph in the textarea that exceeds the viewable area but with no carriage returns ie never press enter what is the expected output what do you see instead i would expect that the textarea would auto grow however it doesn t it appears that the textarea only auto expands if you press enter after each line of text what version of the product are you using on what operating system windows xp tested on firefox and ie please provide any additional information below original issue reported on code google com by michael gmail com on sep at | 1 |
340,918 | 30,553,651,795 | IssuesEvent | 2023-07-20 10:10:26 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Failing test: Chrome UI Functional Tests.test/functional/apps/management/_files·ts - management Files management should render an empty prompt | failed-test Team:SharedUX | A test failed on a tracked branch
```
Error: expected 'Files\nManage files stored in Kibana.\nStatistics\nName A-Z\nTags\nName\nSize\nNo files matched your search.' to contain 'No files found'
at Assertion.assert (expect.js:100:11)
at Assertion.contain (expect.js:442:10)
at Context.<anonymous> (_files.ts:26:27)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
at Object.apply (wrap_function.js:73:16)
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/31779#0188dea0-2746-4ecc-a8a4-c2516646d39f)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome UI Functional Tests.test/functional/apps/management/_files·ts","test.name":"management Files management should render an empty prompt","test.failCount":5}} --> | 1.0 | Failing test: Chrome UI Functional Tests.test/functional/apps/management/_files·ts - management Files management should render an empty prompt - A test failed on a tracked branch
```
Error: expected 'Files\nManage files stored in Kibana.\nStatistics\nName A-Z\nTags\nName\nSize\nNo files matched your search.' to contain 'No files found'
at Assertion.assert (expect.js:100:11)
at Assertion.contain (expect.js:442:10)
at Context.<anonymous> (_files.ts:26:27)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
at Object.apply (wrap_function.js:73:16)
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/31779#0188dea0-2746-4ecc-a8a4-c2516646d39f)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome UI Functional Tests.test/functional/apps/management/_files·ts","test.name":"management Files management should render an empty prompt","test.failCount":5}} --> | non_defect | failing test chrome ui functional tests test functional apps management files·ts management files management should render an empty prompt a test failed on a tracked branch error expected files nmanage files stored in kibana nstatistics nname a z ntags nname nsize nno files matched your search to contain no files found at assertion assert expect js at assertion contain expect js at context files ts at runmicrotasks at processticksandrejections node internal process task queues at object apply wrap function js first failure | 0 |
8,339 | 6,520,547,088 | IssuesEvent | 2017-08-28 16:56:58 | mysticfall/Alensia | https://api.github.com/repos/mysticfall/Alensia | closed | Fix delay before opening main menu | performance | Currently, there's a noticeable delay before the main menu shows up for the first time. We need to profile the process to see if there's potential bottleneck and remove the overhead. | True | Fix delay before opening main menu - Currently, there's a noticeable delay before the main menu shows up for the first time. We need to profile the process to see if there's potential bottleneck and remove the overhead. | non_defect | fix delay before opening main menu currently there s a noticeable delay before the main menu shows up for the first time we need to profile the process to see if there s potential bottleneck and remove the overhead | 0 |
34,028 | 7,327,854,641 | IssuesEvent | 2018-03-04 15:01:43 | cython/cython | https://api.github.com/repos/cython/cython | closed | segfault with asyncio | defect | On macOS 10.12.4, Cython 0.25.2, Python 3.6.1, if you await asyncio.sleep() <= Cython <= Python chain, it will segfault. It won't die if you await Cython future right away, without first using Python.
Here is an example that reproduces this bug:
main.py:
```
WANT_CRASH = True
import asyncio
import cy_test
async def main():
await cy_test.say()
loop = asyncio.get_event_loop()
if WANT_CRASH:
loop.run_until_complete(main())
else:
loop.run_until_complete(cy_test.say())
loop.close()
```
cy_test.pyx:
```
import asyncio
from py_test import py_async
async def cy_async():
print("- this one is from Cython")
async def say():
await cb()
async def cb(): # renaming this to "say" would prevent crash
print("Async functions can be awaited without problem:")
await cy_async()
await py_async()
print("And this will actually sleep for 3 seconds before crashing")
await asyncio.sleep(3)
print("I will be never output, because app has crashed already")
```
py_test.py:
```
async def py_async():
print("- and this one is from Python")
```
Setup.py:
```
from distutils.core import setup
from distutils.extension import Extension
from Cython.Build import cythonize
setup(
name='asyncio_cb',
ext_modules=cythonize([
Extension("cy_test", ["cy_test.pyx"]),
]),
)
```
Makefile:
```
all:
python3 Setup.py build_ext --inplace
test: all
python3 main.py
clean:
@echo Cleaning
@rm -f cy_*.c *.o *.so *~ core
@rm -rf build
```
| 1.0 | segfault with asyncio - On macOS 10.12.4, Cython 0.25.2, Python 3.6.1, if you await asyncio.sleep() <= Cython <= Python chain, it will segfault. It won't die if you await Cython future right away, without first using Python.
Here is an example that reproduces this bug:
main.py:
```
WANT_CRASH = True
import asyncio
import cy_test
async def main():
await cy_test.say()
loop = asyncio.get_event_loop()
if WANT_CRASH:
loop.run_until_complete(main())
else:
loop.run_until_complete(cy_test.say())
loop.close()
```
cy_test.pyx:
```
import asyncio
from py_test import py_async
async def cy_async():
print("- this one is from Cython")
async def say():
await cb()
async def cb(): # renaming this to "say" would prevent crash
print("Async functions can be awaited without problem:")
await cy_async()
await py_async()
print("And this will actually sleep for 3 seconds before crashing")
await asyncio.sleep(3)
print("I will be never output, because app has crashed already")
```
py_test.py:
```
async def py_async():
print("- and this one is from Python")
```
Setup.py:
```
from distutils.core import setup
from distutils.extension import Extension
from Cython.Build import cythonize
setup(
name='asyncio_cb',
ext_modules=cythonize([
Extension("cy_test", ["cy_test.pyx"]),
]),
)
```
Makefile:
```
all:
python3 Setup.py build_ext --inplace
test: all
python3 main.py
clean:
@echo Cleaning
@rm -f cy_*.c *.o *.so *~ core
@rm -rf build
```
| defect | segfault with asyncio on macos cython python if you await asyncio sleep cython python chain it will segfault it won t die if you await cython future right away without first using python here is an example that reproduces this bug main py want crash true import asyncio import cy test async def main await cy test say loop asyncio get event loop if want crash loop run until complete main else loop run until complete cy test say loop close cy test pyx import asyncio from py test import py async async def cy async print this one is from cython async def say await cb async def cb renaming this to say would prevent crash print async functions can be awaited without problem await cy async await py async print and this will actually sleep for seconds before crashing await asyncio sleep print i will be never output because app has crashed already py test py async def py async print and this one is from python setup py from distutils core import setup from distutils extension import extension from cython build import cythonize setup name asyncio cb ext modules cythonize extension cy test makefile all setup py build ext inplace test all main py clean echo cleaning rm f cy c o so core rm rf build | 1 |
5,795 | 2,794,353,781 | IssuesEvent | 2015-05-11 16:13:55 | OryxProject/oryx | https://api.github.com/repos/OryxProject/oryx | closed | Explain / fix non-fatal exceptions during integration tests | enhancement Tests | There are several types of exceptions that appear regularly in the logs when integration tests run. They do not appear to affect the result, and tests pass. They may just be corner cases, from ZK and Kafka being rapidly started and stopped. This is a to-do to see if we can ever figure them out.
Updated to include what's being observed most recently. It's almost all:
```
java.io.IOException: Broken pipe
at sun.nio.ch.FileDispatcherImpl.write0(Native Method)
at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:47)
at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:93)
at sun.nio.ch.IOUtil.write(IOUtil.java:65)
at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:470)
at kafka.api.TopicDataSend.writeTo(FetchResponse.scala:123)
at kafka.network.MultiSend.writeTo(Transmission.scala:101)
at kafka.api.FetchResponseSend.writeTo(FetchResponse.scala:231)
at kafka.network.Processor.write(SocketServer.scala:472)
at kafka.network.Processor.run(SocketServer.scala:342)
at java.lang.Thread.run(Thread.java:745)
```
Rarely:
```
java.nio.channels.ClosedByInterruptException
at java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
at sun.nio.ch.SocketChannelImpl.poll(SocketChannelImpl.java:956)
at sun.nio.ch.SocketAdaptor$SocketInputStream.read(SocketAdaptor.java:204)
at sun.nio.ch.ChannelInputStream.read(ChannelInputStream.java:103)
at java.nio.channels.Channels$ReadableByteChannelImpl.read(Channels.java:385)
at kafka.utils.Utils$.read(Utils.scala:380)
at kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.scala:54)
at kafka.network.Receive$class.readCompletely(Transmission.scala:56)
at kafka.network.BoundedByteBufferReceive.readCompletely(BoundedByteBufferReceive.scala:29)
at kafka.network.BlockingChannel.receive(BlockingChannel.scala:111)
at kafka.producer.SyncProducer.liftedTree1$1(SyncProducer.scala:75)
at kafka.producer.SyncProducer.kafka$producer$SyncProducer$$doSend(SyncProducer.scala:72)
at kafka.producer.SyncProducer.send(SyncProducer.scala:113)
at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:58)
at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:93)
at kafka.consumer.ConsumerFetcherManager$LeaderFinderThread.doWork(ConsumerFetcherManager.scala:66)
at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:60)
```
and a few in the web tests:
```
java.lang.IllegalStateException: ServiceLocatorImpl(__HK2_Generated_73,74,527137253) has been shut down
at org.glassfish.grizzly.servlet.FilterChainImpl.doFilter(FilterChainImpl.java:151)
at org.glassfish.grizzly.servlet.FilterChainImpl.invokeFilterChain(FilterChainImpl.java:106)
at org.glassfish.grizzly.servlet.ServletHandler.doServletService(ServletHandler.java:221)
at org.glassfish.grizzly.servlet.ServletHandler.service(ServletHandler.java:169)
at org.glassfish.grizzly.http.server.HttpHandler$1.run(HttpHandler.java:219)
at org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.doWork(AbstractThreadPool.java:565)
at org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.run(AbstractThreadPool.java:545)
at java.lang.Thread.run(Thread.java:745)
```
Others seem to have gone. | 1.0 | Explain / fix non-fatal exceptions during integration tests - There are several types of exceptions that appear regularly in the logs when integration tests run. They do not appear to affect the result, and tests pass. They may just be corner cases, from ZK and Kafka being rapidly started and stopped. This is a to-do to see if we can ever figure them out.
Updated to include what's being observed most recently. It's almost all:
```
java.io.IOException: Broken pipe
at sun.nio.ch.FileDispatcherImpl.write0(Native Method)
at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:47)
at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:93)
at sun.nio.ch.IOUtil.write(IOUtil.java:65)
at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:470)
at kafka.api.TopicDataSend.writeTo(FetchResponse.scala:123)
at kafka.network.MultiSend.writeTo(Transmission.scala:101)
at kafka.api.FetchResponseSend.writeTo(FetchResponse.scala:231)
at kafka.network.Processor.write(SocketServer.scala:472)
at kafka.network.Processor.run(SocketServer.scala:342)
at java.lang.Thread.run(Thread.java:745)
```
Rarely:
```
java.nio.channels.ClosedByInterruptException
at java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
at sun.nio.ch.SocketChannelImpl.poll(SocketChannelImpl.java:956)
at sun.nio.ch.SocketAdaptor$SocketInputStream.read(SocketAdaptor.java:204)
at sun.nio.ch.ChannelInputStream.read(ChannelInputStream.java:103)
at java.nio.channels.Channels$ReadableByteChannelImpl.read(Channels.java:385)
at kafka.utils.Utils$.read(Utils.scala:380)
at kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.scala:54)
at kafka.network.Receive$class.readCompletely(Transmission.scala:56)
at kafka.network.BoundedByteBufferReceive.readCompletely(BoundedByteBufferReceive.scala:29)
at kafka.network.BlockingChannel.receive(BlockingChannel.scala:111)
at kafka.producer.SyncProducer.liftedTree1$1(SyncProducer.scala:75)
at kafka.producer.SyncProducer.kafka$producer$SyncProducer$$doSend(SyncProducer.scala:72)
at kafka.producer.SyncProducer.send(SyncProducer.scala:113)
at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:58)
at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:93)
at kafka.consumer.ConsumerFetcherManager$LeaderFinderThread.doWork(ConsumerFetcherManager.scala:66)
at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:60)
```
and a few in the web tests:
```
java.lang.IllegalStateException: ServiceLocatorImpl(__HK2_Generated_73,74,527137253) has been shut down
at org.glassfish.grizzly.servlet.FilterChainImpl.doFilter(FilterChainImpl.java:151)
at org.glassfish.grizzly.servlet.FilterChainImpl.invokeFilterChain(FilterChainImpl.java:106)
at org.glassfish.grizzly.servlet.ServletHandler.doServletService(ServletHandler.java:221)
at org.glassfish.grizzly.servlet.ServletHandler.service(ServletHandler.java:169)
at org.glassfish.grizzly.http.server.HttpHandler$1.run(HttpHandler.java:219)
at org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.doWork(AbstractThreadPool.java:565)
at org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.run(AbstractThreadPool.java:545)
at java.lang.Thread.run(Thread.java:745)
```
Others seem to have gone. | non_defect | explain fix non fatal exceptions during integration tests there are several types of exceptions that appear regularly in the logs when integration tests run they do not appear to affect the result and tests pass they may just be corner cases from zk and kafka being rapidly started and stopped this is a to do to see if we can ever figure them out updated to include what s being observed most recently it s almost all java io ioexception broken pipe at sun nio ch filedispatcherimpl native method at sun nio ch socketdispatcher write socketdispatcher java at sun nio ch ioutil writefromnativebuffer ioutil java at sun nio ch ioutil write ioutil java at sun nio ch socketchannelimpl write socketchannelimpl java at kafka api topicdatasend writeto fetchresponse scala at kafka network multisend writeto transmission scala at kafka api fetchresponsesend writeto fetchresponse scala at kafka network processor write socketserver scala at kafka network processor run socketserver scala at java lang thread run thread java rarely java nio channels closedbyinterruptexception at java nio channels spi abstractinterruptiblechannel end abstractinterruptiblechannel java at sun nio ch socketchannelimpl poll socketchannelimpl java at sun nio ch socketadaptor socketinputstream read socketadaptor java at sun nio ch channelinputstream read channelinputstream java at java nio channels channels readablebytechannelimpl read channels java at kafka utils utils read utils scala at kafka network boundedbytebufferreceive readfrom boundedbytebufferreceive scala at kafka network receive class readcompletely transmission scala at kafka network boundedbytebufferreceive readcompletely boundedbytebufferreceive scala at kafka network blockingchannel receive blockingchannel scala at kafka producer syncproducer syncproducer scala at kafka producer syncproducer kafka producer syncproducer dosend syncproducer scala at kafka producer syncproducer send syncproducer scala at kafka client clientutils fetchtopicmetadata clientutils scala at kafka client clientutils fetchtopicmetadata clientutils scala at kafka consumer consumerfetchermanager leaderfinderthread dowork consumerfetchermanager scala at kafka utils shutdownablethread run shutdownablethread scala and a few in the web tests java lang illegalstateexception servicelocatorimpl generated has been shut down at org glassfish grizzly servlet filterchainimpl dofilter filterchainimpl java at org glassfish grizzly servlet filterchainimpl invokefilterchain filterchainimpl java at org glassfish grizzly servlet servlethandler doservletservice servlethandler java at org glassfish grizzly servlet servlethandler service servlethandler java at org glassfish grizzly http server httphandler run httphandler java at org glassfish grizzly threadpool abstractthreadpool worker dowork abstractthreadpool java at org glassfish grizzly threadpool abstractthreadpool worker run abstractthreadpool java at java lang thread run thread java others seem to have gone | 0 |
16,229 | 10,453,582,181 | IssuesEvent | 2019-09-19 16:56:19 | HackIllinois/api | https://api.github.com/repos/HackIllinois/api | opened | Brainstorm & implement Slack integrations | Feature New Service Notifications | This is an idea that never made it into the API, but considering how much we use Slack for the event, it makes sense to build in some integration.
An obvious use case would be to send out notifications alongside mobile pushes.
https://github.com/acm-uiuc/groot-notification-service/ might be a good place to look.
This could either be built as a standalone service or as an additional target for sending notifications. If we go with the latter, this is a good opportunity to abstract out SNS as one of several notification delivery systems we can use. | 1.0 | Brainstorm & implement Slack integrations - This is an idea that never made it into the API, but considering how much we use Slack for the event, it makes sense to build in some integration.
An obvious use case would be to send out notifications alongside mobile pushes.
https://github.com/acm-uiuc/groot-notification-service/ might be a good place to look.
This could either be built as a standalone service or as an additional target for sending notifications. If we go with the latter, this is a good opportunity to abstract out SNS as one of several notification delivery systems we can use. | non_defect | brainstorm implement slack integrations this is an idea that never made it into the api but considering how much we use slack for the event it makes sense to build in some integration an obvious use case would be to send out notifications alongside mobile pushes might be a good place to look this could either be built as a standalone service or as an additional target for sending notifications if we go with the latter this is a good opportunity to abstract out sns as one of several notification delivery systems we can use | 0 |
12,355 | 2,693,621,476 | IssuesEvent | 2015-04-01 15:40:39 | akvo/akvo-flow | https://api.github.com/repos/akvo/akvo-flow | opened | Even when it fails, processor always sends a fileProcessed message to the dashboard | 1 - Defect | ref #1185
Calls to http://uat1.akvoflow.org/processor?action=submit&fileName=non-existent-file.zip always seem to return a successful message in the dashboard - regardless of their actual status:

| 1.0 | Even when it fails, processor always sends a fileProcessed message to the dashboard - ref #1185
Calls to http://uat1.akvoflow.org/processor?action=submit&fileName=non-existent-file.zip always seem to return a successful message in the dashboard - regardless of their actual status:

| defect | even when it fails processor always sends a fileprocessed message to the dashboard ref calls to always seem to return a successful message in the dashboard regardless of their actual status | 1 |
67,823 | 21,180,414,699 | IssuesEvent | 2022-04-08 07:20:39 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Vulnerability in Postgresql used by hazelcast-jet-cdc-postgres [HZ-1050] | Type: Defect Source: Internal dependencies Module: Jet security severity:critical to-jira Team: Platform | `hazelcast-jet-cdc-postgres` uses `debezium-connector-postgres` in version `1.2.5.Final` and it uses `postgresql` in version `42.2.14` with following CVE:
- CVE-2022-21724 - https://nvd.nist.gov/vuln/detail/CVE-2022-21724 | 1.0 | Vulnerability in Postgresql used by hazelcast-jet-cdc-postgres [HZ-1050] - `hazelcast-jet-cdc-postgres` uses `debezium-connector-postgres` in version `1.2.5.Final` and it uses `postgresql` in version `42.2.14` with following CVE:
- CVE-2022-21724 - https://nvd.nist.gov/vuln/detail/CVE-2022-21724 | defect | vulnerability in postgresql used by hazelcast jet cdc postgres hazelcast jet cdc postgres uses debezium connector postgres in version final and it uses postgresql in version with following cve cve | 1 |
53,747 | 13,262,226,309 | IssuesEvent | 2020-08-20 21:20:50 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | SnDAQ 2ndbin bug in rates reported to moni (Trac #1999) | Migrated from Trac defect other | There is a bug in how SnDAQ reports rates to moni, causing apparent spikes in event rate shortly after run transitions (the "2ndbin" bug).
This is not critical for run start but conditionals have been written in moni to suppress alerts caused by the bug.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1999">https://code.icecube.wisc.edu/projects/icecube/ticket/1999</a>, reported by sybenzviand owned by gmoment</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:44",
"_ts": "1550067284370534",
"description": "There is a bug in how SnDAQ reports rates to moni, causing apparent spikes in event rate shortly after run transitions (the \"2ndbin\" bug).\n\nThis is not critical for run start but conditionals have been written in moni to suppress alerts caused by the bug.",
"reporter": "sybenzvi",
"cc": "gmoment, rcross, mfrere",
"resolution": "fixed",
"time": "2017-05-05T20:29:05",
"component": "other",
"summary": "SnDAQ 2ndbin bug in rates reported to moni",
"priority": "major",
"keywords": "sndaq, moni, 2ndbin",
"milestone": "",
"owner": "gmoment",
"type": "defect"
}
```
</p>
</details>
| 1.0 | SnDAQ 2ndbin bug in rates reported to moni (Trac #1999) - There is a bug in how SnDAQ reports rates to moni, causing apparent spikes in event rate shortly after run transitions (the "2ndbin" bug).
This is not critical for run start but conditionals have been written in moni to suppress alerts caused by the bug.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1999">https://code.icecube.wisc.edu/projects/icecube/ticket/1999</a>, reported by sybenzviand owned by gmoment</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:44",
"_ts": "1550067284370534",
"description": "There is a bug in how SnDAQ reports rates to moni, causing apparent spikes in event rate shortly after run transitions (the \"2ndbin\" bug).\n\nThis is not critical for run start but conditionals have been written in moni to suppress alerts caused by the bug.",
"reporter": "sybenzvi",
"cc": "gmoment, rcross, mfrere",
"resolution": "fixed",
"time": "2017-05-05T20:29:05",
"component": "other",
"summary": "SnDAQ 2ndbin bug in rates reported to moni",
"priority": "major",
"keywords": "sndaq, moni, 2ndbin",
"milestone": "",
"owner": "gmoment",
"type": "defect"
}
```
</p>
</details>
| defect | sndaq bug in rates reported to moni trac there is a bug in how sndaq reports rates to moni causing apparent spikes in event rate shortly after run transitions the bug this is not critical for run start but conditionals have been written in moni to suppress alerts caused by the bug migrated from json status closed changetime ts description there is a bug in how sndaq reports rates to moni causing apparent spikes in event rate shortly after run transitions the bug n nthis is not critical for run start but conditionals have been written in moni to suppress alerts caused by the bug reporter sybenzvi cc gmoment rcross mfrere resolution fixed time component other summary sndaq bug in rates reported to moni priority major keywords sndaq moni milestone owner gmoment type defect | 1 |
56,307 | 15,020,012,251 | IssuesEvent | 2021-02-01 14:14:09 | mozilla-lockwise/lockwise-ios | https://api.github.com/repos/mozilla-lockwise/lockwise-ios | reopened | Welcome screen is briefly displayed after locking the app while syncing | archived defect priority-P2 | Build: 1.6.0(3339)
Device: iPhone X
iOS: 11.4
Steps to reproduce:
1. Launch Lockwise
2. While syncing, go to settings and lock the app
Actual results:
- The welcome screen will be briefly displayed.
Video:
- [ScreenRecording_05-22-2019 16-14-35.MP4.zip](https://github.com/mozilla-lockwise/lockwise-ios/files/3207510/ScreenRecording_05-22-2019.16-14-35.MP4.zip)
| 1.0 | Welcome screen is briefly displayed after locking the app while syncing - Build: 1.6.0(3339)
Device: iPhone X
iOS: 11.4
Steps to reproduce:
1. Launch Lockwise
2. While syncing, go to settings and lock the app
Actual results:
- The welcome screen will be briefly displayed.
Video:
- [ScreenRecording_05-22-2019 16-14-35.MP4.zip](https://github.com/mozilla-lockwise/lockwise-ios/files/3207510/ScreenRecording_05-22-2019.16-14-35.MP4.zip)
| defect | welcome screen is briefly displayed after locking the app while syncing build device iphone x ios steps to reproduce launch lockwise while syncing go to settings and lock the app actual results the welcome screen will be briefly displayed video | 1 |
64,103 | 18,213,964,270 | IssuesEvent | 2021-09-30 00:13:24 | idaholab/moose | https://api.github.com/repos/idaholab/moose | closed | Unknown material property in ParsedMaterial is not an error | C: MOOSE T: defect P: normal | ## Bug Description
When using ParsedMaterial, if the material property listed in material_property_names is not defined, it becomes defined with a value of zero. I think it would make more sense for this to be an error. As it is now, input files will merrily run if someone mistypes a material property name.
## Steps to Reproduce
Use ParsedMaterial and supply a bogus material property in material_property_names.
## Impact
This behavior allows input file errors that are difficult to detect. | 1.0 | Unknown material property in ParsedMaterial is not an error - ## Bug Description
When using ParsedMaterial, if the material property listed in material_property_names is not defined, it becomes defined with a value of zero. I think it would make more sense for this to be an error. As it is now, input files will merrily run if someone mistypes a material property name.
## Steps to Reproduce
Use ParsedMaterial and supply a bogus material property in material_property_names.
## Impact
This behavior allows input file errors that are difficult to detect. | defect | unknown material property in parsedmaterial is not an error bug description when using parsedmaterial if the material property listed in material property names is not defined it becomes defined with a value of zero i think it would make more sense for this to be an error as it is now input files will merrily run if someone mistypes a material property name steps to reproduce use parsedmaterial and supply a bogus material property in material property names impact this behavior allows input file errors that are difficult to detect | 1 |
8,844 | 2,612,907,146 | IssuesEvent | 2015-02-27 17:26:15 | chrsmith/windows-package-manager | https://api.github.com/repos/chrsmith/windows-package-manager | closed | Gimp 2.8.2 link | auto-migrated Milestone-End_Of_Month Type-Defect | ```
http://downloads.sourceforge.net/project/gimp-win/GIMP %2B GTK%2B %28stable
release%29/GIMP 2.8.2/gimp-2.8.2-setup.exe
doesn't exist anymore.
```
Original issue reported on code.google.com by `igi...@gmail.com` on 28 Aug 2012 at 5:51 | 1.0 | Gimp 2.8.2 link - ```
http://downloads.sourceforge.net/project/gimp-win/GIMP %2B GTK%2B %28stable
release%29/GIMP 2.8.2/gimp-2.8.2-setup.exe
doesn't exist anymore.
```
Original issue reported on code.google.com by `igi...@gmail.com` on 28 Aug 2012 at 5:51 | defect | gimp link gtk release gimp gimp setup exe doesn t exist anymore original issue reported on code google com by igi gmail com on aug at | 1 |
35,150 | 7,611,656,279 | IssuesEvent | 2018-05-01 14:45:04 | gwaldron/osgearth | https://api.github.com/repos/gwaldron/osgearth | closed | Disabled layers initialize upon exit | defect | I put an mgrs_graticule layer in an earth file with enabled="false". Upon quitting the viewer, the layer initializes. | 1.0 | Disabled layers initialize upon exit - I put an mgrs_graticule layer in an earth file with enabled="false". Upon quitting the viewer, the layer initializes. | defect | disabled layers initialize upon exit i put an mgrs graticule layer in an earth file with enabled false upon quitting the viewer the layer initializes | 1 |
73,409 | 24,617,002,990 | IssuesEvent | 2022-10-15 12:52:50 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Can't connect session to key backup | T-Defect | ### Steps to reproduce
- Go to Safety & Privacy settings
- Encryption→Secure Backup says: "This session is not backing up your keys, but you do have an existing backup you can restore from and add to going forward. Connect this session to key backup before signing out to avoid losing any keys that may only be on this session."
- Click "Connect this session to Key Backup"
- It restores the keys from the backup. (Why? I'd expect to actually backup my keys _to_ the backup, rather than restoring from it.)
- After the process is finished, nothing has changed. Encryption→Secure Backup still says: "This session is not backing up your keys, but you do have an existing backup you can restore from and add to going forward. Connect this session to key backup before signing out to avoid losing any keys that may only be on this session."
### Outcome
#### What did you expect?
A message like "This session is backing up your keys"
#### What happened instead?
"This session is not backing up your keys"
### Operating system
Ubuntu 22.04.1
### Application version
1.11.10
### How did you install the app?
flatpak install flathub im.riot.Riot
### Homeserver
Synapse 1.68.0
### Will you send logs?
No | 1.0 | Can't connect session to key backup - ### Steps to reproduce
- Go to Safety & Privacy settings
- Encryption→Secure Backup says: "This session is not backing up your keys, but you do have an existing backup you can restore from and add to going forward. Connect this session to key backup before signing out to avoid losing any keys that may only be on this session."
- Click "Connect this session to Key Backup"
- It restores the keys from the backup. (Why? I'd expect to actually backup my keys _to_ the backup, rather than restoring from it.)
- After the process is finished, nothing has changed. Encryption→Secure Backup still says: "This session is not backing up your keys, but you do have an existing backup you can restore from and add to going forward. Connect this session to key backup before signing out to avoid losing any keys that may only be on this session."
### Outcome
#### What did you expect?
A message like "This session is backing up your keys"
#### What happened instead?
"This session is not backing up your keys"
### Operating system
Ubuntu 22.04.1
### Application version
1.11.10
### How did you install the app?
flatpak install flathub im.riot.Riot
### Homeserver
Synapse 1.68.0
### Will you send logs?
No | defect | can t connect session to key backup steps to reproduce go to safety privacy settings encryption→secure backup says this session is not backing up your keys but you do have an existing backup you can restore from and add to going forward connect this session to key backup before signing out to avoid losing any keys that may only be on this session click connect this session to key backup it restores the keys from the backup why i d expect to actually backup my keys to the backup rather than restoring from it after the process is finished nothing has changed encryption→secure backup still says this session is not backing up your keys but you do have an existing backup you can restore from and add to going forward connect this session to key backup before signing out to avoid losing any keys that may only be on this session outcome what did you expect a message like this session is backing up your keys what happened instead this session is not backing up your keys operating system ubuntu application version how did you install the app flatpak install flathub im riot riot homeserver synapse will you send logs no | 1 |
279,594 | 21,176,613,699 | IssuesEvent | 2022-04-08 01:03:09 | opendp/opendp | https://api.github.com/repos/opendp/opendp | closed | Document feasible range for epsilon in `base_gaussian` | CATEGORY: Documentation good first issue OpenDP Core Effort 1 - Small :coffee: | Any `epsilon` greater than 1 is clipped to 1, because of limitations of the mechanism. This should be more clearly documented, because it is easy to mistake the relation as broken, if it continually doesn't pass for arbitrarily large epsilon. | 1.0 | Document feasible range for epsilon in `base_gaussian` - Any `epsilon` greater than 1 is clipped to 1, because of limitations of the mechanism. This should be more clearly documented, because it is easy to mistake the relation as broken, if it continually doesn't pass for arbitrarily large epsilon. | non_defect | document feasible range for epsilon in base gaussian any epsilon greater than is clipped to because of limitations of the mechanism this should be more clearly documented because it is easy to mistake the relation as broken if it continually doesn t pass for arbitrarily large epsilon | 0 |
41,950 | 10,833,308,569 | IssuesEvent | 2019-11-11 12:38:11 | opencv/opencv | https://api.github.com/repos/opencv/opencv | closed | OpenCV linkage fails | category: build/install incomplete platform: ios/osx | <!--
If you have a question rather than reporting a bug please go to http://answers.opencv.org where you get much faster responses.
If you need further assistance please read [How To Contribute](https://github.com/opencv/opencv/wiki/How_to_contribute).
This is a template helping you to create an issue which can be processed as quickly as possible. This is the bug reporting section for the OpenCV library.
-->
##### System information (version)
<!-- Example
- OpenCV => 3.1
- Operating System / Platform => Windows 64 Bit
- Compiler => Visual Studio 2015
-->
- OpenCV => 4.1.2-dev
- Operating System / Platform => macOS 10.15.1
- Compiler => clang 11.0.0
##### Detailed description
OpenCV compiles successfully, but linkage fails with:
```
Undefined symbols for architecture x86_64:
"_ippicviSet_32f_C1MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicvsRound_64f", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviSet_32s_C3MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicvsConvert_64f32f", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviSet_32f_C4MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviSet_8u_C3MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviSet_16u_C4MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviSet_16u_C3MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviSet_32s_C1MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviCopy_8u_C3P3R", referenced from:
_llwiCopySplit in libippiw.a(iw_image_op_copy_split.c.o)
"_ippicviCopy_16u_C3P3R", referenced from:
_llwiCopySplit in libippiw.a(iw_image_op_copy_split.c.o)
"_ippicviCopy_16u_C4P4R", referenced from:
_llwiCopySplit in libippiw.a(iw_image_op_copy_split.c.o)
"_ippicviCopy_32f_C4P4R", referenced from:
_llwiCopySplit in libippiw.a(iw_image_op_copy_split.c.o)
"_ippicviCopy_16u_P3C3R", referenced from:
_llwiCopyMerge in libippiw.a(iw_image_op_copy_merge.c.o)
"_ippicviCopy_32f_P3C3R", referenced from:
_llwiCopyMerge in libippiw.a(iw_image_op_copy_merge.c.o)
"_ippicviCopy_8u_P4C4R", referenced from:
_llwiCopyMerge in libippiw.a(iw_image_op_copy_merge.c.o)
"_ippicviCopy_16u_P4C4R", referenced from:
_llwiCopyMerge in libippiw.a(iw_image_op_copy_merge.c.o)
"_ippicviCopy_32f_P4C4R", referenced from:
_llwiCopyMerge in libippiw.a(iw_image_op_copy_merge.c.o)
"_ippicviCopy_16u_C4MR", referenced from:
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviCopy_32f_C4MR", referenced from:
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviCopy_32f_C3MR", referenced from:
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviSet_32f_C3MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviCopy_8u_P3C3R", referenced from:
_llwiCopyMerge in libippiw.a(iw_image_op_copy_merge.c.o)
"_ippicviCopy_32f_C1C3R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_8u_C3C1R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_16u_C3C1R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_8u_C4C1R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_8u_C4CR", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_32f_C4CR", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviDCTInvGetSize_32f", referenced from:
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviMirror_8u_C1R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviMirror_8u_C1IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviMirror_32f_C1R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviCopy_16u_C1C4R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_16u_C4CR", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviMirror_16u_C1R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicvsMaxEvery_16u", referenced from:
cv::hal::max16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviMirror_32f_C1IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviMirror_16u_C1IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviDFTGetSize_R_32f", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::Dft_R_IPPLoop_Invoker<cv::IPPDFT_R_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviMirror_16u_C4R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviDFTInv_PackToR_32f_C1R", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
"_ippicviMirror_16u_C3R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviNormDiff_L1_16u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviMirror_32f_C3IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicvsFree", referenced from:
_iwiImage_Release in libippiw.a(iw_image.c.o)
"_ippicvInit", referenced from:
cv::ipp::IPPInitSingleton::IPPInitSingleton() in system.cpp.o
"_ippicviCopy_32f_C3P3R", referenced from:
_llwiCopySplit in libippiw.a(iw_image_op_copy_split.c.o)
"_ippicvGetEnabledCpuFeatures", referenced from:
cv::ipp::IPPInitSingleton::IPPInitSingleton() in system.cpp.o
"_ippicviCountInRange_8u_C1R", referenced from:
cv::countNonZero(cv::_InputArray const&) in count_non_zero.dispatch.cpp.o
"_ippicviNormDiff_Inf_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_Inf_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L1_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L1_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsExp_32f_A21", referenced from:
cv::hal::exp32f(float const*, float*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicvsLn_32f_A21", referenced from:
cv::hal::log32f(float const*, float*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicviNormRel_L2_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L1_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L1_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviXor_8u_C1R", referenced from:
cv::hal::xor8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNormDiff_L2_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L1_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_Inf_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_Inf_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviMirror_32f_C3R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviNormRel_Inf_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L2_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCopy_16u_C1C3R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviNorm_L1_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviMirror_32f_C4IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviNorm_L1_16u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviDCTFwdGetSize_32f", referenced from:
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviNormRel_L1_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCopy_32s_C1MR", referenced from:
cv::copyMask32s(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
"_ippicviNormRel_Inf_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviMirror_8u_C4IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviNormDiff_Inf_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L1_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsExp_64f_A50", referenced from:
cv::hal::exp64f(double const*, double*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicvsLn_64f_A50", referenced from:
cv::hal::log64f(double const*, double*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicviNormRel_L2_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_Inf_16u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_Inf_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L1_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviSet_32s_C4MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviNormRel_L2_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_Inf_32f_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L1_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_16u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCopy_8u_C4MR", referenced from:
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviNormDiff_Inf_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_8u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsMaxEvery_32f", referenced from:
cv::hal::max32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNorm_L1_32f_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L1_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviSub_16s_C1RSfs", referenced from:
cv::hal::sub16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsDFTGetSize_C_64fc", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicviNorm_L2_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviSub_8u_C1RSfs", referenced from:
cv::hal::sub8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNorm_L1_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L2_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L1_8u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L1_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_8u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L2_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviSet_16u_C1MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviNormDiff_L1_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviAdd_16u_C1RSfs", referenced from:
cv::hal::add16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsPolarToCart_64f", referenced from:
cv::polarToCart(cv::_InputArray const&, cv::_InputArray const&, cv::_OutputArray const&, cv::_OutputArray const&, bool) in mathfuncs.cpp.o
"_ippicviCopy_8u_C4P4R", referenced from:
_llwiCopySplit in libippiw.a(iw_image_op_copy_split.c.o)
"_ippicviCompare_16u_C1R", referenced from:
cv::hal::cmp16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNorm_Inf_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L2_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L2_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L2_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCopy_32s_C3MR", referenced from:
cv::copyMask32sC3(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
"_ippicvsDFTGetSize_R_32f", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicviNorm_Inf_32f_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L1_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_16u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L2_32f_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L2_8u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCopy_16u_C4C1R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviMul_8u_C1RSfs", referenced from:
cv::hal::mul8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviAbsDiff_32f_C1R", referenced from:
cv::hal::absdiff32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviDFTFwd_RToPack_32f_C1R", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
"_ippicviMirror_8u_C3R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicvsInvSqrt_32f_A21", referenced from:
cv::hal::invSqrt32f(float const*, float*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicviNormDiff_Inf_8u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviSet_8u_C1MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviCopy_8u_C1R_L", referenced from:
cv::Mat::copyTo(cv::_OutputArray const&) const in copy.cpp.o
"_ippicviMirror_8u_C3IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicvsInvSqrt_64f_A50", referenced from:
cv::hal::invSqrt64f(double const*, double*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicvsMagnitude_64f", referenced from:
cv::hal::magnitude64f(double const*, double const*, double*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicviNot_8u_C1R", referenced from:
cv::hal::not8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviGetLibVersion", referenced from:
cv::ipp::IPPInitSingleton::IPPInitSingleton() in system.cpp.o
"_ippicvsMagnitude_32f", referenced from:
cv::hal::magnitude32f(float const*, float const*, float*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicviDCTInvInit_32f", referenced from:
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicvMalloc_L", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::Dft_C_IPPLoop_Invoker<cv::IPPDFT_C_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
cv::Dft_R_IPPLoop_Invoker<cv::IPPDFT_R_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviCountInRange_32f_C1R", referenced from:
cv::countNonZero(cv::_InputArray const&) in count_non_zero.dispatch.cpp.o
"_ippicvsDFTGetSize_C_32fc", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicviDFTInv_CToC_32fc_C1R", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
"_ippicviSet_8u_C4MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviNorm_L2_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviDCTFwd_32f_C1R", referenced from:
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicvsDFTInit_C_64fc", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicviDFTInit_R_32f", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::Dft_R_IPPLoop_Invoker<cv::IPPDFT_R_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviNorm_L2_16u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L1_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviDFTInit_C_32fc", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::Dft_C_IPPLoop_Invoker<cv::IPPDFT_C_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviDFTFwd_CToC_32fc_C1R", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
"_ippicviMirror_8u_C4R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviNormRel_Inf_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviOr_8u_C1R", referenced from:
cv::hal::or8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsDFTInv_PackToR_64f", referenced from:
void cv::CCSIDFT<double>(cv::OcvDftOptions const&, double const*, double*) in dxt.cpp.o
"_ippicvsDFTFwd_RToPack_64f", referenced from:
void cv::RealDFT<double>(cv::OcvDftOptions const&, double const*, double*) in dxt.cpp.o
"_ippicviMirror_32f_C4R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviCopy_16u_C3CR", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_16u_C1MR", referenced from:
cv::copyMask16u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicvsDFTInv_CToC_64fc", referenced from:
void cv::DFT<double>(cv::OcvDftOptions const&, cv::Complex<double> const*, cv::Complex<double>*) in dxt.cpp.o
"_ippicvsDFTInv_PackToR_32f", referenced from:
void cv::CCSIDFT<float>(cv::OcvDftOptions const&, float const*, float*) in dxt.cpp.o
"_ippicviNormRel_Inf_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L1_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsDFTFwd_RToPack_32f", referenced from:
void cv::RealDFT<float>(cv::OcvDftOptions const&, float const*, float*) in dxt.cpp.o
"_ippicvsDFTInv_CToC_32fc", referenced from:
void cv::DFT<float>(cv::OcvDftOptions const&, cv::Complex<float> const*, cv::Complex<float>*) in dxt.cpp.o
"_ippicvsDFTFwd_CToC_32fc", referenced from:
void cv::DFT<float>(cv::OcvDftOptions const&, cv::Complex<float> const*, cv::Complex<float>*) in dxt.cpp.o
"_ippicviCompare_16s_C1R", referenced from:
cv::hal::cmp16s(short const*, unsigned long, short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsDFTInit_C_32fc", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicvFree", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::Dft_C_IPPLoop_Invoker<cv::IPPDFT_C_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
cv::Dft_R_IPPLoop_Invoker<cv::IPPDFT_R_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviNormDiff_L1_32f_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_Inf_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsDFTFwd_CToC_64fc", referenced from:
void cv::DFT<double>(cv::OcvDftOptions const&, cv::Complex<double> const*, cv::Complex<double>*) in dxt.cpp.o
"_ippicviNormRel_L2_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCopy_8u_C1C3R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviNorm_L2_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsMinEvery_64f", referenced from:
cv::hal::min64f(double const*, unsigned long, double const*, unsigned long, double*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviCopy_8u_C3MR", referenced from:
cv::copyMask8uC3(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviCopy_8u_C3CR", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_8u_C1MR", referenced from:
cv::copyMask8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviNormDiff_L2_32f_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviMul_32f_C1R", referenced from:
cv::hal::mul32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviMul_16u_C1RSfs", referenced from:
cv::hal::mul16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNormRel_Inf_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviDCTFwdInit_32f", referenced from:
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviNormDiff_L1_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCompare_32f_C1R", referenced from:
cv::hal::cmp32f(float const*, unsigned long, float const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNormDiff_L2_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L1_8u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviAbsDiff_8u_C1R", referenced from:
cv::hal::absdiff8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNorm_L1_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L2_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsPolarToCart_32f", referenced from:
cv::polarToCart(cv::_InputArray const&, cv::_InputArray const&, cv::_OutputArray const&, cv::_OutputArray const&, bool) in mathfuncs.cpp.o
"_ippicviDFTGetSize_C_32fc", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::Dft_C_IPPLoop_Invoker<cv::IPPDFT_C_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviCopy_32f_C1MR", referenced from:
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviCopy_32f_C3CR", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviMirror_16u_C4IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviNormDiff_L1_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsMaxEvery_8u", referenced from:
cv::hal::max8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsDFTGetSize_R_64f", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicviCompare_8u_C1R", referenced from:
cv::hal::cmp8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNormRel_L1_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviAdd_8u_C1RSfs", referenced from:
cv::hal::add8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviCopy_32f_C1C4R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviNormRel_Inf_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsMinEvery_32f", referenced from:
cv::hal::min32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsMinEvery_8u", referenced from:
cv::hal::min8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviAbsDiff_16u_C1R", referenced from:
cv::hal::absdiff16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsMaxEvery_64f", referenced from:
cv::hal::max64f(double const*, unsigned long, double const*, unsigned long, double*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsDFTInit_R_64f", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicvsMinEvery_16u", referenced from:
cv::hal::min16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviAdd_32f_C1R", referenced from:
cv::hal::add32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviMul_16s_C1RSfs", referenced from:
cv::hal::mul16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviDCTInv_32f_C1R", referenced from:
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviSub_32f_C1R", referenced from:
cv::hal::sub32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvSetCpuFeatures", referenced from:
cv::ipp::IPPInitSingleton::IPPInitSingleton() in system.cpp.o
"_ippicviAdd_16s_C1RSfs", referenced from:
cv::hal::add16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviCopy_8u_C1C4R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviSub_16u_C1RSfs", referenced from:
cv::hal::sub16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviCopy_32s_C4MR", referenced from:
cv::copyMask32sC4(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
"_ippicvsDFTInit_R_32f", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicviCopy_32f_C4C1R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicvGetCpuFeatures", referenced from:
cv::ipp::IPPInitSingleton::IPPInitSingleton() in system.cpp.o
"_ippicviCopy_16u_C3MR", referenced from:
cv::copyMask16uC3(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviCopy_32f_C3C1R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviAnd_8u_C1R", referenced from:
cv::hal::and8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviMirror_16u_C3IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
ld: symbol(s) not found for architecture x86_64
```
##### Steps to reproduce
- clone opencv repo master at commit 5dd3e6052ed9e614cf4303eee64d092fefbabd8e
- clone opencv_contrib master at commit 61cbb44539c5a8b07315c9a6c38f5350c0e91
- cd opencv/build
- mkdir tmp
- cmake -DOPENCV_EXTRA_MODULES_PATH=<path to>/opencv_contrib/modules ..
- make -j2
| 1.0 | OpenCV linkage fails - <!--
If you have a question rather than reporting a bug please go to http://answers.opencv.org where you get much faster responses.
If you need further assistance please read [How To Contribute](https://github.com/opencv/opencv/wiki/How_to_contribute).
This is a template helping you to create an issue which can be processed as quickly as possible. This is the bug reporting section for the OpenCV library.
-->
##### System information (version)
<!-- Example
- OpenCV => 3.1
- Operating System / Platform => Windows 64 Bit
- Compiler => Visual Studio 2015
-->
- OpenCV => 4.1.2-dev
- Operating System / Platform => macOS 10.15.1
- Compiler => clang 11.0.0
##### Detailed description
OpenCV compiles successfully, but linkage fails with:
```
Undefined symbols for architecture x86_64:
"_ippicviSet_32f_C1MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicvsRound_64f", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviSet_32s_C3MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicvsConvert_64f32f", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviSet_32f_C4MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviSet_8u_C3MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviSet_16u_C4MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviSet_16u_C3MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviSet_32s_C1MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviCopy_8u_C3P3R", referenced from:
_llwiCopySplit in libippiw.a(iw_image_op_copy_split.c.o)
"_ippicviCopy_16u_C3P3R", referenced from:
_llwiCopySplit in libippiw.a(iw_image_op_copy_split.c.o)
"_ippicviCopy_16u_C4P4R", referenced from:
_llwiCopySplit in libippiw.a(iw_image_op_copy_split.c.o)
"_ippicviCopy_32f_C4P4R", referenced from:
_llwiCopySplit in libippiw.a(iw_image_op_copy_split.c.o)
"_ippicviCopy_16u_P3C3R", referenced from:
_llwiCopyMerge in libippiw.a(iw_image_op_copy_merge.c.o)
"_ippicviCopy_32f_P3C3R", referenced from:
_llwiCopyMerge in libippiw.a(iw_image_op_copy_merge.c.o)
"_ippicviCopy_8u_P4C4R", referenced from:
_llwiCopyMerge in libippiw.a(iw_image_op_copy_merge.c.o)
"_ippicviCopy_16u_P4C4R", referenced from:
_llwiCopyMerge in libippiw.a(iw_image_op_copy_merge.c.o)
"_ippicviCopy_32f_P4C4R", referenced from:
_llwiCopyMerge in libippiw.a(iw_image_op_copy_merge.c.o)
"_ippicviCopy_16u_C4MR", referenced from:
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviCopy_32f_C4MR", referenced from:
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviCopy_32f_C3MR", referenced from:
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviSet_32f_C3MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviCopy_8u_P3C3R", referenced from:
_llwiCopyMerge in libippiw.a(iw_image_op_copy_merge.c.o)
"_ippicviCopy_32f_C1C3R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_8u_C3C1R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_16u_C3C1R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_8u_C4C1R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_8u_C4CR", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_32f_C4CR", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviDCTInvGetSize_32f", referenced from:
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviMirror_8u_C1R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviMirror_8u_C1IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviMirror_32f_C1R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviCopy_16u_C1C4R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_16u_C4CR", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviMirror_16u_C1R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicvsMaxEvery_16u", referenced from:
cv::hal::max16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviMirror_32f_C1IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviMirror_16u_C1IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviDFTGetSize_R_32f", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::Dft_R_IPPLoop_Invoker<cv::IPPDFT_R_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviMirror_16u_C4R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviDFTInv_PackToR_32f_C1R", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
"_ippicviMirror_16u_C3R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviNormDiff_L1_16u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviMirror_32f_C3IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicvsFree", referenced from:
_iwiImage_Release in libippiw.a(iw_image.c.o)
"_ippicvInit", referenced from:
cv::ipp::IPPInitSingleton::IPPInitSingleton() in system.cpp.o
"_ippicviCopy_32f_C3P3R", referenced from:
_llwiCopySplit in libippiw.a(iw_image_op_copy_split.c.o)
"_ippicvGetEnabledCpuFeatures", referenced from:
cv::ipp::IPPInitSingleton::IPPInitSingleton() in system.cpp.o
"_ippicviCountInRange_8u_C1R", referenced from:
cv::countNonZero(cv::_InputArray const&) in count_non_zero.dispatch.cpp.o
"_ippicviNormDiff_Inf_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_Inf_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L1_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L1_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsExp_32f_A21", referenced from:
cv::hal::exp32f(float const*, float*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicvsLn_32f_A21", referenced from:
cv::hal::log32f(float const*, float*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicviNormRel_L2_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L1_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L1_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviXor_8u_C1R", referenced from:
cv::hal::xor8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNormDiff_L2_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L1_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_Inf_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_Inf_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviMirror_32f_C3R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviNormRel_Inf_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L2_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCopy_16u_C1C3R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviNorm_L1_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviMirror_32f_C4IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviNorm_L1_16u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviDCTFwdGetSize_32f", referenced from:
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviNormRel_L1_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCopy_32s_C1MR", referenced from:
cv::copyMask32s(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
"_ippicviNormRel_Inf_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviMirror_8u_C4IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviNormDiff_Inf_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L1_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsExp_64f_A50", referenced from:
cv::hal::exp64f(double const*, double*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicvsLn_64f_A50", referenced from:
cv::hal::log64f(double const*, double*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicviNormRel_L2_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_Inf_16u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_Inf_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L1_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviSet_32s_C4MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviNormRel_L2_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_Inf_32f_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L1_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_16u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCopy_8u_C4MR", referenced from:
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviNormDiff_Inf_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_8u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsMaxEvery_32f", referenced from:
cv::hal::max32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNorm_L1_32f_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L1_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviSub_16s_C1RSfs", referenced from:
cv::hal::sub16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsDFTGetSize_C_64fc", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicviNorm_L2_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviSub_8u_C1RSfs", referenced from:
cv::hal::sub8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNorm_L1_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L2_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L1_8u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L1_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_8u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L2_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviSet_16u_C1MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviNormDiff_L1_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviAdd_16u_C1RSfs", referenced from:
cv::hal::add16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsPolarToCart_64f", referenced from:
cv::polarToCart(cv::_InputArray const&, cv::_InputArray const&, cv::_OutputArray const&, cv::_OutputArray const&, bool) in mathfuncs.cpp.o
"_ippicviCopy_8u_C4P4R", referenced from:
_llwiCopySplit in libippiw.a(iw_image_op_copy_split.c.o)
"_ippicviCompare_16u_C1R", referenced from:
cv::hal::cmp16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNorm_Inf_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L2_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormRel_L2_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L2_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCopy_32s_C3MR", referenced from:
cv::copyMask32sC3(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
"_ippicvsDFTGetSize_R_32f", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicviNorm_Inf_32f_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L1_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_16u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L2_32f_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L2_8u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCopy_16u_C4C1R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviMul_8u_C1RSfs", referenced from:
cv::hal::mul8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviAbsDiff_32f_C1R", referenced from:
cv::hal::absdiff32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviDFTFwd_RToPack_32f_C1R", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
"_ippicviMirror_8u_C3R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicvsInvSqrt_32f_A21", referenced from:
cv::hal::invSqrt32f(float const*, float*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicviNormDiff_Inf_8u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviSet_8u_C1MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviCopy_8u_C1R_L", referenced from:
cv::Mat::copyTo(cv::_OutputArray const&) const in copy.cpp.o
"_ippicviMirror_8u_C3IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicvsInvSqrt_64f_A50", referenced from:
cv::hal::invSqrt64f(double const*, double*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicvsMagnitude_64f", referenced from:
cv::hal::magnitude64f(double const*, double const*, double*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicviNot_8u_C1R", referenced from:
cv::hal::not8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviGetLibVersion", referenced from:
cv::ipp::IPPInitSingleton::IPPInitSingleton() in system.cpp.o
"_ippicvsMagnitude_32f", referenced from:
cv::hal::magnitude32f(float const*, float const*, float*, int) in mathfuncs_core.dispatch.cpp.o
"_ippicviDCTInvInit_32f", referenced from:
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicvMalloc_L", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::Dft_C_IPPLoop_Invoker<cv::IPPDFT_C_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
cv::Dft_R_IPPLoop_Invoker<cv::IPPDFT_R_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviCountInRange_32f_C1R", referenced from:
cv::countNonZero(cv::_InputArray const&) in count_non_zero.dispatch.cpp.o
"_ippicvsDFTGetSize_C_32fc", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicviDFTInv_CToC_32fc_C1R", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
"_ippicviSet_8u_C4MR", referenced from:
_llwiSetMask in libippiw.a(iw_image_op_set.c.o)
"_ippicviNorm_L2_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviDCTFwd_32f_C1R", referenced from:
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicvsDFTInit_C_64fc", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicviDFTInit_R_32f", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::Dft_R_IPPLoop_Invoker<cv::IPPDFT_R_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviNorm_L2_16u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L1_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviDFTInit_C_32fc", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::Dft_C_IPPLoop_Invoker<cv::IPPDFT_C_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviDFTFwd_CToC_32fc_C1R", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
"_ippicviMirror_8u_C4R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviNormRel_Inf_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviOr_8u_C1R", referenced from:
cv::hal::or8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsDFTInv_PackToR_64f", referenced from:
void cv::CCSIDFT<double>(cv::OcvDftOptions const&, double const*, double*) in dxt.cpp.o
"_ippicvsDFTFwd_RToPack_64f", referenced from:
void cv::RealDFT<double>(cv::OcvDftOptions const&, double const*, double*) in dxt.cpp.o
"_ippicviMirror_32f_C4R", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviCopy_16u_C3CR", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_16u_C1MR", referenced from:
cv::copyMask16u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicvsDFTInv_CToC_64fc", referenced from:
void cv::DFT<double>(cv::OcvDftOptions const&, cv::Complex<double> const*, cv::Complex<double>*) in dxt.cpp.o
"_ippicvsDFTInv_PackToR_32f", referenced from:
void cv::CCSIDFT<float>(cv::OcvDftOptions const&, float const*, float*) in dxt.cpp.o
"_ippicviNormRel_Inf_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L1_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_Inf_16u_C1R", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsDFTFwd_RToPack_32f", referenced from:
void cv::RealDFT<float>(cv::OcvDftOptions const&, float const*, float*) in dxt.cpp.o
"_ippicvsDFTInv_CToC_32fc", referenced from:
void cv::DFT<float>(cv::OcvDftOptions const&, cv::Complex<float> const*, cv::Complex<float>*) in dxt.cpp.o
"_ippicvsDFTFwd_CToC_32fc", referenced from:
void cv::DFT<float>(cv::OcvDftOptions const&, cv::Complex<float> const*, cv::Complex<float>*) in dxt.cpp.o
"_ippicviCompare_16s_C1R", referenced from:
cv::hal::cmp16s(short const*, unsigned long, short const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsDFTInit_C_32fc", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicvFree", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::Dft_C_IPPLoop_Invoker<cv::IPPDFT_C_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
cv::Dft_R_IPPLoop_Invoker<cv::IPPDFT_R_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviNormDiff_L1_32f_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_Inf_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L2_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsDFTFwd_CToC_64fc", referenced from:
void cv::DFT<double>(cv::OcvDftOptions const&, cv::Complex<double> const*, cv::Complex<double>*) in dxt.cpp.o
"_ippicviNormRel_L2_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCopy_8u_C1C3R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviNorm_L2_8u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsMinEvery_64f", referenced from:
cv::hal::min64f(double const*, unsigned long, double const*, unsigned long, double*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviCopy_8u_C3MR", referenced from:
cv::copyMask8uC3(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviCopy_8u_C3CR", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviCopy_8u_C1MR", referenced from:
cv::copyMask8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviNormDiff_L2_32f_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviMul_32f_C1R", referenced from:
cv::hal::mul32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviMul_16u_C1RSfs", referenced from:
cv::hal::mul16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNormRel_Inf_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviDCTFwdInit_32f", referenced from:
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviNormDiff_L1_32f_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviCompare_32f_C1R", referenced from:
cv::hal::cmp32f(float const*, unsigned long, float const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNormDiff_L2_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNormDiff_L1_8u_C3CMR", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviAbsDiff_8u_C1R", referenced from:
cv::hal::absdiff8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNorm_L1_16u_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviNorm_L2_32f_C1MR", referenced from:
cv::norm(cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsPolarToCart_32f", referenced from:
cv::polarToCart(cv::_InputArray const&, cv::_InputArray const&, cv::_OutputArray const&, cv::_OutputArray const&, bool) in mathfuncs.cpp.o
"_ippicviDFTGetSize_C_32fc", referenced from:
cv::OcvDftImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::Dft_C_IPPLoop_Invoker<cv::IPPDFT_C_Functor>::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviCopy_32f_C1MR", referenced from:
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviCopy_32f_C3CR", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviMirror_16u_C4IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
"_ippicviNormDiff_L1_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsMaxEvery_8u", referenced from:
cv::hal::max8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsDFTGetSize_R_64f", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicviCompare_8u_C1R", referenced from:
cv::hal::cmp8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviNormRel_L1_8u_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicviAdd_8u_C1RSfs", referenced from:
cv::hal::add8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviCopy_32f_C1C4R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviNormRel_Inf_16s_C1R", referenced from:
cv::norm(cv::_InputArray const&, cv::_InputArray const&, int, cv::_InputArray const&) in norm.cpp.o
"_ippicvsMinEvery_32f", referenced from:
cv::hal::min32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsMinEvery_8u", referenced from:
cv::hal::min8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviAbsDiff_16u_C1R", referenced from:
cv::hal::absdiff16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsMaxEvery_64f", referenced from:
cv::hal::max64f(double const*, unsigned long, double const*, unsigned long, double*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvsDFTInit_R_64f", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicvsMinEvery_16u", referenced from:
cv::hal::min16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviAdd_32f_C1R", referenced from:
cv::hal::add32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviMul_16s_C1RSfs", referenced from:
cv::hal::mul16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviDCTInv_32f_C1R", referenced from:
cv::OcvDctImpl::apply(unsigned char const*, unsigned long, unsigned char*, unsigned long) in dxt.cpp.o
cv::DctIPPLoop_Invoker::operator()(cv::Range const&) const in dxt.cpp.o
"_ippicviSub_32f_C1R", referenced from:
cv::hal::sub32f(float const*, unsigned long, float const*, unsigned long, float*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicvSetCpuFeatures", referenced from:
cv::ipp::IPPInitSingleton::IPPInitSingleton() in system.cpp.o
"_ippicviAdd_16s_C1RSfs", referenced from:
cv::hal::add16s(short const*, unsigned long, short const*, unsigned long, short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviCopy_8u_C1C4R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviSub_16u_C1RSfs", referenced from:
cv::hal::sub16u(unsigned short const*, unsigned long, unsigned short const*, unsigned long, unsigned short*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviCopy_32s_C4MR", referenced from:
cv::copyMask32sC4(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
"_ippicvsDFTInit_R_32f", referenced from:
cv::OcvDftBasicImpl::init(int, int, int, int, bool*) in dxt.cpp.o
"_ippicviCopy_32f_C4C1R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicvGetCpuFeatures", referenced from:
cv::ipp::IPPInitSingleton::IPPInitSingleton() in system.cpp.o
"_ippicviCopy_16u_C3MR", referenced from:
cv::copyMask16uC3(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, cv::Size_<int>, void*) in copy.cpp.o
_llwiCopyMask in libippiw.a(iw_image_op_copy.c.o)
"_ippicviCopy_32f_C3C1R", referenced from:
_llwiCopyChannel in libippiw.a(iw_image_op_copy_channel.c.o)
"_ippicviAnd_8u_C1R", referenced from:
cv::hal::and8u(unsigned char const*, unsigned long, unsigned char const*, unsigned long, unsigned char*, unsigned long, int, int, void*) in arithm.dispatch.cpp.o
"_ippicviMirror_16u_C3IR", referenced from:
_llwiMirror in libippiw.a(iw_image_transform_mirror.c.o)
ld: symbol(s) not found for architecture x86_64
```
##### Steps to reproduce
- clone opencv repo master at commit 5dd3e6052ed9e614cf4303eee64d092fefbabd8e
- clone opencv_contrib master at commit 61cbb44539c5a8b07315c9a6c38f5350c0e91
- cd opencv/build
- mkdir tmp
- cmake -DOPENCV_EXTRA_MODULES_PATH=<path to>/opencv_contrib/modules ..
- make -j2
| non_defect | opencv linkage fails if you have a question rather than reporting a bug please go to where you get much faster responses if you need further assistance please read this is a template helping you to create an issue which can be processed as quickly as possible this is the bug reporting section for the opencv library system information version example opencv operating system platform windows bit compiler visual studio opencv dev operating system platform macos compiler clang detailed description opencv compiles successfully but linkage fails with undefined symbols for architecture ippicviset referenced from llwisetmask in libippiw a iw image op set c o ippicvsround referenced from llwisetmask in libippiw a iw image op set c o ippicviset referenced from llwisetmask in libippiw a iw image op set c o ippicvsconvert referenced from llwisetmask in libippiw a iw image op set c o ippicviset referenced from llwisetmask in libippiw a iw image op set c o ippicviset referenced from llwisetmask in libippiw a iw image op set c o ippicviset referenced from llwisetmask in libippiw a iw image op set c o ippicviset referenced from llwisetmask in libippiw a iw image op set c o ippicviset referenced from llwisetmask in libippiw a iw image op set c o ippicvicopy referenced from llwicopysplit in libippiw a iw image op copy split c o ippicvicopy referenced from llwicopysplit in libippiw a iw image op copy split c o ippicvicopy referenced from llwicopysplit in libippiw a iw image op copy split c o ippicvicopy referenced from llwicopysplit in libippiw a iw image op copy split c o ippicvicopy referenced from llwicopymerge in libippiw a iw image op copy merge c o ippicvicopy referenced from llwicopymerge in libippiw a iw image op copy merge c o ippicvicopy referenced from llwicopymerge in libippiw a iw image op copy merge c o ippicvicopy referenced from llwicopymerge in libippiw a iw image op copy merge c o ippicvicopy referenced from llwicopymerge in libippiw a iw image op copy merge c o ippicvicopy referenced from llwicopymask in libippiw a iw image op copy c o ippicvicopy referenced from llwicopymask in libippiw a iw image op copy c o ippicvicopy referenced from llwicopymask in libippiw a iw image op copy c o ippicviset referenced from llwisetmask in libippiw a iw image op set c o ippicvicopy referenced from llwicopymerge in libippiw a iw image op copy merge c o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvidctinvgetsize referenced from cv ocvdctimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dctipploop invoker operator cv range const const in dxt cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvsmaxevery referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in arithm dispatch cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvidftgetsize r referenced from cv ocvdftimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dft r ipploop invoker operator cv range const const in dxt cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvidftinv packtor referenced from cv ocvdftimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvsfree referenced from iwiimage release in libippiw a iw image c o ippicvinit referenced from cv ipp ippinitsingleton ippinitsingleton in system cpp o ippicvicopy referenced from llwicopysplit in libippiw a iw image op copy split c o ippicvgetenabledcpufeatures referenced from cv ipp ippinitsingleton ippinitsingleton in system cpp o ippicvicountinrange referenced from cv countnonzero cv inputarray const in count non zero dispatch cpp o ippicvinormdiff inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvsexp referenced from cv hal float const float int in mathfuncs core dispatch cpp o ippicvsln referenced from cv hal float const float int in mathfuncs core dispatch cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvixor referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormrel inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvinormrel inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvidctfwdgetsize referenced from cv ocvdctimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dctipploop invoker operator cv range const const in dxt cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvicopy referenced from cv unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long cv size void in copy cpp o ippicvinormrel inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvinormdiff inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvsexp referenced from cv hal double const double int in mathfuncs core dispatch cpp o ippicvsln referenced from cv hal double const double int in mathfuncs core dispatch cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicviset referenced from llwisetmask in libippiw a iw image op set c o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvicopy referenced from llwicopymask in libippiw a iw image op copy c o ippicvinormdiff inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinorm inf referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvsmaxevery referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in arithm dispatch cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm inf referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvisub referenced from cv hal short const unsigned long short const unsigned long short unsigned long int int void in arithm dispatch cpp o ippicvsdftgetsize c referenced from cv ocvdftbasicimpl init int int int int bool in dxt cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm inf referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvisub referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicviset referenced from llwisetmask in libippiw a iw image op set c o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicviadd referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in arithm dispatch cpp o ippicvspolartocart referenced from cv polartocart cv inputarray const cv inputarray const cv outputarray const cv outputarray const bool in mathfuncs cpp o ippicvicopy referenced from llwicopysplit in libippiw a iw image op copy split c o ippicvicompare referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicvinorm inf referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvicopy referenced from cv unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long cv size void in copy cpp o ippicvsdftgetsize r referenced from cv ocvdftbasicimpl init int int int int bool in dxt cpp o ippicvinorm inf referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm inf referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinorm inf referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm inf referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvimul referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicviabsdiff referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in arithm dispatch cpp o ippicvidftfwd rtopack referenced from cv ocvdftimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvsinvsqrt referenced from cv hal float const float int in mathfuncs core dispatch cpp o ippicvinormdiff inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicviset referenced from llwisetmask in libippiw a iw image op set c o ippicvicopy l referenced from cv mat copyto cv outputarray const const in copy cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvsinvsqrt referenced from cv hal double const double int in mathfuncs core dispatch cpp o ippicvsmagnitude referenced from cv hal double const double const double int in mathfuncs core dispatch cpp o ippicvinot referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicvigetlibversion referenced from cv ipp ippinitsingleton ippinitsingleton in system cpp o ippicvsmagnitude referenced from cv hal float const float const float int in mathfuncs core dispatch cpp o ippicvidctinvinit referenced from cv ocvdctimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dctipploop invoker operator cv range const const in dxt cpp o ippicvmalloc l referenced from cv ocvdftimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dft c ipploop invoker operator cv range const const in dxt cpp o cv dft r ipploop invoker operator cv range const const in dxt cpp o cv ocvdctimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dctipploop invoker operator cv range const const in dxt cpp o ippicvicountinrange referenced from cv countnonzero cv inputarray const in count non zero dispatch cpp o ippicvsdftgetsize c referenced from cv ocvdftbasicimpl init int int int int bool in dxt cpp o ippicvidftinv ctoc referenced from cv ocvdftimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o ippicviset referenced from llwisetmask in libippiw a iw image op set c o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvidctfwd referenced from cv ocvdctimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dctipploop invoker operator cv range const const in dxt cpp o ippicvsdftinit c referenced from cv ocvdftbasicimpl init int int int int bool in dxt cpp o ippicvidftinit r referenced from cv ocvdftimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dft r ipploop invoker operator cv range const const in dxt cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm inf referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvidftinit c referenced from cv ocvdftimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dft c ipploop invoker operator cv range const const in dxt cpp o ippicvidftfwd ctoc referenced from cv ocvdftimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvinormrel inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvior referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicvsdftinv packtor referenced from void cv ccsidft cv ocvdftoptions const double const double in dxt cpp o ippicvsdftfwd rtopack referenced from void cv realdft cv ocvdftoptions const double const double in dxt cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvicopy referenced from cv unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long cv size void in copy cpp o llwicopymask in libippiw a iw image op copy c o ippicvsdftinv ctoc referenced from void cv dft cv ocvdftoptions const cv complex const cv complex in dxt cpp o ippicvsdftinv packtor referenced from void cv ccsidft cv ocvdftoptions const float const float in dxt cpp o ippicvinormrel inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm inf referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvsdftfwd rtopack referenced from void cv realdft cv ocvdftoptions const float const float in dxt cpp o ippicvsdftinv ctoc referenced from void cv dft cv ocvdftoptions const cv complex const cv complex in dxt cpp o ippicvsdftfwd ctoc referenced from void cv dft cv ocvdftoptions const cv complex const cv complex in dxt cpp o ippicvicompare referenced from cv hal short const unsigned long short const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicvsdftinit c referenced from cv ocvdftbasicimpl init int int int int bool in dxt cpp o ippicvfree referenced from cv ocvdftimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dft c ipploop invoker operator cv range const const in dxt cpp o cv dft r ipploop invoker operator cv range const const in dxt cpp o cv ocvdctimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dctipploop invoker operator cv range const const in dxt cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvsdftfwd ctoc referenced from void cv dft cv ocvdftoptions const cv complex const cv complex in dxt cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvsminevery referenced from cv hal double const unsigned long double const unsigned long double unsigned long int int void in arithm dispatch cpp o ippicvicopy referenced from cv unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long cv size void in copy cpp o llwicopymask in libippiw a iw image op copy c o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvicopy referenced from cv unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long cv size void in copy cpp o llwicopymask in libippiw a iw image op copy c o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvimul referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in arithm dispatch cpp o ippicvimul referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in arithm dispatch cpp o ippicvinormrel inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvidctfwdinit referenced from cv ocvdctimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dctipploop invoker operator cv range const const in dxt cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvicompare referenced from cv hal float const unsigned long float const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicviabsdiff referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvinorm referenced from cv norm cv inputarray const int cv inputarray const in norm cpp o ippicvspolartocart referenced from cv polartocart cv inputarray const cv inputarray const cv outputarray const cv outputarray const bool in mathfuncs cpp o ippicvidftgetsize c referenced from cv ocvdftimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dft c ipploop invoker operator cv range const const in dxt cpp o ippicvicopy referenced from llwicopymask in libippiw a iw image op copy c o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ippicvinormdiff referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvsmaxevery referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicvsdftgetsize r referenced from cv ocvdftbasicimpl init int int int int bool in dxt cpp o ippicvicompare referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicvinormrel referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicviadd referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvinormrel inf referenced from cv norm cv inputarray const cv inputarray const int cv inputarray const in norm cpp o ippicvsminevery referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in arithm dispatch cpp o ippicvsminevery referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicviabsdiff referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in arithm dispatch cpp o ippicvsmaxevery referenced from cv hal double const unsigned long double const unsigned long double unsigned long int int void in arithm dispatch cpp o ippicvsdftinit r referenced from cv ocvdftbasicimpl init int int int int bool in dxt cpp o ippicvsminevery referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in arithm dispatch cpp o ippicviadd referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in arithm dispatch cpp o ippicvimul referenced from cv hal short const unsigned long short const unsigned long short unsigned long int int void in arithm dispatch cpp o ippicvidctinv referenced from cv ocvdctimpl apply unsigned char const unsigned long unsigned char unsigned long in dxt cpp o cv dctipploop invoker operator cv range const const in dxt cpp o ippicvisub referenced from cv hal float const unsigned long float const unsigned long float unsigned long int int void in arithm dispatch cpp o ippicvsetcpufeatures referenced from cv ipp ippinitsingleton ippinitsingleton in system cpp o ippicviadd referenced from cv hal short const unsigned long short const unsigned long short unsigned long int int void in arithm dispatch cpp o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvisub referenced from cv hal unsigned short const unsigned long unsigned short const unsigned long unsigned short unsigned long int int void in arithm dispatch cpp o ippicvicopy referenced from cv unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long cv size void in copy cpp o ippicvsdftinit r referenced from cv ocvdftbasicimpl init int int int int bool in dxt cpp o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicvgetcpufeatures referenced from cv ipp ippinitsingleton ippinitsingleton in system cpp o ippicvicopy referenced from cv unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long cv size void in copy cpp o llwicopymask in libippiw a iw image op copy c o ippicvicopy referenced from llwicopychannel in libippiw a iw image op copy channel c o ippicviand referenced from cv hal unsigned char const unsigned long unsigned char const unsigned long unsigned char unsigned long int int void in arithm dispatch cpp o ippicvimirror referenced from llwimirror in libippiw a iw image transform mirror c o ld symbol s not found for architecture steps to reproduce clone opencv repo master at commit clone opencv contrib master at commit cd opencv build mkdir tmp cmake dopencv extra modules path opencv contrib modules make | 0 |
37,707 | 8,481,832,374 | IssuesEvent | 2018-10-25 16:46:21 | CenturyLinkCloud/mdw | https://api.github.com/repos/CenturyLinkCloud/mdw | opened | Subprocess activities in Hub definition view = issues when ' v' included in name | defect | The `processname` attribute causes problems identifying the configured subprocess due to this logic in configurator.js:
```javascript
Configurator.prototype.initSubprocBindings = function(widget, subproc) {
var spaceV = subproc.lastIndexOf(' v');
if (spaceV > 0)
subproc = subproc(0, spaceV);
``` | 1.0 | Subprocess activities in Hub definition view = issues when ' v' included in name - The `processname` attribute causes problems identifying the configured subprocess due to this logic in configurator.js:
```javascript
Configurator.prototype.initSubprocBindings = function(widget, subproc) {
var spaceV = subproc.lastIndexOf(' v');
if (spaceV > 0)
subproc = subproc(0, spaceV);
``` | defect | subprocess activities in hub definition view issues when v included in name the processname attribute causes problems identifying the configured subprocess due to this logic in configurator js javascript configurator prototype initsubprocbindings function widget subproc var spacev subproc lastindexof v if spacev subproc subproc spacev | 1 |
25,612 | 5,187,032,909 | IssuesEvent | 2017-01-20 15:45:50 | spring-projects/spring-boot | https://api.github.com/repos/spring-projects/spring-boot | closed | Document how to insert test data with flyway | priority: normal status: ideal-for-contribution type: documentation | When I include the Flyway library in my classpath, Spring Boot correctly locates and executes my migrations. It does not, however, make use of my data.sql script, which I use to import fixture data for unit tests. Per the documentation, I am setting the `jpa.hibernate.ddl-auto` property to `false` so that Hibernate doesn't try to generate the schema itself. If I remove this property, then Spring boot _does_ load the data.sql file, but Hibernate is also clearly trying to generate the schema redundantly, which is at the very least unnecessary but also probably error prone.
Is there a way to use Flyway along with the normal Spring Boot datasource initialization features?
```
2015-03-30 16:04:09.851 INFO 56097 --- [ Test worker] o.f.core.internal.util.VersionPrinter : Flyway 3.2.1 by Boxfuse
2015-03-30 16:04:10.113 INFO 56097 --- [ Test worker] o.f.c.i.dbsupport.DbSupportFactory : Database: jdbc:h2:mem:testdb (H2 1.4)
2015-03-30 16:04:10.169 INFO 56097 --- [ Test worker] o.f.core.internal.command.DbValidate : Validated 2 migrations (execution time 00:00.013s)
2015-03-30 16:04:10.183 INFO 56097 --- [ Test worker] o.f.c.i.metadatatable.MetaDataTableImpl : Creating Metadata table: "PUBLIC"."schema_version"
2015-03-30 16:04:10.209 INFO 56097 --- [ Test worker] o.f.core.internal.command.DbMigrate : Current version of schema "PUBLIC": << Empty Schema >>
2015-03-30 16:04:10.209 INFO 56097 --- [ Test worker] o.f.core.internal.command.DbMigrate : Migrating schema "PUBLIC" to version 1 - create user info tables
2015-03-30 16:04:10.242 INFO 56097 --- [ Test worker] o.f.core.internal.command.DbMigrate : Migrating schema "PUBLIC" to version 2 - create recipe tables
2015-03-30 16:04:10.251 INFO 56097 --- [ Test worker] o.f.core.internal.command.DbMigrate : Successfully applied 2 migrations to schema "PUBLIC" (execution time 00:00.070s).
2015-03-30 16:04:10.364 INFO 56097 --- [ Test worker] j.LocalContainerEntityManagerFactoryBean : Building JPA container EntityManagerFactory for persistence unit 'default'
2015-03-30 16:04:10.378 INFO 56097 --- [ Test worker] o.hibernate.jpa.internal.util.LogHelper : HHH000204: Processing PersistenceUnitInfo [
name: default
...]
2015-03-30 16:04:10.490 INFO 56097 --- [ Test worker] org.hibernate.Version : HHH000412: Hibernate Core {4.3.8.Final}
2015-03-30 16:04:10.492 INFO 56097 --- [ Test worker] org.hibernate.cfg.Environment : HHH000206: hibernate.properties not found
2015-03-30 16:04:10.495 INFO 56097 --- [ Test worker] org.hibernate.cfg.Environment : HHH000021: Bytecode provider name : javassist
2015-03-30 16:04:10.795 INFO 56097 --- [ Test worker] o.hibernate.annotations.common.Version : HCANN000001: Hibernate Commons Annotations {4.0.5.Final}
2015-03-30 16:04:10.853 INFO 56097 --- [ Test worker] org.hibernate.dialect.Dialect : HHH000400: Using dialect: org.hibernate.dialect.H2Dialect
2015-03-30 16:04:10.990 INFO 56097 --- [ Test worker] o.h.h.i.ast.ASTQueryTranslatorFactory : HHH000397: Using ASTQueryTranslatorFactory
2015-03-30 16:04:11.255 INFO 56097 --- [ Test worker] org.hibernate.tool.hbm2ddl.SchemaExport : HHH000227: Running hbm2ddl schema export
2015-03-30 16:04:11.267 INFO 56097 --- [ Test worker] org.hibernate.tool.hbm2ddl.SchemaExport : HHH000230: Schema export complete
2015-03-30 16:04:11.348 INFO 56097 --- [ Test worker] o.s.jdbc.datasource.init.ScriptUtils : Executing SQL script from URL [file:/redacted/main/data.sql]
```
| 1.0 | Document how to insert test data with flyway - When I include the Flyway library in my classpath, Spring Boot correctly locates and executes my migrations. It does not, however, make use of my data.sql script, which I use to import fixture data for unit tests. Per the documentation, I am setting the `jpa.hibernate.ddl-auto` property to `false` so that Hibernate doesn't try to generate the schema itself. If I remove this property, then Spring boot _does_ load the data.sql file, but Hibernate is also clearly trying to generate the schema redundantly, which is at the very least unnecessary but also probably error prone.
Is there a way to use Flyway along with the normal Spring Boot datasource initialization features?
```
2015-03-30 16:04:09.851 INFO 56097 --- [ Test worker] o.f.core.internal.util.VersionPrinter : Flyway 3.2.1 by Boxfuse
2015-03-30 16:04:10.113 INFO 56097 --- [ Test worker] o.f.c.i.dbsupport.DbSupportFactory : Database: jdbc:h2:mem:testdb (H2 1.4)
2015-03-30 16:04:10.169 INFO 56097 --- [ Test worker] o.f.core.internal.command.DbValidate : Validated 2 migrations (execution time 00:00.013s)
2015-03-30 16:04:10.183 INFO 56097 --- [ Test worker] o.f.c.i.metadatatable.MetaDataTableImpl : Creating Metadata table: "PUBLIC"."schema_version"
2015-03-30 16:04:10.209 INFO 56097 --- [ Test worker] o.f.core.internal.command.DbMigrate : Current version of schema "PUBLIC": << Empty Schema >>
2015-03-30 16:04:10.209 INFO 56097 --- [ Test worker] o.f.core.internal.command.DbMigrate : Migrating schema "PUBLIC" to version 1 - create user info tables
2015-03-30 16:04:10.242 INFO 56097 --- [ Test worker] o.f.core.internal.command.DbMigrate : Migrating schema "PUBLIC" to version 2 - create recipe tables
2015-03-30 16:04:10.251 INFO 56097 --- [ Test worker] o.f.core.internal.command.DbMigrate : Successfully applied 2 migrations to schema "PUBLIC" (execution time 00:00.070s).
2015-03-30 16:04:10.364 INFO 56097 --- [ Test worker] j.LocalContainerEntityManagerFactoryBean : Building JPA container EntityManagerFactory for persistence unit 'default'
2015-03-30 16:04:10.378 INFO 56097 --- [ Test worker] o.hibernate.jpa.internal.util.LogHelper : HHH000204: Processing PersistenceUnitInfo [
name: default
...]
2015-03-30 16:04:10.490 INFO 56097 --- [ Test worker] org.hibernate.Version : HHH000412: Hibernate Core {4.3.8.Final}
2015-03-30 16:04:10.492 INFO 56097 --- [ Test worker] org.hibernate.cfg.Environment : HHH000206: hibernate.properties not found
2015-03-30 16:04:10.495 INFO 56097 --- [ Test worker] org.hibernate.cfg.Environment : HHH000021: Bytecode provider name : javassist
2015-03-30 16:04:10.795 INFO 56097 --- [ Test worker] o.hibernate.annotations.common.Version : HCANN000001: Hibernate Commons Annotations {4.0.5.Final}
2015-03-30 16:04:10.853 INFO 56097 --- [ Test worker] org.hibernate.dialect.Dialect : HHH000400: Using dialect: org.hibernate.dialect.H2Dialect
2015-03-30 16:04:10.990 INFO 56097 --- [ Test worker] o.h.h.i.ast.ASTQueryTranslatorFactory : HHH000397: Using ASTQueryTranslatorFactory
2015-03-30 16:04:11.255 INFO 56097 --- [ Test worker] org.hibernate.tool.hbm2ddl.SchemaExport : HHH000227: Running hbm2ddl schema export
2015-03-30 16:04:11.267 INFO 56097 --- [ Test worker] org.hibernate.tool.hbm2ddl.SchemaExport : HHH000230: Schema export complete
2015-03-30 16:04:11.348 INFO 56097 --- [ Test worker] o.s.jdbc.datasource.init.ScriptUtils : Executing SQL script from URL [file:/redacted/main/data.sql]
```
| non_defect | document how to insert test data with flyway when i include the flyway library in my classpath spring boot correctly locates and executes my migrations it does not however make use of my data sql script which i use to import fixture data for unit tests per the documentation i am setting the jpa hibernate ddl auto property to false so that hibernate doesn t try to generate the schema itself if i remove this property then spring boot does load the data sql file but hibernate is also clearly trying to generate the schema redundantly which is at the very least unnecessary but also probably error prone is there a way to use flyway along with the normal spring boot datasource initialization features info o f core internal util versionprinter flyway by boxfuse info o f c i dbsupport dbsupportfactory database jdbc mem testdb info o f core internal command dbvalidate validated migrations execution time info o f c i metadatatable metadatatableimpl creating metadata table public schema version info o f core internal command dbmigrate current version of schema public info o f core internal command dbmigrate migrating schema public to version create user info tables info o f core internal command dbmigrate migrating schema public to version create recipe tables info o f core internal command dbmigrate successfully applied migrations to schema public execution time info j localcontainerentitymanagerfactorybean building jpa container entitymanagerfactory for persistence unit default info o hibernate jpa internal util loghelper processing persistenceunitinfo name default info org hibernate version hibernate core final info org hibernate cfg environment hibernate properties not found info org hibernate cfg environment bytecode provider name javassist info o hibernate annotations common version hibernate commons annotations final info org hibernate dialect dialect using dialect org hibernate dialect info o h h i ast astquerytranslatorfactory using astquerytranslatorfactory info org hibernate tool schemaexport running schema export info org hibernate tool schemaexport schema export complete info o s jdbc datasource init scriptutils executing sql script from url | 0 |
90,256 | 26,024,561,196 | IssuesEvent | 2022-12-21 15:18:51 | numpy/numpy | https://api.github.com/repos/numpy/numpy | closed | Fails to import with numpy 1.23.5 after compiling on raspbian arm with clang 13 for python 3.11 | 00 - Bug 36 - Build | ### Steps to reproduce:
Pip install build from source as there is no wheel for python 3.11
```
(homeassistant-shadow) homeassistant@lazaro:/srv/homeassistant-shadow$ pip install numpy
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Collecting numpy
Downloading numpy-1.23.5.tar.gz (10.7 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 10.7/10.7 MB 8.4 MB/s eta 0:00:00
Installing build dependencies ... done
Getting requirements to build wheel ... done
Preparing metadata (pyproject.toml) ... done
Building wheels for collected packages: numpy
Building wheel for numpy (pyproject.toml) ... done
Created wheel for numpy: filename=numpy-1.23.5-cp311-cp311-linux_armv7l.whl size=10681800 sha256=38fff5fe9473a06d14419d01d7e1321906d59765930bd20be43ce30f42e282b6
Stored in directory: /home/homeassistant/.cache/pip/wheels/32/6f/c1/c2fda686f19bc32d64822ae92e3dc4922f9a999c703d998abc
Successfully built numpy
Installing collected packages: numpy
Successfully installed numpy-1.23.5
```
### Error message:
```shell
Traceback (most recent call last):
File "/srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/core/__init__.py", line 23, in <module>
from . import multiarray
File "/srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/core/multiarray.py", line 10, in <module>
from . import overrides
File "/srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/core/overrides.py", line 6, in <module>
from numpy.core._multiarray_umath import (
ImportError: /srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/core/_multiarray_umath.cpython-311-arm-linux-gnueabihf.so: undefined symbol: __mulodi4
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/__init__.py", line 140, in <module>
from . import core
File "/srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/core/__init__.py", line 49, in <module>
raise ImportError(msg)
ImportError:
IMPORTANT: PLEASE READ THIS FOR ADVICE ON HOW TO SOLVE THIS ISSUE!
Importing the numpy C-extensions failed. This error can happen for
many reasons, often due to issues with your setup or how NumPy was
installed.
We have compiled some common reasons and troubleshooting tips at:
https://numpy.org/devdocs/user/troubleshooting-importerror.html
Please note and check the following:
* The Python version is: Python3.11 from "/srv/homeassistant-shadow/bin/python"
* The NumPy version is: "1.23.5"
and make sure that they are the versions you expect.
Please carefully study the documentation linked above for further help.
Original error was: /srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/core/_multiarray_umath.cpython-311-arm-linux-gnueabihf.so: undefined symbol: __mulodi4
```
### Additional information:
Both python and numpy where compiled on the raspberry pi 4 running bullseye using clang-13
In particular, for compiling python the following was run
```
export CC=/usr/bin/clang-13
export CXX=/usrbin/clang++-13
export LDFLAGS=" -fuse-ld=lld"
ln -s /usr/bin/llvm-arm-13 /usr/local/bin/llv-arm
ln -s /usr/bin/llvm-profdata-13 /usr/local/bin/llvm-profdata
./configure --prefix /usr/local/cpython-3.11 --with-lto=full --enable-optimizations
```
python -m test passes correctly.
The error showed up inside a venv created with the previously compiled and installed python 3.11.
I'm running another build to get the build log and will attach it when it's done. | 1.0 | Fails to import with numpy 1.23.5 after compiling on raspbian arm with clang 13 for python 3.11 - ### Steps to reproduce:
Pip install build from source as there is no wheel for python 3.11
```
(homeassistant-shadow) homeassistant@lazaro:/srv/homeassistant-shadow$ pip install numpy
Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple
Collecting numpy
Downloading numpy-1.23.5.tar.gz (10.7 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 10.7/10.7 MB 8.4 MB/s eta 0:00:00
Installing build dependencies ... done
Getting requirements to build wheel ... done
Preparing metadata (pyproject.toml) ... done
Building wheels for collected packages: numpy
Building wheel for numpy (pyproject.toml) ... done
Created wheel for numpy: filename=numpy-1.23.5-cp311-cp311-linux_armv7l.whl size=10681800 sha256=38fff5fe9473a06d14419d01d7e1321906d59765930bd20be43ce30f42e282b6
Stored in directory: /home/homeassistant/.cache/pip/wheels/32/6f/c1/c2fda686f19bc32d64822ae92e3dc4922f9a999c703d998abc
Successfully built numpy
Installing collected packages: numpy
Successfully installed numpy-1.23.5
```
### Error message:
```shell
Traceback (most recent call last):
File "/srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/core/__init__.py", line 23, in <module>
from . import multiarray
File "/srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/core/multiarray.py", line 10, in <module>
from . import overrides
File "/srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/core/overrides.py", line 6, in <module>
from numpy.core._multiarray_umath import (
ImportError: /srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/core/_multiarray_umath.cpython-311-arm-linux-gnueabihf.so: undefined symbol: __mulodi4
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/__init__.py", line 140, in <module>
from . import core
File "/srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/core/__init__.py", line 49, in <module>
raise ImportError(msg)
ImportError:
IMPORTANT: PLEASE READ THIS FOR ADVICE ON HOW TO SOLVE THIS ISSUE!
Importing the numpy C-extensions failed. This error can happen for
many reasons, often due to issues with your setup or how NumPy was
installed.
We have compiled some common reasons and troubleshooting tips at:
https://numpy.org/devdocs/user/troubleshooting-importerror.html
Please note and check the following:
* The Python version is: Python3.11 from "/srv/homeassistant-shadow/bin/python"
* The NumPy version is: "1.23.5"
and make sure that they are the versions you expect.
Please carefully study the documentation linked above for further help.
Original error was: /srv/homeassistant-shadow/lib/python3.11/site-packages/numpy/core/_multiarray_umath.cpython-311-arm-linux-gnueabihf.so: undefined symbol: __mulodi4
```
### Additional information:
Both python and numpy where compiled on the raspberry pi 4 running bullseye using clang-13
In particular, for compiling python the following was run
```
export CC=/usr/bin/clang-13
export CXX=/usrbin/clang++-13
export LDFLAGS=" -fuse-ld=lld"
ln -s /usr/bin/llvm-arm-13 /usr/local/bin/llv-arm
ln -s /usr/bin/llvm-profdata-13 /usr/local/bin/llvm-profdata
./configure --prefix /usr/local/cpython-3.11 --with-lto=full --enable-optimizations
```
python -m test passes correctly.
The error showed up inside a venv created with the previously compiled and installed python 3.11.
I'm running another build to get the build log and will attach it when it's done. | non_defect | fails to import with numpy after compiling on raspbian arm with clang for python steps to reproduce pip install build from source as there is no wheel for python homeassistant shadow homeassistant lazaro srv homeassistant shadow pip install numpy looking in indexes collecting numpy downloading numpy tar gz mb ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ mb mb s eta installing build dependencies done getting requirements to build wheel done preparing metadata pyproject toml done building wheels for collected packages numpy building wheel for numpy pyproject toml done created wheel for numpy filename numpy linux whl size stored in directory home homeassistant cache pip wheels successfully built numpy installing collected packages numpy successfully installed numpy error message shell traceback most recent call last file srv homeassistant shadow lib site packages numpy core init py line in from import multiarray file srv homeassistant shadow lib site packages numpy core multiarray py line in from import overrides file srv homeassistant shadow lib site packages numpy core overrides py line in from numpy core multiarray umath import importerror srv homeassistant shadow lib site packages numpy core multiarray umath cpython arm linux gnueabihf so undefined symbol during handling of the above exception another exception occurred traceback most recent call last file line in file srv homeassistant shadow lib site packages numpy init py line in from import core file srv homeassistant shadow lib site packages numpy core init py line in raise importerror msg importerror important please read this for advice on how to solve this issue importing the numpy c extensions failed this error can happen for many reasons often due to issues with your setup or how numpy was installed we have compiled some common reasons and troubleshooting tips at please note and check the following the python version is from srv homeassistant shadow bin python the numpy version is and make sure that they are the versions you expect please carefully study the documentation linked above for further help original error was srv homeassistant shadow lib site packages numpy core multiarray umath cpython arm linux gnueabihf so undefined symbol additional information both python and numpy where compiled on the raspberry pi running bullseye using clang in particular for compiling python the following was run export cc usr bin clang export cxx usrbin clang export ldflags fuse ld lld ln s usr bin llvm arm usr local bin llv arm ln s usr bin llvm profdata usr local bin llvm profdata configure prefix usr local cpython with lto full enable optimizations python m test passes correctly the error showed up inside a venv created with the previously compiled and installed python i m running another build to get the build log and will attach it when it s done | 0 |
79,547 | 28,367,954,287 | IssuesEvent | 2023-04-12 15:00:25 | vector-im/element-ios | https://api.github.com/repos/vector-im/element-ios | opened | User suggestion list height doesn't match container on iOS 16+ | T-Defect | ### Steps to reproduce
1. Go to a room
2. Type @... with a pattern that would matches just a couple/few users
### Outcome
#### What did you expect?
User suggestion list displays multiple users nicely, given that there is enough room for it
#### What happened instead?
User suggestion list height is bigger than the container's (container size is "manually" calculated). This is probably due to the fact that on iOS 16, SwiftUI's `List` component switches from using a `UITableView` to a `UICollectionView`. The implemented calculation matches the expected standard vertical spacing of the previous component.

### Your phone model
Any
### Operating system version
iOS 16.0+
### Application version
develop
### Homeserver
Any
### Will you send logs?
No | 1.0 | User suggestion list height doesn't match container on iOS 16+ - ### Steps to reproduce
1. Go to a room
2. Type @... with a pattern that would matches just a couple/few users
### Outcome
#### What did you expect?
User suggestion list displays multiple users nicely, given that there is enough room for it
#### What happened instead?
User suggestion list height is bigger than the container's (container size is "manually" calculated). This is probably due to the fact that on iOS 16, SwiftUI's `List` component switches from using a `UITableView` to a `UICollectionView`. The implemented calculation matches the expected standard vertical spacing of the previous component.

### Your phone model
Any
### Operating system version
iOS 16.0+
### Application version
develop
### Homeserver
Any
### Will you send logs?
No | defect | user suggestion list height doesn t match container on ios steps to reproduce go to a room type with a pattern that would matches just a couple few users outcome what did you expect user suggestion list displays multiple users nicely given that there is enough room for it what happened instead user suggestion list height is bigger than the container s container size is manually calculated this is probably due to the fact that on ios swiftui s list component switches from using a uitableview to a uicollectionview the implemented calculation matches the expected standard vertical spacing of the previous component your phone model any operating system version ios application version develop homeserver any will you send logs no | 1 |
71,921 | 23,854,486,273 | IssuesEvent | 2022-09-06 21:25:34 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Outdated rooms shouldn't show up in room searches | T-Defect P1 A-Room-Directory Z-Synapse | When searching for this room by clicking on the plus and typing some relevant text (e.g. 'Desktop', or even 'iOS'), the first hit is the outdated version.

| 1.0 | Outdated rooms shouldn't show up in room searches - When searching for this room by clicking on the plus and typing some relevant text (e.g. 'Desktop', or even 'iOS'), the first hit is the outdated version.

| defect | outdated rooms shouldn t show up in room searches when searching for this room by clicking on the plus and typing some relevant text e g desktop or even ios the first hit is the outdated version | 1 |
21,892 | 3,575,858,540 | IssuesEvent | 2016-01-27 17:21:55 | antang/NewCapstoneProject | https://api.github.com/repos/antang/NewCapstoneProject | closed | Defect Edit in "Danh Sách Cuộc Thi" | Defect | Double click in Contest, when you click button "Tiếp" step number will count up, but click button "Lùi" it don't count down, click button "Tiếp" again, step number don't count down too. | 1.0 | Defect Edit in "Danh Sách Cuộc Thi" - Double click in Contest, when you click button "Tiếp" step number will count up, but click button "Lùi" it don't count down, click button "Tiếp" again, step number don't count down too. | defect | defect edit in danh sách cuộc thi double click in contest when you click button tiếp step number will count up but click button lùi it don t count down click button tiếp again step number don t count down too | 1 |
365,354 | 25,531,256,471 | IssuesEvent | 2022-11-29 08:37:18 | MinBZK/regels.overheid.nl | https://api.github.com/repos/MinBZK/regels.overheid.nl | closed | Fine-tune and document and publish agile software development process | documentation current sprint | Describe how we do our development process and document it extensively.
- [ ] Epics
- [ ] Issues
- [ ] Labels
- [ ] Pipeline
- [ ] Sprints
- [ ] Releases | 1.0 | Fine-tune and document and publish agile software development process - Describe how we do our development process and document it extensively.
- [ ] Epics
- [ ] Issues
- [ ] Labels
- [ ] Pipeline
- [ ] Sprints
- [ ] Releases | non_defect | fine tune and document and publish agile software development process describe how we do our development process and document it extensively epics issues labels pipeline sprints releases | 0 |
34,058 | 28,141,520,026 | IssuesEvent | 2023-04-02 01:19:55 | GoodsPartner/GoodsPartner | https://api.github.com/repos/GoodsPartner/GoodsPartner | opened | [CI\CD] Improve deployment pipeline | Investigation infrastructure | 1. Add environment
2. Investigate possibility to occupy port on server without running script under sudo permission | 1.0 | [CI\CD] Improve deployment pipeline - 1. Add environment
2. Investigate possibility to occupy port on server without running script under sudo permission | non_defect | improve deployment pipeline add environment investigate possibility to occupy port on server without running script under sudo permission | 0 |
231,173 | 25,491,012,204 | IssuesEvent | 2022-11-27 03:31:05 | amccool/AM.Extensions.Logging.ElasticSearch | https://api.github.com/repos/amccool/AM.Extensions.Logging.ElasticSearch | reopened | CVE-2018-8292 (High) detected in system.net.http.4.3.0.nupkg | security vulnerability | ## CVE-2018-8292 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers.
</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /src/ElasticLogger.Test/ElasticLogger.Test.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- microsoft.net.test.sdk.15.6.0.nupkg (Root Library)
- microsoft.testplatform.testhost.15.6.0.nupkg
- microsoft.testplatform.objectmodel.15.6.0.nupkg
- netstandard.library.1.6.1.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/amccool/AM.Extensions.Logging.ElasticSearch/commit/95d09cecdcfd2cd0d44becc30d4d21d470049160">95d09cecdcfd2cd0d44becc30d4d21d470049160</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0.
<p>Publish Date: 2018-10-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8292>CVE-2018-8292</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-10-10</p>
<p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-8292 (High) detected in system.net.http.4.3.0.nupkg - ## CVE-2018-8292 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that allow applications to consume web services over HTTP and HTTP components that can be used by both clients and servers for parsing HTTP headers.
</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /src/ElasticLogger.Test/ElasticLogger.Test.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- microsoft.net.test.sdk.15.6.0.nupkg (Root Library)
- microsoft.testplatform.testhost.15.6.0.nupkg
- microsoft.testplatform.objectmodel.15.6.0.nupkg
- netstandard.library.1.6.1.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/amccool/AM.Extensions.Logging.ElasticSearch/commit/95d09cecdcfd2cd0d44becc30d4d21d470049160">95d09cecdcfd2cd0d44becc30d4d21d470049160</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0.
<p>Publish Date: 2018-10-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8292>CVE-2018-8292</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-10-10</p>
<p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in system net http nupkg cve high severity vulnerability vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that allow applications to consume web services over http and http components that can be used by both clients and servers for parsing http headers library home page a href path to dependency file src elasticlogger test elasticlogger test csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg dependency hierarchy microsoft net test sdk nupkg root library microsoft testplatform testhost nupkg microsoft testplatform objectmodel nupkg netstandard library nupkg x system net http nupkg vulnerable library found in head commit a href found in base branch master vulnerability details an information disclosure vulnerability exists in net core when authentication information is inadvertently exposed in a redirect aka net core information disclosure vulnerability this affects net core net core net core powershell core publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution system net http microsoft powershell commands utility rc step up your open source security game with mend | 0 |
1,400 | 2,603,846,955 | IssuesEvent | 2015-02-24 18:16:13 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳病毒疱疹检查 | auto-migrated Priority-Medium Type-Defect | ```
沈阳病毒疱疹检查〓沈陽軍區政治部醫院性病〓TEL:024-3102330
8〓成立于1946年,68年專注于性傳播疾病的研究和治療。位于�
��陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷�
��悠久、設備精良、技術權威、專家云集,是預防、保健、醫
療、科研康復為一體的綜合性醫院。是國家首批公立甲等部��
�醫院、全國首批醫療規范定點單位,是第四軍醫大學、東南�
��學等知名高等院校的教學醫院。曾被中國人民解放軍空軍后
勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二等��
�。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:14 | 1.0 | 沈阳病毒疱疹检查 - ```
沈阳病毒疱疹检查〓沈陽軍區政治部醫院性病〓TEL:024-3102330
8〓成立于1946年,68年專注于性傳播疾病的研究和治療。位于�
��陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷�
��悠久、設備精良、技術權威、專家云集,是預防、保健、醫
療、科研康復為一體的綜合性醫院。是國家首批公立甲等部��
�醫院、全國首批醫療規范定點單位,是第四軍醫大學、東南�
��學等知名高等院校的教學醫院。曾被中國人民解放軍空軍后
勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二等��
�。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:14 | defect | 沈阳病毒疱疹检查 沈阳病毒疱疹检查〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位于� �� 。是一所與新中國同建立共輝煌的歷� ��悠久、設備精良、技術權威、專家云集,是預防、保健、醫 療、科研康復為一體的綜合性醫院。是國家首批公立甲等部�� �醫院、全國首批醫療規范定點單位,是第四軍醫大學、東南� ��學等知名高等院校的教學醫院。曾被中國人民解放軍空軍后 勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二等�� �。 original issue reported on code google com by gmail com on jun at | 1 |
59,881 | 17,023,276,901 | IssuesEvent | 2021-07-03 01:11:12 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | German language graphicbugs | Component: potlatch (flash editor) Priority: major Resolution: duplicate Type: defect | **[Submitted to the original trac issue database at 11.47am, Wednesday, 23rd July 2008]**
In german version if have problems with the buttons and many sentences.
Look at the screenshot:
http://picload.org/image/e87517eddbdf9ee2d2da772df82e8894/bildschirmfoto.png
| 1.0 | German language graphicbugs - **[Submitted to the original trac issue database at 11.47am, Wednesday, 23rd July 2008]**
In german version if have problems with the buttons and many sentences.
Look at the screenshot:
http://picload.org/image/e87517eddbdf9ee2d2da772df82e8894/bildschirmfoto.png
| defect | german language graphicbugs in german version if have problems with the buttons and many sentences look at the screenshot | 1 |
588,794 | 17,671,681,239 | IssuesEvent | 2021-08-23 07:10:21 | status-im/status-desktop | https://api.github.com/repos/status-im/status-desktop | closed | add unfurling support for tenor | Chat feature priority F1: mandatory | the gif selector sends messages with the tenor url, we should add tenor.com to the list, probably needs changes in status-go

marking as mandatory because it's not obvious to the use there is an advanced setting to display any image | 1.0 | add unfurling support for tenor - the gif selector sends messages with the tenor url, we should add tenor.com to the list, probably needs changes in status-go

marking as mandatory because it's not obvious to the use there is an advanced setting to display any image | non_defect | add unfurling support for tenor the gif selector sends messages with the tenor url we should add tenor com to the list probably needs changes in status go marking as mandatory because it s not obvious to the use there is an advanced setting to display any image | 0 |
793,352 | 27,992,092,544 | IssuesEvent | 2023-03-27 05:12:12 | AY2223S2-CS2103T-W12-4/tp | https://api.github.com/repos/AY2223S2-CS2103T-W12-4/tp | closed | Find appointments during a particular time period | priority.High type.Appointment | As a doctor, I want to view the number of appointments I have during a particular time period so that I can gauge my workload. | 1.0 | Find appointments during a particular time period - As a doctor, I want to view the number of appointments I have during a particular time period so that I can gauge my workload. | non_defect | find appointments during a particular time period as a doctor i want to view the number of appointments i have during a particular time period so that i can gauge my workload | 0 |
315,110 | 27,046,469,276 | IssuesEvent | 2023-02-13 10:07:47 | getsentry/sentry-javascript | https://api.github.com/repos/getsentry/sentry-javascript | closed | Improve NextJS integration tests | Status: In Progress Type: Improvement Package: Nextjs Type: Tests | ### Problem Statement
NextJS integration tests are definitely our most flaky set of tests, and they are also much harder to debug since they don't use jest under the hood.
https://sentry.io/organizations/sentry/dashboard/24245/?project=5899451&statsPeriod=3d
<img width="627" alt="image" src="https://user-images.githubusercontent.com/18689448/198138338-743ddc57-74e1-43f4-b6a8-39aac4474d0b.png">
<img width="1140" alt="image" src="https://user-images.githubusercontent.com/18689448/198138395-6a1a1b18-2990-411a-8526-606a546e34e9.png">
We should update these tests!
### Solution Brainstorm
These probably involves a couple of changes:
- [x] Instead of using puppeteer, use playwright like our browser and remix integration tests. This is as puppeteer causes [some problems](https://github.com/puppeteer/puppeteer/issues/7740) for fresh installations, which was ran into by @drguthals [here](https://github.com/getsentry/sentry-javascript/discussions/6052). -> #6718
- [x] Use jest for the node integration tests, like our remix integration tests do. This should make things much more consistent -> #6909
- [ ] Figure out how to make session tests less flaky
- [ ] Rewrite main test script in TS rather than bash
- [ ] Run most recent node version integration test on Windows | 1.0 | Improve NextJS integration tests - ### Problem Statement
NextJS integration tests are definitely our most flaky set of tests, and they are also much harder to debug since they don't use jest under the hood.
https://sentry.io/organizations/sentry/dashboard/24245/?project=5899451&statsPeriod=3d
<img width="627" alt="image" src="https://user-images.githubusercontent.com/18689448/198138338-743ddc57-74e1-43f4-b6a8-39aac4474d0b.png">
<img width="1140" alt="image" src="https://user-images.githubusercontent.com/18689448/198138395-6a1a1b18-2990-411a-8526-606a546e34e9.png">
We should update these tests!
### Solution Brainstorm
These probably involves a couple of changes:
- [x] Instead of using puppeteer, use playwright like our browser and remix integration tests. This is as puppeteer causes [some problems](https://github.com/puppeteer/puppeteer/issues/7740) for fresh installations, which was ran into by @drguthals [here](https://github.com/getsentry/sentry-javascript/discussions/6052). -> #6718
- [x] Use jest for the node integration tests, like our remix integration tests do. This should make things much more consistent -> #6909
- [ ] Figure out how to make session tests less flaky
- [ ] Rewrite main test script in TS rather than bash
- [ ] Run most recent node version integration test on Windows | non_defect | improve nextjs integration tests problem statement nextjs integration tests are definitely our most flaky set of tests and they are also much harder to debug since they don t use jest under the hood img width alt image src img width alt image src we should update these tests solution brainstorm these probably involves a couple of changes instead of using puppeteer use playwright like our browser and remix integration tests this is as puppeteer causes for fresh installations which was ran into by drguthals use jest for the node integration tests like our remix integration tests do this should make things much more consistent figure out how to make session tests less flaky rewrite main test script in ts rather than bash run most recent node version integration test on windows | 0 |
42,687 | 11,218,223,820 | IssuesEvent | 2020-01-07 10:58:11 | zotonic/zotonic | https://api.github.com/repos/zotonic/zotonic | closed | mod_backup: Cannot backup site created from web GUI | defect | [master branch]
When I create a site from the web gui with default values (using settings from ~/.zotonic/1/zotonic.config I cannot make a backup.
```22:22:00.442 [info] z_module_manager:237 [creado2] info @ z_module_manager:237 Module mod_backup activated by 1 ()
(zotonic001@backv1)1> 22:24:34.974 [warning] mod_backup:365 pg_dump: [archiver (db)] connection to database "zotonic" failed: invalid port number: "0"
22:24:34.975 [error] Error in process <0.2422.0> on node zotonic001@backv1 with exit value:
{{badmatch,{error,"pg_dump: [archiver (db)] connection to database \"zotonic\" failed: invalid port number: \"0\"\n"}},[{mod_backup,do_backup_process,3,[{file,"/home/mikael/work/zotonic/apps/zotonic_mod_backup/src/mod_backup.erl"},{line,301}]}]}
| 1.0 | mod_backup: Cannot backup site created from web GUI - [master branch]
When I create a site from the web gui with default values (using settings from ~/.zotonic/1/zotonic.config I cannot make a backup.
```22:22:00.442 [info] z_module_manager:237 [creado2] info @ z_module_manager:237 Module mod_backup activated by 1 ()
(zotonic001@backv1)1> 22:24:34.974 [warning] mod_backup:365 pg_dump: [archiver (db)] connection to database "zotonic" failed: invalid port number: "0"
22:24:34.975 [error] Error in process <0.2422.0> on node zotonic001@backv1 with exit value:
{{badmatch,{error,"pg_dump: [archiver (db)] connection to database \"zotonic\" failed: invalid port number: \"0\"\n"}},[{mod_backup,do_backup_process,3,[{file,"/home/mikael/work/zotonic/apps/zotonic_mod_backup/src/mod_backup.erl"},{line,301}]}]}
| defect | mod backup cannot backup site created from web gui when i create a site from the web gui with default values using settings from zotonic zotonic config i cannot make a backup z module manager info z module manager module mod backup activated by mod backup pg dump connection to database zotonic failed invalid port number error in process on node with exit value badmatch error pg dump connection to database zotonic failed invalid port number n | 1 |
15,831 | 9,104,431,230 | IssuesEvent | 2019-02-20 18:09:18 | tidyverse/dplyr | https://api.github.com/repos/tidyverse/dplyr | closed | group_by + summarise in dplyr v8.0 is so much slower than previous version | performance :rocket: | Simple operations on fairly large data frames now take forever to complete
---
I have a tibble
```r
> pepData
# A tibble: 1,496,130 x 6
id Sequence Modifications Protein Gene Description
<chr> <chr> <chr> <chr> <chr> <chr>
```
This operation used to complete in 2sec
```r
> pepData %>% group_by(Protein) %>% summarise(n=n_distinct(id))
# A tibble: 6,614 x 2
Protein n
<chr> <int>
```
With new dplyr v8 it takes 5min or more. Adding .drop=TRUE to group_by doesn't help.
| True | group_by + summarise in dplyr v8.0 is so much slower than previous version - Simple operations on fairly large data frames now take forever to complete
---
I have a tibble
```r
> pepData
# A tibble: 1,496,130 x 6
id Sequence Modifications Protein Gene Description
<chr> <chr> <chr> <chr> <chr> <chr>
```
This operation used to complete in 2sec
```r
> pepData %>% group_by(Protein) %>% summarise(n=n_distinct(id))
# A tibble: 6,614 x 2
Protein n
<chr> <int>
```
With new dplyr v8 it takes 5min or more. Adding .drop=TRUE to group_by doesn't help.
| non_defect | group by summarise in dplyr is so much slower than previous version simple operations on fairly large data frames now take forever to complete i have a tibble r pepdata a tibble x id sequence modifications protein gene description this operation used to complete in r pepdata group by protein summarise n n distinct id a tibble x protein n with new dplyr it takes or more adding drop true to group by doesn t help | 0 |
39,019 | 19,672,814,658 | IssuesEvent | 2022-01-11 09:16:37 | MicrosoftDocs/windowsserverdocs | https://api.github.com/repos/MicrosoftDocs/windowsserverdocs | closed | Blog links on page are dead, please update | Pri2 windows-server/prod performance-tuning-guide/tech | attempting to reference information on performance counters and what they indicate, there are blog links on the page at https://docs.microsoft.com/en-us/windows-server/administration/performance-tuning/role/file-server/smb-file-server but they are dead links.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7eff7ae9-34b9-5047-b707-50bbd05dbe5a
* Version Independent ID: 8e6ae745-afb9-89d0-77ad-3a9059aac122
* Content: [Performance Tuning for SMB File Servers](https://docs.microsoft.com/en-us/windows-server/administration/performance-tuning/role/file-server/smb-file-server#feedback)
* Content Source: [WindowsServerDocs/administration/performance-tuning/role/file-server/smb-file-server.md](https://github.com/MicrosoftDocs/windowsserverdocs/blob/master/WindowsServerDocs/administration/performance-tuning/role/file-server/smb-file-server.md)
* Product: **windows-server**
* Technology: **performance-tuning-guide**
* GitHub Login: @phstee
* Microsoft Alias: **NedPyle; Danlo; DKruse** | True | Blog links on page are dead, please update - attempting to reference information on performance counters and what they indicate, there are blog links on the page at https://docs.microsoft.com/en-us/windows-server/administration/performance-tuning/role/file-server/smb-file-server but they are dead links.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7eff7ae9-34b9-5047-b707-50bbd05dbe5a
* Version Independent ID: 8e6ae745-afb9-89d0-77ad-3a9059aac122
* Content: [Performance Tuning for SMB File Servers](https://docs.microsoft.com/en-us/windows-server/administration/performance-tuning/role/file-server/smb-file-server#feedback)
* Content Source: [WindowsServerDocs/administration/performance-tuning/role/file-server/smb-file-server.md](https://github.com/MicrosoftDocs/windowsserverdocs/blob/master/WindowsServerDocs/administration/performance-tuning/role/file-server/smb-file-server.md)
* Product: **windows-server**
* Technology: **performance-tuning-guide**
* GitHub Login: @phstee
* Microsoft Alias: **NedPyle; Danlo; DKruse** | non_defect | blog links on page are dead please update attempting to reference information on performance counters and what they indicate there are blog links on the page at but they are dead links document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product windows server technology performance tuning guide github login phstee microsoft alias nedpyle danlo dkruse | 0 |
500,857 | 14,516,299,395 | IssuesEvent | 2020-12-13 15:28:47 | Zettlr/Zettlr | https://api.github.com/repos/Zettlr/Zettlr | closed | When linking files, ID is duplicated | bug [non-critical] pinned priority:mid | ## Description
<!-- Below, please describe what the Bug does in one or two short sentences. -->
My files are created like this: "ID Name.md". For example:"20201121220427 Menos es mas.md" .
In Preferences/Zettelkasten "When linking files, add the filename...", "always" option is selected.
<!-- Below, please describe what Zettlr should do instead -->
Usually, when creating an internal link, suggestion list shows files as "20201121220427 Menos es mas"

and created link shows the same

Now I'm getting this behaviour

and created link shows repeated the ID

## Reproducing
<!-- Below, describe the steps necessary to reproduce the unwanted behaviour -->
## Expected behaviour
<!-- If helpful, add Screenshots to further narrow down the issue -->
<!-- Please provide information on your operating system -->
## Platform
- OS and version: Windows 10
- Zettlr Version: 1.8.0
## Additional information
<!-- If there is anything else that might be of interest, please provide it here -->
| 1.0 | When linking files, ID is duplicated - ## Description
<!-- Below, please describe what the Bug does in one or two short sentences. -->
My files are created like this: "ID Name.md". For example:"20201121220427 Menos es mas.md" .
In Preferences/Zettelkasten "When linking files, add the filename...", "always" option is selected.
<!-- Below, please describe what Zettlr should do instead -->
Usually, when creating an internal link, suggestion list shows files as "20201121220427 Menos es mas"

and created link shows the same

Now I'm getting this behaviour

and created link shows repeated the ID

## Reproducing
<!-- Below, describe the steps necessary to reproduce the unwanted behaviour -->
## Expected behaviour
<!-- If helpful, add Screenshots to further narrow down the issue -->
<!-- Please provide information on your operating system -->
## Platform
- OS and version: Windows 10
- Zettlr Version: 1.8.0
## Additional information
<!-- If there is anything else that might be of interest, please provide it here -->
| non_defect | when linking files id is duplicated description my files are created like this id name md for example menos es mas md in preferences zettelkasten when linking files add the filename always option is selected usually when creating an internal link suggestion list shows files as menos es mas and created link shows the same now i m getting this behaviour and created link shows repeated the id reproducing expected behaviour platform os and version windows zettlr version additional information | 0 |
5,478 | 2,610,188,446 | IssuesEvent | 2015-02-26 18:59:42 | chrsmith/quchuseban | https://api.github.com/repos/chrsmith/quchuseban | opened | 资讯如何自然祛色斑 | auto-migrated Priority-Medium Type-Defect | ```
《摘要》
我不放弃依恋这样的冬日,喧闹的幸福,似乎已经遥远。曾��
�的她,至今已不再出现。我依旧不舍与这样的季节挥手。记�
��里,伤心的她,比这个季节更冰凉。独守破败的窗台,幻想
着她的经过。却迎来冬日暖阳,如昨日雪花,静静飘落,一��
�片轻轻柔柔的。一缕白烟,在我眼前升起,那是手中绿茶的�
��吸,也这样般轻柔。再美好也是曾经,脸上的痕迹提醒着我
,我们不会在相聚!如何自然祛色斑,
《客户案例》
不用说,长斑肯定是咱们女人最讨厌的事情,就算再不��
�美的女人,脸上长了斑就是嘴上不说,心里也肯定是难过的�
��在意的,我就是这样,其实是不想让老公和关心我的人担心
,虽然总安慰着他们说没事,没关系,心里多多少少还会想��
�底要怎么样能给去掉。<br>
我也不知道自己怎么长的斑,奇怪的很,说是内分泌失��
�,可造成这些的原因我也没有啊,好像凭空里冒出来似的,�
��初也就是那么几颗,就在鼻子旁边,开始也看到了,照镜子
的是时候还仔细的看了看,老公说我是精神紧张过度,就没��
�么在意了,过了好一段时间,我这人平时也不怎么爱照镜子�
��而且也不怎么化妆,那天老公忽然就盯着我的脸说,亲爱的
,你的脸中什么奖了。我还觉得奇怪呢,他就把镜子拿给我��
�,我一看可吓住了,脸颊什么时候也多了那么斑点,老公赶�
��安慰我说,是不是太忙了?虽然笑笑说没事,可心里还是更��
�过了,女为悦己者容吗?我就是想美美的,谁让它长这些斑了
,我就在心里暗下决心,不行,我一定要把这些斑给去掉,��
�初,我上网查了查,说是精华比较好,我就专门去查了下精�
��的产品,不知道怎么的就看到一个叫「黛芙薇尔精华液」的
商城了,正好去仔细的瞧了瞧,居然有很多人都好了,我又��
�了她们的产品成分,觉得还好,说是国际品牌的,我又仔细�
��去查了,就是查到了,觉得应该不错,至少有保证吧,要不
然怎么会有那么多人就好了,心里这么想着,就在商城上订��
�,总得试试啊,要不然怎么也不会知道结果啊,使用了两个�
��期,才发现斑明显的淡了,心情一下子真的好多了,老公还
开玩笑说,家里的低气压总算过去了,呵呵,「黛芙薇尔精��
�液」还真是最好的美白祛斑产品。
阅读了如何自然祛色斑,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
如何自然祛色斑,同时为您分享祛斑小方法
1、百合:兰州特产,含多种氨基酸、维生素、蛋白质、铁、�
��微量元素。百合花粉、百合块茎,均可作组分配伍。其味甘
性平,消心安神、润肺止咳、补中益气。
2、红花:活血、消肿通经,入心肝经,外用可使面色红艳。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:28 | 1.0 | 资讯如何自然祛色斑 - ```
《摘要》
我不放弃依恋这样的冬日,喧闹的幸福,似乎已经遥远。曾��
�的她,至今已不再出现。我依旧不舍与这样的季节挥手。记�
��里,伤心的她,比这个季节更冰凉。独守破败的窗台,幻想
着她的经过。却迎来冬日暖阳,如昨日雪花,静静飘落,一��
�片轻轻柔柔的。一缕白烟,在我眼前升起,那是手中绿茶的�
��吸,也这样般轻柔。再美好也是曾经,脸上的痕迹提醒着我
,我们不会在相聚!如何自然祛色斑,
《客户案例》
不用说,长斑肯定是咱们女人最讨厌的事情,就算再不��
�美的女人,脸上长了斑就是嘴上不说,心里也肯定是难过的�
��在意的,我就是这样,其实是不想让老公和关心我的人担心
,虽然总安慰着他们说没事,没关系,心里多多少少还会想��
�底要怎么样能给去掉。<br>
我也不知道自己怎么长的斑,奇怪的很,说是内分泌失��
�,可造成这些的原因我也没有啊,好像凭空里冒出来似的,�
��初也就是那么几颗,就在鼻子旁边,开始也看到了,照镜子
的是时候还仔细的看了看,老公说我是精神紧张过度,就没��
�么在意了,过了好一段时间,我这人平时也不怎么爱照镜子�
��而且也不怎么化妆,那天老公忽然就盯着我的脸说,亲爱的
,你的脸中什么奖了。我还觉得奇怪呢,他就把镜子拿给我��
�,我一看可吓住了,脸颊什么时候也多了那么斑点,老公赶�
��安慰我说,是不是太忙了?虽然笑笑说没事,可心里还是更��
�过了,女为悦己者容吗?我就是想美美的,谁让它长这些斑了
,我就在心里暗下决心,不行,我一定要把这些斑给去掉,��
�初,我上网查了查,说是精华比较好,我就专门去查了下精�
��的产品,不知道怎么的就看到一个叫「黛芙薇尔精华液」的
商城了,正好去仔细的瞧了瞧,居然有很多人都好了,我又��
�了她们的产品成分,觉得还好,说是国际品牌的,我又仔细�
��去查了,就是查到了,觉得应该不错,至少有保证吧,要不
然怎么会有那么多人就好了,心里这么想着,就在商城上订��
�,总得试试啊,要不然怎么也不会知道结果啊,使用了两个�
��期,才发现斑明显的淡了,心情一下子真的好多了,老公还
开玩笑说,家里的低气压总算过去了,呵呵,「黛芙薇尔精��
�液」还真是最好的美白祛斑产品。
阅读了如何自然祛色斑,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
如何自然祛色斑,同时为您分享祛斑小方法
1、百合:兰州特产,含多种氨基酸、维生素、蛋白质、铁、�
��微量元素。百合花粉、百合块茎,均可作组分配伍。其味甘
性平,消心安神、润肺止咳、补中益气。
2、红花:活血、消肿通经,入心肝经,外用可使面色红艳。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:28 | defect | 资讯如何自然祛色斑 《摘要》 我不放弃依恋这样的冬日,喧闹的幸福,似乎已经遥远。曾�� �的她,至今已不再出现。我依旧不舍与这样的季节挥手。记� ��里,伤心的她,比这个季节更冰凉。独守破败的窗台,幻想 着她的经过。却迎来冬日暖阳,如昨日雪花,静静飘落,一�� �片轻轻柔柔的。一缕白烟,在我眼前升起,那是手中绿茶的� ��吸,也这样般轻柔。再美好也是曾经,脸上的痕迹提醒着我 ,我们不会在相聚!如何自然祛色斑, 《客户案例》 不用说,长斑肯定是咱们女人最讨厌的事情,就算再不�� �美的女人,脸上长了斑就是嘴上不说,心里也肯定是难过的� ��在意的,我就是这样,其实是不想让老公和关心我的人担心 ,虽然总安慰着他们说没事,没关系,心里多多少少还会想�� �底要怎么样能给去掉。 我也不知道自己怎么长的斑,奇怪的很,说是内分泌失�� �,可造成这些的原因我也没有啊,好像凭空里冒出来似的,� ��初也就是那么几颗,就在鼻子旁边,开始也看到了,照镜子 的是时候还仔细的看了看,老公说我是精神紧张过度,就没�� �么在意了,过了好一段时间,我这人平时也不怎么爱照镜子� ��而且也不怎么化妆,那天老公忽然就盯着我的脸说,亲爱的 ,你的脸中什么奖了。我还觉得奇怪呢,他就把镜子拿给我�� �,我一看可吓住了,脸颊什么时候也多了那么斑点,老公赶� ��安慰我说,是不是太忙了 虽然笑笑说没事,可心里还是更�� �过了,女为悦己者容吗 我就是想美美的,谁让它长这些斑了 ,我就在心里暗下决心,不行,我一定要把这些斑给去掉,�� �初,我上网查了查,说是精华比较好,我就专门去查了下精� ��的产品,不知道怎么的就看到一个叫「黛芙薇尔精华液」的 商城了,正好去仔细的瞧了瞧,居然有很多人都好了,我又�� �了她们的产品成分,觉得还好,说是国际品牌的,我又仔细� ��去查了,就是查到了,觉得应该不错,至少有保证吧,要不 然怎么会有那么多人就好了,心里这么想着,就在商城上订�� �,总得试试啊,要不然怎么也不会知道结果啊,使用了两个� ��期,才发现斑明显的淡了,心情一下子真的好多了,老公还 开玩笑说,家里的低气压总算过去了,呵呵,「黛芙薇尔精�� �液」还真是最好的美白祛斑产品。 阅读了如何自然祛色斑,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 如何自然祛色斑,同时为您分享祛斑小方法 、百合:兰州特产,含多种氨基酸、维生素、蛋白质、铁、� ��微量元素。百合花粉、百合块茎,均可作组分配伍。其味甘 性平,消心安神、润肺止咳、补中益气。 、红花:活血、消肿通经,入心肝经,外用可使面色红艳。 original issue reported on code google com by additive gmail com on jul at | 1 |
9,965 | 2,616,015,538 | IssuesEvent | 2015-03-02 00:58:03 | jasonhall/bwapi | https://api.github.com/repos/jasonhall/bwapi | closed | Tau Cross analyzing, and some other UMS. | auto-migrated Component-Logic Priority-Critical Type-Defect Usability | ```
Exception Address:
Exception Type:
Starcraft error log (See Starcraft\Errors) or debug information:
Not sure what these mean, but i have the screen shot:
http://i.imgur.com/q1Msu.png and error log attached.
What version of the product are you using? On what operating system?
BWAPI 3.4, Win7 64 bit.
```
Original issue reported on code.google.com by `hirsc...@uci.edu` on 13 Dec 2010 at 8:18
Attachments:
* [David160101.ERR](https://storage.googleapis.com/google-code-attachments/bwapi/issue-332/comment-0/David160101.ERR)
| 1.0 | Tau Cross analyzing, and some other UMS. - ```
Exception Address:
Exception Type:
Starcraft error log (See Starcraft\Errors) or debug information:
Not sure what these mean, but i have the screen shot:
http://i.imgur.com/q1Msu.png and error log attached.
What version of the product are you using? On what operating system?
BWAPI 3.4, Win7 64 bit.
```
Original issue reported on code.google.com by `hirsc...@uci.edu` on 13 Dec 2010 at 8:18
Attachments:
* [David160101.ERR](https://storage.googleapis.com/google-code-attachments/bwapi/issue-332/comment-0/David160101.ERR)
| defect | tau cross analyzing and some other ums exception address exception type starcraft error log see starcraft errors or debug information not sure what these mean but i have the screen shot and error log attached what version of the product are you using on what operating system bwapi bit original issue reported on code google com by hirsc uci edu on dec at attachments | 1 |
74,476 | 25,138,433,814 | IssuesEvent | 2022-11-09 20:43:29 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | closed | p-calendar | maxDate & minDate are broken in month-picker | defect | basically, the p-calendar when using the month-picker, changes the years, and it has the maxDate-minDate boundaries, but it doesn't apply the p-disabled class to the months
StackBlitz with reproducing the issue: [https://stackblitz.com/edit/primeng-tablescroll-demo-wdclen?file=src%2Fapp%2Fapp.component.ts](url) | 1.0 | p-calendar | maxDate & minDate are broken in month-picker - basically, the p-calendar when using the month-picker, changes the years, and it has the maxDate-minDate boundaries, but it doesn't apply the p-disabled class to the months
StackBlitz with reproducing the issue: [https://stackblitz.com/edit/primeng-tablescroll-demo-wdclen?file=src%2Fapp%2Fapp.component.ts](url) | defect | p calendar maxdate mindate are broken in month picker basically the p calendar when using the month picker changes the years and it has the maxdate mindate boundaries but it doesn t apply the p disabled class to the months stackblitz with reproducing the issue url | 1 |
119,536 | 12,034,075,155 | IssuesEvent | 2020-04-13 15:23:30 | Chadori/Construct-Master-Collection | https://api.github.com/repos/Chadori/Construct-Master-Collection | closed | Mobile IronSource Collection - New 6.16.0 (latest) features. | Documentation Enhancement Mobile Monetization Native Cordova | ## Request Type
**Feature Request**
## Information
Mobile IronSource Collection - New **6.16.0** (latest) features.
Updated the **Native Cordova** plugins for the new IronSource Android SDK (**6.16.0**).
Changes:
```
implementation 'com.ironsource.sdk:mediationsdk:+'
```
And, I've also added these:
```
implementation 'com.google.android.gms:play-services-ads-identifier:+'
implementation 'com.google.android.gms:play-services-basement:+'
```
## Notable features
- **Google Admob Ads** is no longer required to retrieve the `advertisingid`. Hence, **User ID** is automatically filled, even without the **Google Admob Ads** addon.
- Website documentation are now updated.
- Website projects are now updated.
- New **6.16.0** features.
- Reference: https://developers.ironsrc.com/ironsource-mobile/android/sdk-change-log/
## Construct Editor
- Construct 3
## Construct Runtime
- Construct 3
| 1.0 | Mobile IronSource Collection - New 6.16.0 (latest) features. - ## Request Type
**Feature Request**
## Information
Mobile IronSource Collection - New **6.16.0** (latest) features.
Updated the **Native Cordova** plugins for the new IronSource Android SDK (**6.16.0**).
Changes:
```
implementation 'com.ironsource.sdk:mediationsdk:+'
```
And, I've also added these:
```
implementation 'com.google.android.gms:play-services-ads-identifier:+'
implementation 'com.google.android.gms:play-services-basement:+'
```
## Notable features
- **Google Admob Ads** is no longer required to retrieve the `advertisingid`. Hence, **User ID** is automatically filled, even without the **Google Admob Ads** addon.
- Website documentation are now updated.
- Website projects are now updated.
- New **6.16.0** features.
- Reference: https://developers.ironsrc.com/ironsource-mobile/android/sdk-change-log/
## Construct Editor
- Construct 3
## Construct Runtime
- Construct 3
| non_defect | mobile ironsource collection new latest features request type feature request information mobile ironsource collection new latest features updated the native cordova plugins for the new ironsource android sdk changes implementation com ironsource sdk mediationsdk and i ve also added these implementation com google android gms play services ads identifier implementation com google android gms play services basement notable features google admob ads is no longer required to retrieve the advertisingid hence user id is automatically filled even without the google admob ads addon website documentation are now updated website projects are now updated new features reference construct editor construct construct runtime construct | 0 |
425 | 7,880,242,148 | IssuesEvent | 2018-06-26 15:24:50 | Automattic/wp-calypso | https://api.github.com/repos/Automattic/wp-calypso | opened | People: When viewing/editing person, show details of how they were added/invited to site | People Management [Type] Enhancement | It would be useful when viewing/editing a person at `people/edit` to show details of how that person was added/invited to the site originally. Basically, show the information that is currently shown for Invites details at `people/invites`.
<img width="749" alt="screencapture at tue jun 26 11 17 55 edt 2018" src="https://user-images.githubusercontent.com/2098816/41922368-eefe0e00-7932-11e8-9c2e-42c637330527.png">
Internal discussion: p58i-75d-p2 | 1.0 | People: When viewing/editing person, show details of how they were added/invited to site - It would be useful when viewing/editing a person at `people/edit` to show details of how that person was added/invited to the site originally. Basically, show the information that is currently shown for Invites details at `people/invites`.
<img width="749" alt="screencapture at tue jun 26 11 17 55 edt 2018" src="https://user-images.githubusercontent.com/2098816/41922368-eefe0e00-7932-11e8-9c2e-42c637330527.png">
Internal discussion: p58i-75d-p2 | non_defect | people when viewing editing person show details of how they were added invited to site it would be useful when viewing editing a person at people edit to show details of how that person was added invited to the site originally basically show the information that is currently shown for invites details at people invites img width alt screencapture at tue jun edt src internal discussion | 0 |
6,418 | 5,415,608,957 | IssuesEvent | 2017-03-01 22:01:19 | twilio/twilio-video.js | https://api.github.com/repos/twilio/twilio-video.js | closed | Video freezes or browser crash with 3 or more participants | performance question | Hello,
We have been trying to use Twilio Video conference feature for one of our application.
We are facing few of the below mentioned issues in our application as well as in sample demo app provided by Twilio deployed on Heroku
One or more participant video stream freezes in ongoing video session randomly
One or more participant browser crashes randomly showing blank window. Once blank window pops up we are not even able to right click in window.
This issue happens if we have 3 or more participants joining video call.
Issue can be replicated in demo app deployed in Heroku
URL: https://twilio-demo-1.herokuapp.com/
Organizer can generate three sample access tokens in Twilio account and invite two participants to join the call using identifier of access tokens
Can somebody please help to resolve this issue.
Thanks,
| True | Video freezes or browser crash with 3 or more participants - Hello,
We have been trying to use Twilio Video conference feature for one of our application.
We are facing few of the below mentioned issues in our application as well as in sample demo app provided by Twilio deployed on Heroku
One or more participant video stream freezes in ongoing video session randomly
One or more participant browser crashes randomly showing blank window. Once blank window pops up we are not even able to right click in window.
This issue happens if we have 3 or more participants joining video call.
Issue can be replicated in demo app deployed in Heroku
URL: https://twilio-demo-1.herokuapp.com/
Organizer can generate three sample access tokens in Twilio account and invite two participants to join the call using identifier of access tokens
Can somebody please help to resolve this issue.
Thanks,
| non_defect | video freezes or browser crash with or more participants hello we have been trying to use twilio video conference feature for one of our application we are facing few of the below mentioned issues in our application as well as in sample demo app provided by twilio deployed on heroku one or more participant video stream freezes in ongoing video session randomly one or more participant browser crashes randomly showing blank window once blank window pops up we are not even able to right click in window this issue happens if we have or more participants joining video call issue can be replicated in demo app deployed in heroku url organizer can generate three sample access tokens in twilio account and invite two participants to join the call using identifier of access tokens can somebody please help to resolve this issue thanks | 0 |
80,772 | 30,523,532,395 | IssuesEvent | 2023-07-19 09:39:51 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Matthew not able to see my messages | T-Defect X-Needs-Info | ### Steps to reproduce
1. Where are you starting? What can you see?
2. What do you click?
3. More steps…
### Outcome
#### What did you expect?
#### What happened instead?
### Operating system
MacOs
### Application version
Version 1.11.31 (1.11.31)
### How did you install the app?
i dont know
### Homeserver
https://matrix-client.matrix.org
### Will you send logs?
Yes | 1.0 | Matthew not able to see my messages - ### Steps to reproduce
1. Where are you starting? What can you see?
2. What do you click?
3. More steps…
### Outcome
#### What did you expect?
#### What happened instead?
### Operating system
MacOs
### Application version
Version 1.11.31 (1.11.31)
### How did you install the app?
i dont know
### Homeserver
https://matrix-client.matrix.org
### Will you send logs?
Yes | defect | matthew not able to see my messages steps to reproduce where are you starting what can you see what do you click more steps… outcome what did you expect what happened instead operating system macos application version version how did you install the app i dont know homeserver will you send logs yes | 1 |
31,971 | 7,471,212,509 | IssuesEvent | 2018-04-03 08:32:19 | Microsoft/vscode-arduino | https://api.github.com/repos/Microsoft/vscode-arduino | closed | Add support to multi-root workspaces | code ready duplicate feature-request | Please add support to multi-root workspaces as explained [here](https://code.visualstudio.com/docs/editor/multi-root-workspaces).
If i enter some root project the arduino.json should be loaded with board and port information for instance.
Thank you, nuno407 | 1.0 | Add support to multi-root workspaces - Please add support to multi-root workspaces as explained [here](https://code.visualstudio.com/docs/editor/multi-root-workspaces).
If i enter some root project the arduino.json should be loaded with board and port information for instance.
Thank you, nuno407 | non_defect | add support to multi root workspaces please add support to multi root workspaces as explained if i enter some root project the arduino json should be loaded with board and port information for instance thank you | 0 |
100,755 | 16,490,391,283 | IssuesEvent | 2021-05-25 02:15:37 | jinuem/Parse-SDK-JS | https://api.github.com/repos/jinuem/Parse-SDK-JS | opened | CVE-2020-7598 (Medium) detected in multiple libraries | security vulnerability | ## CVE-2020-7598 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-0.0.10.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary>
<p>
<details><summary><b>minimist-0.0.8.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p>
<p>Path to dependency file: /Parse-SDK-JS/package.json</p>
<p>Path to vulnerable library: Parse-SDK-JS/node_modules/mkdirp/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- jest-cli-0.5.10.tgz (Root Library)
- mkdirp-0.5.1.tgz
- :x: **minimist-0.0.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimist-0.0.10.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.10.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.10.tgz</a></p>
<p>Path to dependency file: /Parse-SDK-JS/package.json</p>
<p>Path to vulnerable library: Parse-SDK-JS/node_modules/optimist/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- jest-cli-0.5.10.tgz (Root Library)
- optimist-0.6.1.tgz
- :x: **minimist-0.0.10.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimist-1.2.0.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p>
<p>Path to dependency file: /Parse-SDK-JS/package.json</p>
<p>Path to vulnerable library: Parse-SDK-JS/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- gulp-3.9.1.tgz (Root Library)
- :x: **minimist-1.2.0.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload.
<p>Publish Date: 2020-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94">https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94</a></p>
<p>Release Date: 2020-03-11</p>
<p>Fix Resolution: minimist - 0.2.1,1.2.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7598 (Medium) detected in multiple libraries - ## CVE-2020-7598 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-0.0.10.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary>
<p>
<details><summary><b>minimist-0.0.8.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p>
<p>Path to dependency file: /Parse-SDK-JS/package.json</p>
<p>Path to vulnerable library: Parse-SDK-JS/node_modules/mkdirp/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- jest-cli-0.5.10.tgz (Root Library)
- mkdirp-0.5.1.tgz
- :x: **minimist-0.0.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimist-0.0.10.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.10.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.10.tgz</a></p>
<p>Path to dependency file: /Parse-SDK-JS/package.json</p>
<p>Path to vulnerable library: Parse-SDK-JS/node_modules/optimist/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- jest-cli-0.5.10.tgz (Root Library)
- optimist-0.6.1.tgz
- :x: **minimist-0.0.10.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimist-1.2.0.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p>
<p>Path to dependency file: /Parse-SDK-JS/package.json</p>
<p>Path to vulnerable library: Parse-SDK-JS/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- gulp-3.9.1.tgz (Root Library)
- :x: **minimist-1.2.0.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload.
<p>Publish Date: 2020-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94">https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94</a></p>
<p>Release Date: 2020-03-11</p>
<p>Fix Resolution: minimist - 0.2.1,1.2.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries minimist tgz minimist tgz minimist tgz minimist tgz parse argument options library home page a href path to dependency file parse sdk js package json path to vulnerable library parse sdk js node modules mkdirp node modules minimist package json dependency hierarchy jest cli tgz root library mkdirp tgz x minimist tgz vulnerable library minimist tgz parse argument options library home page a href path to dependency file parse sdk js package json path to vulnerable library parse sdk js node modules optimist node modules minimist package json dependency hierarchy jest cli tgz root library optimist tgz x minimist tgz vulnerable library minimist tgz parse argument options library home page a href path to dependency file parse sdk js package json path to vulnerable library parse sdk js node modules minimist package json dependency hierarchy gulp tgz root library x minimist tgz vulnerable library vulnerability details minimist before could be tricked into adding or modifying properties of object prototype using a constructor or proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution minimist step up your open source security game with whitesource | 0 |
89,200 | 25,601,992,129 | IssuesEvent | 2022-12-01 21:07:04 | NixOS/nixpkgs | https://api.github.com/repos/NixOS/nixpkgs | closed | Baserow is broken | 0.kind: build failure | ### Steps To Reproduce
Steps to reproduce the behavior:
1. build `baserow`
2. it will err out with missing dependency: `uvloop`
### Build log
```
[...]
Finished executing setuptoolsBuildPhase
installing
Executing pipInstallPhase
/build/source/backend/dist /build/source/backend
Processing ./baserow-1.10.2-py3-none-any.whl
ERROR: Could not find a version that satisfies the requirement uvloop (from baserow) (from versions: none)
ERROR: No matching distribution found for uvloop
```
### Additional context
It seems I am building the 1.10.2, but it was initialized at 1.10.1 in 6323269208f036bc6c4140fc72251f8505bf8073 and I do not find the upgrade commit, I do not understand if it was a mistake.
### Notify maintainers
@onny
### Metadata
Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result.
```console
❯ nsp nix-info --run "nix-info -m"
- system: `"x86_64-linux"`
- host os: `Linux 5.15.58, NixOS, 22.11 (Raccoon), 22.11pre402830.5e804cd8a27`
- multi-user?: `yes`
- sandbox: `yes`
- version: `nix-env (Nix) 2.10.3`
- channels(root): `"home-manager, nixos, sops-nix"`
- channels(raito): `"home-manager, nixgl, nixpkgs-21.11pre319254.b5182c214fa"`
- nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos`
``` | 1.0 | Baserow is broken - ### Steps To Reproduce
Steps to reproduce the behavior:
1. build `baserow`
2. it will err out with missing dependency: `uvloop`
### Build log
```
[...]
Finished executing setuptoolsBuildPhase
installing
Executing pipInstallPhase
/build/source/backend/dist /build/source/backend
Processing ./baserow-1.10.2-py3-none-any.whl
ERROR: Could not find a version that satisfies the requirement uvloop (from baserow) (from versions: none)
ERROR: No matching distribution found for uvloop
```
### Additional context
It seems I am building the 1.10.2, but it was initialized at 1.10.1 in 6323269208f036bc6c4140fc72251f8505bf8073 and I do not find the upgrade commit, I do not understand if it was a mistake.
### Notify maintainers
@onny
### Metadata
Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result.
```console
❯ nsp nix-info --run "nix-info -m"
- system: `"x86_64-linux"`
- host os: `Linux 5.15.58, NixOS, 22.11 (Raccoon), 22.11pre402830.5e804cd8a27`
- multi-user?: `yes`
- sandbox: `yes`
- version: `nix-env (Nix) 2.10.3`
- channels(root): `"home-manager, nixos, sops-nix"`
- channels(raito): `"home-manager, nixgl, nixpkgs-21.11pre319254.b5182c214fa"`
- nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos`
``` | non_defect | baserow is broken steps to reproduce steps to reproduce the behavior build baserow it will err out with missing dependency uvloop build log finished executing setuptoolsbuildphase installing executing pipinstallphase build source backend dist build source backend processing baserow none any whl error could not find a version that satisfies the requirement uvloop from baserow from versions none error no matching distribution found for uvloop additional context it seems i am building the but it was initialized at in and i do not find the upgrade commit i do not understand if it was a mistake notify maintainers onny metadata please run nix shell p nix info run nix info m and paste the result console ❯ nsp nix info run nix info m system linux host os linux nixos raccoon multi user yes sandbox yes version nix env nix channels root home manager nixos sops nix channels raito home manager nixgl nixpkgs nixpkgs nix var nix profiles per user root channels nixos | 0 |
56,949 | 6,535,001,655 | IssuesEvent | 2017-08-31 13:10:27 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | Extended.[k8s.io] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should not deadlock when a pod's predecessor fails | component/kubernetes kind/test-flake priority/P0 | ```
[k8s.io] StatefulSet
/go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:656
[k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
/go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:656
should not deadlock when a pod's predecessor fails [It]
/go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/statefulset.go:249
Jul 22 15:24:24.738: Failed waiting for pods to enter running: timed out waiting for the condition
/go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/statefulset_utils.go:303
```
from https://ci.openshift.redhat.com/jenkins/job/merge_pull_request_origin/1349/
possibly a dupe of
https://github.com/openshift/origin/issues/15408
and
https://github.com/openshift/origin/issues/15401
Note that this was seen on a merge to the 3.6 branch (ie pre-rebase)
| 1.0 | Extended.[k8s.io] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should not deadlock when a pod's predecessor fails - ```
[k8s.io] StatefulSet
/go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:656
[k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
/go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:656
should not deadlock when a pod's predecessor fails [It]
/go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/statefulset.go:249
Jul 22 15:24:24.738: Failed waiting for pods to enter running: timed out waiting for the condition
/go/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/statefulset_utils.go:303
```
from https://ci.openshift.redhat.com/jenkins/job/merge_pull_request_origin/1349/
possibly a dupe of
https://github.com/openshift/origin/issues/15408
and
https://github.com/openshift/origin/issues/15401
Note that this was seen on a merge to the 3.6 branch (ie pre-rebase)
| non_defect | extended statefulset basic statefulset functionality should not deadlock when a pod s predecessor fails statefulset go src github com openshift origin output local go src github com openshift origin vendor io kubernetes test framework framework go basic statefulset functionality go src github com openshift origin output local go src github com openshift origin vendor io kubernetes test framework framework go should not deadlock when a pod s predecessor fails go src github com openshift origin output local go src github com openshift origin vendor io kubernetes test statefulset go jul failed waiting for pods to enter running timed out waiting for the condition go src github com openshift origin output local go src github com openshift origin vendor io kubernetes test framework statefulset utils go from possibly a dupe of and note that this was seen on a merge to the branch ie pre rebase | 0 |
7,401 | 2,610,366,656 | IssuesEvent | 2015-02-26 19:58:26 | chrsmith/scribefire-chrome | https://api.github.com/repos/chrsmith/scribefire-chrome | opened | No Offline save of multiple posts | auto-migrated Priority-Medium Type-Defect | ```
What's the problem?
When I'm offline, I'm not able to save many posts to my computer.
I've made the test many times, and always only the last post is saved.
Probably, if the feature isn't available, a warning message should precise it!
What browser are you using?
Chrome 27.0.1453.110 m
What version of ScribeFire are you running?
4.1
```
-----
Original issue reported on code.google.com by `mathieu....@gmail.com` on 12 Jun 2013 at 9:45 | 1.0 | No Offline save of multiple posts - ```
What's the problem?
When I'm offline, I'm not able to save many posts to my computer.
I've made the test many times, and always only the last post is saved.
Probably, if the feature isn't available, a warning message should precise it!
What browser are you using?
Chrome 27.0.1453.110 m
What version of ScribeFire are you running?
4.1
```
-----
Original issue reported on code.google.com by `mathieu....@gmail.com` on 12 Jun 2013 at 9:45 | defect | no offline save of multiple posts what s the problem when i m offline i m not able to save many posts to my computer i ve made the test many times and always only the last post is saved probably if the feature isn t available a warning message should precise it what browser are you using chrome m what version of scribefire are you running original issue reported on code google com by mathieu gmail com on jun at | 1 |
19,564 | 6,736,355,464 | IssuesEvent | 2017-10-19 03:25:04 | ngageoint/hootenanny | https://api.github.com/repos/ngageoint/hootenanny | closed | Use OpenJDK-1.8 on CentOS 7 | Category: Build Category: Core Category: Install Priority: Medium Type: Maintenance Type: Support | OpenJDK-1.8 is available via default RPM repos on CentOS 7 making using it simple to install. | 1.0 | Use OpenJDK-1.8 on CentOS 7 - OpenJDK-1.8 is available via default RPM repos on CentOS 7 making using it simple to install. | non_defect | use openjdk on centos openjdk is available via default rpm repos on centos making using it simple to install | 0 |
57,831 | 16,092,983,258 | IssuesEvent | 2021-04-26 19:11:19 | primefaces/primereact | https://api.github.com/repos/primefaces/primereact | closed | Treetable break after toggle columns | defect | ### There is no guarantee in receiving an immediate response in GitHub Issue Tracker, If you'd like to secure our response, you may consider *PrimeReact PRO Support* where support is provided within 4 business hours
**I'm submitting a ...** (check one with "x")
```
[ ] bug report
[ ] feature request
[ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57
```
**Codesandbox Case (Bug Reports)**
Please fork the codesandbox below and create a case demonstrating your bug report. Issues without a codesandbox have much less possibility to be reviewed.
https://codesandbox.io/s/friendly-williams-2s6ts
**Current behavior**
1) Drag and drop (swap) two columns (size and type)
2) toggle any column from toggle menulist
3) treetable break with error props of undefined
**Expected behavior**
Treetable should not break after swap two column and toggle it.
**Minimal reproduction of the problem with instructions**
<!--
If the current behavior is a bug or you can illustrate your feature request better with an example,
please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via
https://codesandbox.io or similar (you can use this template as a starting point: https://codesandbox.io/s/qjx332qq4).
-->
**Please tell us about your environment:**
1) chrome 2) windows 3) vscode
* **React version:**
1) v17.0.2
* **PrimeReact version:**
1) v6.3.1
* **Browser:** [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ]
<!-- All browsers where this could be reproduced -->
* **Language:** [all | TypeScript X.X | ES6/7 | ES5]
| 1.0 | Treetable break after toggle columns - ### There is no guarantee in receiving an immediate response in GitHub Issue Tracker, If you'd like to secure our response, you may consider *PrimeReact PRO Support* where support is provided within 4 business hours
**I'm submitting a ...** (check one with "x")
```
[ ] bug report
[ ] feature request
[ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57
```
**Codesandbox Case (Bug Reports)**
Please fork the codesandbox below and create a case demonstrating your bug report. Issues without a codesandbox have much less possibility to be reviewed.
https://codesandbox.io/s/friendly-williams-2s6ts
**Current behavior**
1) Drag and drop (swap) two columns (size and type)
2) toggle any column from toggle menulist
3) treetable break with error props of undefined
**Expected behavior**
Treetable should not break after swap two column and toggle it.
**Minimal reproduction of the problem with instructions**
<!--
If the current behavior is a bug or you can illustrate your feature request better with an example,
please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via
https://codesandbox.io or similar (you can use this template as a starting point: https://codesandbox.io/s/qjx332qq4).
-->
**Please tell us about your environment:**
1) chrome 2) windows 3) vscode
* **React version:**
1) v17.0.2
* **PrimeReact version:**
1) v6.3.1
* **Browser:** [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ]
<!-- All browsers where this could be reproduced -->
* **Language:** [all | TypeScript X.X | ES6/7 | ES5]
| defect | treetable break after toggle columns there is no guarantee in receiving an immediate response in github issue tracker if you d like to secure our response you may consider primereact pro support where support is provided within business hours i m submitting a check one with x bug report feature request support request please do not submit support request here instead see codesandbox case bug reports please fork the codesandbox below and create a case demonstrating your bug report issues without a codesandbox have much less possibility to be reviewed current behavior drag and drop swap two columns size and type toggle any column from toggle menulist treetable break with error props of undefined expected behavior treetable should not break after swap two column and toggle it minimal reproduction of the problem with instructions if the current behavior is a bug or you can illustrate your feature request better with an example please provide the steps to reproduce and if possible a minimal demo of the problem via or similar you can use this template as a starting point please tell us about your environment chrome windows vscode react version primereact version browser language | 1 |
42,709 | 9,301,433,633 | IssuesEvent | 2019-03-23 21:58:28 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | [4.0] System cPanel view requires internet connection | No Code Attached Yet | ### Steps to reproduce the issue
On localhost (eg XAMPP) install some extensions and try to load the "System" view.
Optionally cut off your internet connection and try to load the "System" view.
### Expected result
Loads instantely
### Actual result
Page loads delayed.
### Additional comments
Apparently, the code checks for updates on pageload. If an update source isn't available (due to internet outage, missing source, proxy setting, whatever), the page isn't loaded until the request times out.
This is obviously bad design. There is no reason to check for updates each time I load that page. Or at least it should be done like on the regular cPanel where the check is done using AJAX (so site loading is not delayed). | 1.0 | [4.0] System cPanel view requires internet connection - ### Steps to reproduce the issue
On localhost (eg XAMPP) install some extensions and try to load the "System" view.
Optionally cut off your internet connection and try to load the "System" view.
### Expected result
Loads instantely
### Actual result
Page loads delayed.
### Additional comments
Apparently, the code checks for updates on pageload. If an update source isn't available (due to internet outage, missing source, proxy setting, whatever), the page isn't loaded until the request times out.
This is obviously bad design. There is no reason to check for updates each time I load that page. Or at least it should be done like on the regular cPanel where the check is done using AJAX (so site loading is not delayed). | non_defect | system cpanel view requires internet connection steps to reproduce the issue on localhost eg xampp install some extensions and try to load the system view optionally cut off your internet connection and try to load the system view expected result loads instantely actual result page loads delayed additional comments apparently the code checks for updates on pageload if an update source isn t available due to internet outage missing source proxy setting whatever the page isn t loaded until the request times out this is obviously bad design there is no reason to check for updates each time i load that page or at least it should be done like on the regular cpanel where the check is done using ajax so site loading is not delayed | 0 |
434,661 | 12,521,518,614 | IssuesEvent | 2020-06-03 17:31:35 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | WorkQueueManager timing out when pulling work from the parent queue | BUG Highest Priority WMAgent scalability | **Impact of the bug**
WMAgents
**Describe the bug**
Issue similar to: https://github.com/dmwm/WMCore/issues/9668
and especially: https://github.com/dmwm/WMCore/issues/9669
where, apparently, WorkQueueManager tries to pull all the GQEs sitting in Available status. Which is almost 400k at the moment and almost .5GB of data.
**How to reproduce it**
hundreds of thousands of docs Available in GQ
**Expected behavior**
The obvious one, it shouldn't crash the cycle.
Second, it should not pull all the documents at once (I was pretty sure it did NOT work this way... needs debugging).
And the whole functionality needs to be kept intacted.
**Additional context and error message**
WorkQueueManager traceback from vocms0256:
```
2020-04-30 15:55:28,388:140101196162816:INFO:BaseWorkerThread:WorkQueueManagerLocationPoller took 28.221 secs to execute
2020-04-30 15:56:12,339:140101213210368:ERROR:WorkQueueManagerWorkPoller:Unable to pull work from parent Error: (28, 'Operation timed out after 270000 milliseconds with 421693713 bytes received')
Traceback (most recent call last):
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMComponent/WorkQueueManager/WorkQueueManagerWorkPoller.py", line 101, in pullWork
work = self.queue.pullWork()
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/WorkQueue/WorkQueue.py", line 813, in pullWork
work = self.getAvailableWorkfromParent(resources, jobCounts)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/WorkQueue/WorkQueue.py", line 791, in getAvailableWorkfromParent
work, _, _ = self.parent_queue.availableWork(resources, jobCounts, self.params['Team'], numElems=numElems)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/WorkQueue/WorkQueueBackend.py", line 368, in availableWork
result = self.db.loadList('WorkQueue', 'workRestrictions', 'availableByPriority', options)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/Database/CMSCouch.py", line 570, in loadList
decode=False)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/Services/Requests.py", line 127, in get
encode, decode, contentType)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/Database/CMSCouch.py", line 122, in makeRequest
encode, decode, contentType)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/Services/Requests.py", line 165, in makeRequest
result, response = self.makeRequest_pycurl(uri, data, verb, headers)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/Services/Requests.py", line 183, in makeRequest_pycurl
ckey=ckey, cert=cert, capath=capath)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/Services/pycurl_manager.py", line 257, in request
curl.perform()
error: (28, 'Operation timed out after 270000 milliseconds with 421693713 bytes received')
2020-04-30 15:56:12,339:140101213210368:INFO:WorkQueueManagerWorkPoller:Obtained 0 unit(s) of work
2020-04-30 15:56:12,420:140101213210368:INFO:WorkQueueManagerWorkPoller:Splitting new work
```
| 1.0 | WorkQueueManager timing out when pulling work from the parent queue - **Impact of the bug**
WMAgents
**Describe the bug**
Issue similar to: https://github.com/dmwm/WMCore/issues/9668
and especially: https://github.com/dmwm/WMCore/issues/9669
where, apparently, WorkQueueManager tries to pull all the GQEs sitting in Available status. Which is almost 400k at the moment and almost .5GB of data.
**How to reproduce it**
hundreds of thousands of docs Available in GQ
**Expected behavior**
The obvious one, it shouldn't crash the cycle.
Second, it should not pull all the documents at once (I was pretty sure it did NOT work this way... needs debugging).
And the whole functionality needs to be kept intacted.
**Additional context and error message**
WorkQueueManager traceback from vocms0256:
```
2020-04-30 15:55:28,388:140101196162816:INFO:BaseWorkerThread:WorkQueueManagerLocationPoller took 28.221 secs to execute
2020-04-30 15:56:12,339:140101213210368:ERROR:WorkQueueManagerWorkPoller:Unable to pull work from parent Error: (28, 'Operation timed out after 270000 milliseconds with 421693713 bytes received')
Traceback (most recent call last):
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMComponent/WorkQueueManager/WorkQueueManagerWorkPoller.py", line 101, in pullWork
work = self.queue.pullWork()
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/WorkQueue/WorkQueue.py", line 813, in pullWork
work = self.getAvailableWorkfromParent(resources, jobCounts)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/WorkQueue/WorkQueue.py", line 791, in getAvailableWorkfromParent
work, _, _ = self.parent_queue.availableWork(resources, jobCounts, self.params['Team'], numElems=numElems)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/WorkQueue/WorkQueueBackend.py", line 368, in availableWork
result = self.db.loadList('WorkQueue', 'workRestrictions', 'availableByPriority', options)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/Database/CMSCouch.py", line 570, in loadList
decode=False)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/Services/Requests.py", line 127, in get
encode, decode, contentType)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/Database/CMSCouch.py", line 122, in makeRequest
encode, decode, contentType)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/Services/Requests.py", line 165, in makeRequest
result, response = self.makeRequest_pycurl(uri, data, verb, headers)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/Services/Requests.py", line 183, in makeRequest_pycurl
ckey=ckey, cert=cert, capath=capath)
File "/data/srv/wmagent/v1.3.3/sw/slc7_amd64_gcc630/cms/wmagent/1.3.3/lib/python2.7/site-packages/WMCore/Services/pycurl_manager.py", line 257, in request
curl.perform()
error: (28, 'Operation timed out after 270000 milliseconds with 421693713 bytes received')
2020-04-30 15:56:12,339:140101213210368:INFO:WorkQueueManagerWorkPoller:Obtained 0 unit(s) of work
2020-04-30 15:56:12,420:140101213210368:INFO:WorkQueueManagerWorkPoller:Splitting new work
```
| non_defect | workqueuemanager timing out when pulling work from the parent queue impact of the bug wmagents describe the bug issue similar to and especially where apparently workqueuemanager tries to pull all the gqes sitting in available status which is almost at the moment and almost of data how to reproduce it hundreds of thousands of docs available in gq expected behavior the obvious one it shouldn t crash the cycle second it should not pull all the documents at once i was pretty sure it did not work this way needs debugging and the whole functionality needs to be kept intacted additional context and error message workqueuemanager traceback from info baseworkerthread workqueuemanagerlocationpoller took secs to execute error workqueuemanagerworkpoller unable to pull work from parent error operation timed out after milliseconds with bytes received traceback most recent call last file data srv wmagent sw cms wmagent lib site packages wmcomponent workqueuemanager workqueuemanagerworkpoller py line in pullwork work self queue pullwork file data srv wmagent sw cms wmagent lib site packages wmcore workqueue workqueue py line in pullwork work self getavailableworkfromparent resources jobcounts file data srv wmagent sw cms wmagent lib site packages wmcore workqueue workqueue py line in getavailableworkfromparent work self parent queue availablework resources jobcounts self params numelems numelems file data srv wmagent sw cms wmagent lib site packages wmcore workqueue workqueuebackend py line in availablework result self db loadlist workqueue workrestrictions availablebypriority options file data srv wmagent sw cms wmagent lib site packages wmcore database cmscouch py line in loadlist decode false file data srv wmagent sw cms wmagent lib site packages wmcore services requests py line in get encode decode contenttype file data srv wmagent sw cms wmagent lib site packages wmcore database cmscouch py line in makerequest encode decode contenttype file data srv wmagent sw cms wmagent lib site packages wmcore services requests py line in makerequest result response self makerequest pycurl uri data verb headers file data srv wmagent sw cms wmagent lib site packages wmcore services requests py line in makerequest pycurl ckey ckey cert cert capath capath file data srv wmagent sw cms wmagent lib site packages wmcore services pycurl manager py line in request curl perform error operation timed out after milliseconds with bytes received info workqueuemanagerworkpoller obtained unit s of work info workqueuemanagerworkpoller splitting new work | 0 |
119,466 | 15,554,523,869 | IssuesEvent | 2021-03-16 04:03:17 | UOA-SE701-Group3-2021/3Lancers | https://api.github.com/repos/UOA-SE701-Group3-2021/3Lancers | closed | Hi-fi To-Do List icon design for the widget drawer | approved design | **Is your feature request related to a problem? Please describe.**
As a frontend developer, I want a detailed prototype of the To-Do List I'm implementing, so that I can finalise the styling and logic.
**Describe the solution you'd like**
A hi-fi prototype of the To-Do List icon for the widget drawer.
**Describe alternatives you've considered**
Start coding right away - this would result in conflict further down the road.
Come up with own styling as we develop the code - this would stand out from the UI and reduce cohesiveness.
**Additional context**
The widget drawer is the collection of widgets that can be placed in the journal. The To-Do List is one of these widgets, and the To-Do List icon is one aspect of this To-Do List. | 1.0 | Hi-fi To-Do List icon design for the widget drawer - **Is your feature request related to a problem? Please describe.**
As a frontend developer, I want a detailed prototype of the To-Do List I'm implementing, so that I can finalise the styling and logic.
**Describe the solution you'd like**
A hi-fi prototype of the To-Do List icon for the widget drawer.
**Describe alternatives you've considered**
Start coding right away - this would result in conflict further down the road.
Come up with own styling as we develop the code - this would stand out from the UI and reduce cohesiveness.
**Additional context**
The widget drawer is the collection of widgets that can be placed in the journal. The To-Do List is one of these widgets, and the To-Do List icon is one aspect of this To-Do List. | non_defect | hi fi to do list icon design for the widget drawer is your feature request related to a problem please describe as a frontend developer i want a detailed prototype of the to do list i m implementing so that i can finalise the styling and logic describe the solution you d like a hi fi prototype of the to do list icon for the widget drawer describe alternatives you ve considered start coding right away this would result in conflict further down the road come up with own styling as we develop the code this would stand out from the ui and reduce cohesiveness additional context the widget drawer is the collection of widgets that can be placed in the journal the to do list is one of these widgets and the to do list icon is one aspect of this to do list | 0 |
77,175 | 26,823,066,150 | IssuesEvent | 2023-02-02 10:50:37 | zed-industries/feedback | https://api.github.com/repos/zed-industries/feedback | opened | Clicking an inactive Zed window doesn't focus the split clicked on | defect triage | ### Check for existing issues
- [X] Completed
### Describe the bug / provide steps to reproduce it
Clicking a split of an unfocused Zed window should focus that split.
Consider a Zed window with two splits, a left and right one, and the following steps:
1. The left split is focused
2. Whole Zed window is unfocused
3. Zed window is re-focused by clicking into the right split
Current behaviour: left split is still active, I have to click again into the left split to activate it
Expected behaviour: the right split is focused right away
I mainly notice this when I have a terminal split open and want to insert/paste something into the terminal. I click into the terminal of an unfocused Zed window and start typing/pasting right away. But this often ends up in another editor split that was focused before.
### Environment
Zed: v0.71.3 (stable)
OS: macOS 13.2.0
Memory: 16 GiB
Architecture: aarch64
### If applicable, add mockups / screenshots to help explain present your vision of the feature
Current behaviour in Zed:
https://user-images.githubusercontent.com/409021/216304443-1fa4689d-9db4-4536-b6e2-c76cd985d86d.mov
Expected behaviour using VSCode as an example:
https://user-images.githubusercontent.com/409021/216304561-745af7ee-28ae-4824-b3d1-c6b8854338dd.mov
Ideal solution: IMO the ideal solution would actually be a mix of both. I dislike that VSCode moves the cursor. I'd rather have it to just focus the split. So clicking a split of an unfocused should IMHO focus the split (but not change the cursor's position).
### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue.
If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000.
_No response_ | 1.0 | Clicking an inactive Zed window doesn't focus the split clicked on - ### Check for existing issues
- [X] Completed
### Describe the bug / provide steps to reproduce it
Clicking a split of an unfocused Zed window should focus that split.
Consider a Zed window with two splits, a left and right one, and the following steps:
1. The left split is focused
2. Whole Zed window is unfocused
3. Zed window is re-focused by clicking into the right split
Current behaviour: left split is still active, I have to click again into the left split to activate it
Expected behaviour: the right split is focused right away
I mainly notice this when I have a terminal split open and want to insert/paste something into the terminal. I click into the terminal of an unfocused Zed window and start typing/pasting right away. But this often ends up in another editor split that was focused before.
### Environment
Zed: v0.71.3 (stable)
OS: macOS 13.2.0
Memory: 16 GiB
Architecture: aarch64
### If applicable, add mockups / screenshots to help explain present your vision of the feature
Current behaviour in Zed:
https://user-images.githubusercontent.com/409021/216304443-1fa4689d-9db4-4536-b6e2-c76cd985d86d.mov
Expected behaviour using VSCode as an example:
https://user-images.githubusercontent.com/409021/216304561-745af7ee-28ae-4824-b3d1-c6b8854338dd.mov
Ideal solution: IMO the ideal solution would actually be a mix of both. I dislike that VSCode moves the cursor. I'd rather have it to just focus the split. So clicking a split of an unfocused should IMHO focus the split (but not change the cursor's position).
### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue.
If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000.
_No response_ | defect | clicking an inactive zed window doesn t focus the split clicked on check for existing issues completed describe the bug provide steps to reproduce it clicking a split of an unfocused zed window should focus that split consider a zed window with two splits a left and right one and the following steps the left split is focused whole zed window is unfocused zed window is re focused by clicking into the right split current behaviour left split is still active i have to click again into the left split to activate it expected behaviour the right split is focused right away i mainly notice this when i have a terminal split open and want to insert paste something into the terminal i click into the terminal of an unfocused zed window and start typing pasting right away but this often ends up in another editor split that was focused before environment zed stable os macos memory gib architecture if applicable add mockups screenshots to help explain present your vision of the feature current behaviour in zed expected behaviour using vscode as an example ideal solution imo the ideal solution would actually be a mix of both i dislike that vscode moves the cursor i d rather have it to just focus the split so clicking a split of an unfocused should imho focus the split but not change the cursor s position if applicable attach your library logs zed zed log file to this issue if you only need the most recent lines you can run the zed open log command palette action to see the last no response | 1 |
76,265 | 14,592,413,965 | IssuesEvent | 2020-12-19 17:31:01 | eeeeaaii/vodka | https://api.github.com/repos/eeeeaaii/vodka | closed | code cleanup for EString and EError | code cleanup | EString and EError implemented an old-style "editor flow" that predated me creating the editor class. The code is pretty messy and there is dead code in there. These two classes need to be overhauled and the editing functionality needs to be moved over into a subclass of Editor. | 1.0 | code cleanup for EString and EError - EString and EError implemented an old-style "editor flow" that predated me creating the editor class. The code is pretty messy and there is dead code in there. These two classes need to be overhauled and the editing functionality needs to be moved over into a subclass of Editor. | non_defect | code cleanup for estring and eerror estring and eerror implemented an old style editor flow that predated me creating the editor class the code is pretty messy and there is dead code in there these two classes need to be overhauled and the editing functionality needs to be moved over into a subclass of editor | 0 |
34,570 | 7,457,398,049 | IssuesEvent | 2018-03-30 04:02:59 | kerdokullamae/test_koik_issued | https://api.github.com/repos/kerdokullamae/test_koik_issued | closed | RDF faili puudumisel kuvada korrektne veateade | C: AVAR P: highest R: fixed T: defect | **Reported by sven syld on 26 Mar 2014 14:56 UTC**
'''Kirjeldus'''
Juhul, kui RDF fail on mingil põhjusel tegemata, kuvatakse viga:
''The Response content must be a string or object implementing __toString(), "boolean" given.''
Kontroll võiks olla nt sellises kohas - \Dira\DescriptionUnitBundle\Controller\Service\Rdf\GetController:
```
$file = $gen->getForObject($descriptionUnit, $locale, false);
$data = file_get_contents($file);
```
'''Todo'''
1) Kontrollida, miks ei loodud RDFi uuesti, kuigi oleks pidanud. Arvestada, et ühtegi kataloogi ega midagi serveris tol hetkel ei olnud
2) Kuvada korrektsem veateade, nt "RDF allalaadimisel tekkis viga". | 1.0 | RDF faili puudumisel kuvada korrektne veateade - **Reported by sven syld on 26 Mar 2014 14:56 UTC**
'''Kirjeldus'''
Juhul, kui RDF fail on mingil põhjusel tegemata, kuvatakse viga:
''The Response content must be a string or object implementing __toString(), "boolean" given.''
Kontroll võiks olla nt sellises kohas - \Dira\DescriptionUnitBundle\Controller\Service\Rdf\GetController:
```
$file = $gen->getForObject($descriptionUnit, $locale, false);
$data = file_get_contents($file);
```
'''Todo'''
1) Kontrollida, miks ei loodud RDFi uuesti, kuigi oleks pidanud. Arvestada, et ühtegi kataloogi ega midagi serveris tol hetkel ei olnud
2) Kuvada korrektsem veateade, nt "RDF allalaadimisel tekkis viga". | defect | rdf faili puudumisel kuvada korrektne veateade reported by sven syld on mar utc kirjeldus juhul kui rdf fail on mingil põhjusel tegemata kuvatakse viga the response content must be a string or object implementing tostring boolean given kontroll võiks olla nt sellises kohas dira descriptionunitbundle controller service rdf getcontroller file gen getforobject descriptionunit locale false data file get contents file todo kontrollida miks ei loodud rdfi uuesti kuigi oleks pidanud arvestada et ühtegi kataloogi ega midagi serveris tol hetkel ei olnud kuvada korrektsem veateade nt rdf allalaadimisel tekkis viga | 1 |
32,458 | 13,839,373,782 | IssuesEvent | 2020-10-14 07:52:04 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Replace "plans" concept with "features" | Pri2 azure-communication-services/svc cxp doc-enhancement triaged | I've gotten feedback to refrain from using the term "Plans" in the UI as it may imply an associated cost and/or fixed limits as it does with traditional consumer phone plans. Instead I updated the UI to refer to capabilities as features. I'd suggest we use this same terminology in our docs.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 641aef05-1d16-ba0d-fb48-1527583a860d
* Version Independent ID: 2ca9c55d-56a7-ec67-80da-7c0337316245
* Content: [Quickstart - Get a phone number from Azure Communication Services](https://docs.microsoft.com/en-us/azure/communication-services/quickstarts/telephony-sms/get-phone-number)
* Content Source: [articles/communication-services/quickstarts/telephony-sms/get-phone-number.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/communication-services/quickstarts/telephony-sms/get-phone-number.md)
* Service: **azure-communication-services**
* GitHub Login: @Prakulka
* Microsoft Alias: **prakulka** | 1.0 | Replace "plans" concept with "features" - I've gotten feedback to refrain from using the term "Plans" in the UI as it may imply an associated cost and/or fixed limits as it does with traditional consumer phone plans. Instead I updated the UI to refer to capabilities as features. I'd suggest we use this same terminology in our docs.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 641aef05-1d16-ba0d-fb48-1527583a860d
* Version Independent ID: 2ca9c55d-56a7-ec67-80da-7c0337316245
* Content: [Quickstart - Get a phone number from Azure Communication Services](https://docs.microsoft.com/en-us/azure/communication-services/quickstarts/telephony-sms/get-phone-number)
* Content Source: [articles/communication-services/quickstarts/telephony-sms/get-phone-number.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/communication-services/quickstarts/telephony-sms/get-phone-number.md)
* Service: **azure-communication-services**
* GitHub Login: @Prakulka
* Microsoft Alias: **prakulka** | non_defect | replace plans concept with features i ve gotten feedback to refrain from using the term plans in the ui as it may imply an associated cost and or fixed limits as it does with traditional consumer phone plans instead i updated the ui to refer to capabilities as features i d suggest we use this same terminology in our docs document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service azure communication services github login prakulka microsoft alias prakulka | 0 |
55,385 | 14,434,208,684 | IssuesEvent | 2020-12-07 06:37:55 | SAP/fundamental-ngx | https://api.github.com/repos/SAP/fundamental-ngx | opened | Feed Input: Not able to load avatar | Defect Hunting Medium bug platform | Description: Not able to load avatar
Expected: Avatar should be visible
Screenshot:

| 1.0 | Feed Input: Not able to load avatar - Description: Not able to load avatar
Expected: Avatar should be visible
Screenshot:

| defect | feed input not able to load avatar description not able to load avatar expected avatar should be visible screenshot | 1 |
59,459 | 17,023,134,455 | IssuesEvent | 2021-07-03 00:31:37 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | leisure=pitch, leisure=common, landuse=forest and landuse=allotments render in wrong colour | Component: mapnik Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 9.30am, Wednesday, 29th November 2006]**
The three main blocks in the middle of this view should all be green:
http://labs.metacarta.com/osm/?lat=6681345.88613&lon=-75271.4368&zoom=15&layers=B
landuse=cemetery looks OK, so colours related to that would be suitable.
landuse=forest is also rendering in grey: I suggest using a darker green for this. | 1.0 | leisure=pitch, leisure=common, landuse=forest and landuse=allotments render in wrong colour - **[Submitted to the original trac issue database at 9.30am, Wednesday, 29th November 2006]**
The three main blocks in the middle of this view should all be green:
http://labs.metacarta.com/osm/?lat=6681345.88613&lon=-75271.4368&zoom=15&layers=B
landuse=cemetery looks OK, so colours related to that would be suitable.
landuse=forest is also rendering in grey: I suggest using a darker green for this. | defect | leisure pitch leisure common landuse forest and landuse allotments render in wrong colour the three main blocks in the middle of this view should all be green landuse cemetery looks ok so colours related to that would be suitable landuse forest is also rendering in grey i suggest using a darker green for this | 1 |
112,547 | 14,264,737,219 | IssuesEvent | 2020-11-20 16:08:27 | FilthyCriminals/game-wip | https://api.github.com/repos/FilthyCriminals/game-wip | opened | Crystal Monster | Creature Design | This is for all general discussion on design/development/assets for the crystal monster that Tony shared | 1.0 | Crystal Monster - This is for all general discussion on design/development/assets for the crystal monster that Tony shared | non_defect | crystal monster this is for all general discussion on design development assets for the crystal monster that tony shared | 0 |
64,728 | 18,848,205,134 | IssuesEvent | 2021-11-11 17:15:32 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | The sidebar icon isn't roundish | T-Defect S-Minor A-User-Settings A-Spaces A-Appearance Z-IA | 
All the other icons are round or roundish and the sidebar icon feels a little out of place, IMO | 1.0 | The sidebar icon isn't roundish - 
All the other icons are round or roundish and the sidebar icon feels a little out of place, IMO | defect | the sidebar icon isn t roundish all the other icons are round or roundish and the sidebar icon feels a little out of place imo | 1 |
154,585 | 19,730,317,640 | IssuesEvent | 2022-01-14 01:11:23 | harrinry/DataflowTemplates | https://api.github.com/repos/harrinry/DataflowTemplates | opened | CVE-2018-11765 (High) detected in hadoop-common-2.8.5.jar | security vulnerability | ## CVE-2018-11765 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hadoop-common-2.8.5.jar</b></p></summary>
<p>Apache Hadoop Common</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /ory/org/apache/hadoop/hadoop-common/2.8.5/hadoop-common-2.8.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **hadoop-common-2.8.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/harrinry/DataflowTemplates/commit/dd7cd6660b3c3d0de5f379d8294b49e38a94ca65">dd7cd6660b3c3d0de5f379d8294b49e38a94ca65</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Apache Hadoop versions 3.0.0-alpha2 to 3.0.0, 2.9.0 to 2.9.2, 2.8.0 to 2.8.5, any users can access some servlets without authentication when Kerberos authentication is enabled and SPNEGO through HTTP is not enabled.
<p>Publish Date: 2020-09-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11765>CVE-2018-11765</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1883549">https://bugzilla.redhat.com/show_bug.cgi?id=1883549</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: org.apache.hadoop:hadoop-common:2.8.6,2.9.3,3.0.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.hadoop","packageName":"hadoop-common","packageVersion":"2.8.5","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.hadoop:hadoop-common:2.8.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.hadoop:hadoop-common:2.8.6,2.9.3,3.0.1","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-11765","vulnerabilityDetails":"In Apache Hadoop versions 3.0.0-alpha2 to 3.0.0, 2.9.0 to 2.9.2, 2.8.0 to 2.8.5, any users can access some servlets without authentication when Kerberos authentication is enabled and SPNEGO through HTTP is not enabled.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11765","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-11765 (High) detected in hadoop-common-2.8.5.jar - ## CVE-2018-11765 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hadoop-common-2.8.5.jar</b></p></summary>
<p>Apache Hadoop Common</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /ory/org/apache/hadoop/hadoop-common/2.8.5/hadoop-common-2.8.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **hadoop-common-2.8.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/harrinry/DataflowTemplates/commit/dd7cd6660b3c3d0de5f379d8294b49e38a94ca65">dd7cd6660b3c3d0de5f379d8294b49e38a94ca65</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Apache Hadoop versions 3.0.0-alpha2 to 3.0.0, 2.9.0 to 2.9.2, 2.8.0 to 2.8.5, any users can access some servlets without authentication when Kerberos authentication is enabled and SPNEGO through HTTP is not enabled.
<p>Publish Date: 2020-09-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11765>CVE-2018-11765</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1883549">https://bugzilla.redhat.com/show_bug.cgi?id=1883549</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: org.apache.hadoop:hadoop-common:2.8.6,2.9.3,3.0.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.hadoop","packageName":"hadoop-common","packageVersion":"2.8.5","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.hadoop:hadoop-common:2.8.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.hadoop:hadoop-common:2.8.6,2.9.3,3.0.1","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-11765","vulnerabilityDetails":"In Apache Hadoop versions 3.0.0-alpha2 to 3.0.0, 2.9.0 to 2.9.2, 2.8.0 to 2.8.5, any users can access some servlets without authentication when Kerberos authentication is enabled and SPNEGO through HTTP is not enabled.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11765","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_defect | cve high detected in hadoop common jar cve high severity vulnerability vulnerable library hadoop common jar apache hadoop common path to dependency file pom xml path to vulnerable library ory org apache hadoop hadoop common hadoop common jar dependency hierarchy x hadoop common jar vulnerable library found in head commit a href found in base branch master vulnerability details in apache hadoop versions to to to any users can access some servlets without authentication when kerberos authentication is enabled and spnego through http is not enabled publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache hadoop hadoop common rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache hadoop hadoop common isminimumfixversionavailable true minimumfixversion org apache hadoop hadoop common isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in apache hadoop versions to to to any users can access some servlets without authentication when kerberos authentication is enabled and spnego through http is not enabled vulnerabilityurl | 0 |
29,244 | 5,626,671,553 | IssuesEvent | 2017-04-04 22:31:58 | jccastillo0007/eFacturaT | https://api.github.com/repos/jccastillo0007/eFacturaT | opened | CCE1.1 - envia los domicilios del receptor y emisor del cfdi estándar, con claves | bug defect | Las claves en el domicilio del receptor y emisor, SOLO APLICAN CUANDO SE INCLUYE EL CCE | 1.0 | CCE1.1 - envia los domicilios del receptor y emisor del cfdi estándar, con claves - Las claves en el domicilio del receptor y emisor, SOLO APLICAN CUANDO SE INCLUYE EL CCE | defect | envia los domicilios del receptor y emisor del cfdi estándar con claves las claves en el domicilio del receptor y emisor solo aplican cuando se incluye el cce | 1 |
78,446 | 27,525,355,997 | IssuesEvent | 2023-03-06 17:38:25 | vector-im/element-x-ios | https://api.github.com/repos/vector-im/element-x-ios | opened | Flash of content whilst entering a room I have just entered | T-Defect | ### Steps to reproduce
1. Go to room A, known history is the last 3 messages
2. Wait for the previous messages to be loaded
3. Go back to home screen
4. Get back in room A
### Outcome
#### What did you expect?
For the room to be shown instantaneously
#### What happened instead?
I see the 3 original messages that were displayed, a spinner above them, and a split second later the entire timeline get shown.
This creates a jarring experience
### Your phone model
_No response_
### Operating system version
_No response_
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | Flash of content whilst entering a room I have just entered - ### Steps to reproduce
1. Go to room A, known history is the last 3 messages
2. Wait for the previous messages to be loaded
3. Go back to home screen
4. Get back in room A
### Outcome
#### What did you expect?
For the room to be shown instantaneously
#### What happened instead?
I see the 3 original messages that were displayed, a spinner above them, and a split second later the entire timeline get shown.
This creates a jarring experience
### Your phone model
_No response_
### Operating system version
_No response_
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No | defect | flash of content whilst entering a room i have just entered steps to reproduce go to room a known history is the last messages wait for the previous messages to be loaded go back to home screen get back in room a outcome what did you expect for the room to be shown instantaneously what happened instead i see the original messages that were displayed a spinner above them and a split second later the entire timeline get shown this creates a jarring experience your phone model no response operating system version no response application version no response homeserver no response will you send logs no | 1 |
52,209 | 13,211,407,432 | IssuesEvent | 2020-08-15 22:55:33 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | [I3OmDb] Not compiling for icerec/V05-00-02 (Trac #1853) | Incomplete Migration Migrated from Trac combo reconstruction defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1853">https://code.icecube.wisc.edu/projects/icecube/ticket/1853</a>, reported by maryonand owned by joertlin</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:24",
"_ts": "1550067204154158",
"description": "Does not compile for me when using clang however gcc works. Forward declaration wrongly executed.\n\n{{{\n[ 88%] Building CXX object STTools/CMakeFiles/STTools-pybindings.dir/private/pybindings/algorithms/seededRT/I3SeededRTConfigurationService.cxx.o\nIn file included from /home/maryon/software/meta-projects/icerec/V05-00-02/src/I3Db/private/I3OmDb/I3OmDbFactory.cxx:2:\n/home/maryon/software/meta-projects/icerec/V05-00-02/src/I3Db/public/I3Db/I3OmDb/I3OmDb.h:1280:8: error: 'Om_s' redeclared with 'private' access\nstruct Om_s \n ^\n/home/maryon/software/meta-projects/icerec/V05-00-02/src/I3Db/public/I3Db/I3OmDb/I3OmDb.h:977:8: note: previously declared 'public' here\nstruct Om_s;\n ^\n1 error generated.\nmake[2]: *** [I3Db/CMakeFiles/I3Db.dir/private/I3OmDb/I3OmDbFactory.cxx.o] Error 1\nmake[2]: Target `I3Db/CMakeFiles/I3Db.dir/build' not remade because of errors.\nmake[1]: *** [I3Db/CMakeFiles/I3Db.dir/all] Error 2\n}}}\n",
"reporter": "maryon",
"cc": "",
"resolution": "fixed",
"time": "2016-09-09T11:50:36",
"component": "combo reconstruction",
"summary": "[I3OmDb] Not compiling for icerec/V05-00-02",
"priority": "blocker",
"keywords": "I3OmDb icerec",
"milestone": "",
"owner": "joertlin",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [I3OmDb] Not compiling for icerec/V05-00-02 (Trac #1853) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1853">https://code.icecube.wisc.edu/projects/icecube/ticket/1853</a>, reported by maryonand owned by joertlin</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:24",
"_ts": "1550067204154158",
"description": "Does not compile for me when using clang however gcc works. Forward declaration wrongly executed.\n\n{{{\n[ 88%] Building CXX object STTools/CMakeFiles/STTools-pybindings.dir/private/pybindings/algorithms/seededRT/I3SeededRTConfigurationService.cxx.o\nIn file included from /home/maryon/software/meta-projects/icerec/V05-00-02/src/I3Db/private/I3OmDb/I3OmDbFactory.cxx:2:\n/home/maryon/software/meta-projects/icerec/V05-00-02/src/I3Db/public/I3Db/I3OmDb/I3OmDb.h:1280:8: error: 'Om_s' redeclared with 'private' access\nstruct Om_s \n ^\n/home/maryon/software/meta-projects/icerec/V05-00-02/src/I3Db/public/I3Db/I3OmDb/I3OmDb.h:977:8: note: previously declared 'public' here\nstruct Om_s;\n ^\n1 error generated.\nmake[2]: *** [I3Db/CMakeFiles/I3Db.dir/private/I3OmDb/I3OmDbFactory.cxx.o] Error 1\nmake[2]: Target `I3Db/CMakeFiles/I3Db.dir/build' not remade because of errors.\nmake[1]: *** [I3Db/CMakeFiles/I3Db.dir/all] Error 2\n}}}\n",
"reporter": "maryon",
"cc": "",
"resolution": "fixed",
"time": "2016-09-09T11:50:36",
"component": "combo reconstruction",
"summary": "[I3OmDb] Not compiling for icerec/V05-00-02",
"priority": "blocker",
"keywords": "I3OmDb icerec",
"milestone": "",
"owner": "joertlin",
"type": "defect"
}
```
</p>
</details>
| defect | not compiling for icerec trac migrated from json status closed changetime ts description does not compile for me when using clang however gcc works forward declaration wrongly executed n n n building cxx object sttools cmakefiles sttools pybindings dir private pybindings algorithms seededrt cxx o nin file included from home maryon software meta projects icerec src private cxx n home maryon software meta projects icerec src public h error om s redeclared with private access nstruct om s n n home maryon software meta projects icerec src public h note previously declared public here nstruct om s n error generated nmake error nmake target cmakefiles dir build not remade because of errors nmake error n n reporter maryon cc resolution fixed time component combo reconstruction summary not compiling for icerec priority blocker keywords icerec milestone owner joertlin type defect | 1 |
72,385 | 3,384,994,200 | IssuesEvent | 2015-11-27 08:52:10 | OCHA-DAP/hdx-ckan | https://api.github.com/repos/OCHA-DAP/hdx-ckan | closed | Create a functional base test class based on HdxBaseTest | InProgress Priority-Low | This will automatically load all the HDX plugins so that new tests won't need to override the _load_plugins() function.
This will still allow the flexibility for each test class to only inherit from HdxBaseTest and still only load the plugins it needs.
**Note** Extensions that we plan to make available for the public should NOT create tests based on this new functional base test class | 1.0 | Create a functional base test class based on HdxBaseTest - This will automatically load all the HDX plugins so that new tests won't need to override the _load_plugins() function.
This will still allow the flexibility for each test class to only inherit from HdxBaseTest and still only load the plugins it needs.
**Note** Extensions that we plan to make available for the public should NOT create tests based on this new functional base test class | non_defect | create a functional base test class based on hdxbasetest this will automatically load all the hdx plugins so that new tests won t need to override the load plugins function this will still allow the flexibility for each test class to only inherit from hdxbasetest and still only load the plugins it needs note extensions that we plan to make available for the public should not create tests based on this new functional base test class | 0 |
29,902 | 5,953,287,513 | IssuesEvent | 2017-05-27 06:08:05 | BigBadaboom/androidsvg | https://api.github.com/repos/BigBadaboom/androidsvg | closed | Pattern fills are rendered with visible grid | auto-migrated Priority-Medium Type-Defect | ```
Hey, first of all I want to thank you! androidsvg is THE BEST android svg
parser/renderer that I came across.
I found small problem if you have a pattern fill there are tiny hairline gaps
between pattern repetitions. Please try attached svg and you will see yourself.
```
Original issue reported on code.google.com by `asyschi...@gmail.com` on 18 Mar 2015 at 1:06
Attachments:
- [drawing2.svg](https://storage.googleapis.com/google-code-attachments/androidsvg/issue-55/comment-0/drawing2.svg)
| 1.0 | Pattern fills are rendered with visible grid - ```
Hey, first of all I want to thank you! androidsvg is THE BEST android svg
parser/renderer that I came across.
I found small problem if you have a pattern fill there are tiny hairline gaps
between pattern repetitions. Please try attached svg and you will see yourself.
```
Original issue reported on code.google.com by `asyschi...@gmail.com` on 18 Mar 2015 at 1:06
Attachments:
- [drawing2.svg](https://storage.googleapis.com/google-code-attachments/androidsvg/issue-55/comment-0/drawing2.svg)
| defect | pattern fills are rendered with visible grid hey first of all i want to thank you androidsvg is the best android svg parser renderer that i came across i found small problem if you have a pattern fill there are tiny hairline gaps between pattern repetitions please try attached svg and you will see yourself original issue reported on code google com by asyschi gmail com on mar at attachments | 1 |
59,401 | 17,023,116,872 | IssuesEvent | 2021-07-03 00:27:02 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | danish æ and å become garbled when downloading from server | Component: josm Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 9.41pm, Wednesday, 17th May 2006]**
Adding a property value containing danish (ae) or (aa) characters (like name=Sndergrds Alle, see coord. 55.4739687619073,8.47640550265594) results in a garbled value when uploading and re-downloading the area. After download the character including the follwing two chracters are replaced by a square. Strangely the character is left untouched.
Seen on both Windows 98 and Windows Server 2003 using java version "1.5.0_06", Java(TM) 2 Runtime Environment, Standard Edition (build 1.5.0_06-b05), Java HotSpot(TM) Client VM (build 1.5.0_06-b05, mixed mode).
Mikkel, | 1.0 | danish æ and å become garbled when downloading from server - **[Submitted to the original trac issue database at 9.41pm, Wednesday, 17th May 2006]**
Adding a property value containing danish (ae) or (aa) characters (like name=Sndergrds Alle, see coord. 55.4739687619073,8.47640550265594) results in a garbled value when uploading and re-downloading the area. After download the character including the follwing two chracters are replaced by a square. Strangely the character is left untouched.
Seen on both Windows 98 and Windows Server 2003 using java version "1.5.0_06", Java(TM) 2 Runtime Environment, Standard Edition (build 1.5.0_06-b05), Java HotSpot(TM) Client VM (build 1.5.0_06-b05, mixed mode).
Mikkel, | defect | danish 㦠and 㥠become garbled when downloading from server adding a property value containing danish ae or aa characters like name sndergrds alle see coord results in a garbled value when uploading and re downloading the area after download the character including the follwing two chracters are replaced by a square strangely the character is left untouched seen on both windows and windows server using java version java tm runtime environment standard edition build java hotspot tm client vm build mixed mode mikkel | 1 |
36,226 | 17,569,802,696 | IssuesEvent | 2021-08-14 12:50:14 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | opened | Getting non-deterministic results on TF1.9 | type:performance | Getting different results on different attempts of execution.
I AM RUNNING IT ON CPU.
1. os.environ['TF_CUDNN_DETERMINISTIC']='1'
2. os.environ['HOROVOD_FUSION_THRESHOLD']='0'
The above 2 lines can be ignored for CPU execution.
Could anyone help me getting determinstic results?
```
import os
import numpy as np
import tensorflow as tf
from sklearn.metrics import precision_recall_fscore_support
import config
from data_helper import batch_index, load_word2id, load_y2id_id2y, load_word2vector, recover_data_from_files
from model.nn_layer import transition_layer, softmax_layer
import random
tf.set_random_seed(42)
np.random.seed(42)
os.environ['PYTHONHASHSEED']=str(42)
random.seed(42)
os.environ['TF_DETERMINISTIC_OPS'] = '1'
os.environ['TF_CUDNN_DETERMINISTIC']='1'
os.environ['HOROVOD_FUSION_THRESHOLD']='0'
class NetAbModel(object):
def __init__(self, domain, flags, filter_list=(3, 4, 5), filter_num=100):
self.config = flags
self.filter_list = filter_list
self.filter_num = filter_num
# placeholder
self.sen_x_batch = None
self.sent_len_batch = None
self.sen_y_batch = None
self.keep_prob1 = None
self.keep_prob2 = None
# embedding
self.add_placeholder()
self.word2id = None
# self.id2word = None
self.vocab_size = None
self.embedding = None
inputs = self.add_embedding(domain)
# model
self.sen_logits, self.sen_logits2 = self.netAb(inputs)
# noisy-loss
self.loss = self.add_loss(self.sen_logits)
self.accuracy, self.accuracy_num = self.add_accuracy(self.sen_logits)
self.train_op = self.add_train_op(self.loss)
# clean-loss
self.loss2 = self.add_loss(self.sen_logits2)
self.accuracy2, self.accuracy_num2 = self.add_accuracy(self.sen_logits2)
self.train_op2 = self.add_train_op(self.loss2)
def add_placeholder(self):
self.sen_x_batch = tf.placeholder(tf.int32, [None, self.config.max_sentence_len])
self.sent_len_batch = tf.placeholder(tf.int32, [None])
self.sen_y_batch = tf.placeholder(tf.float32, [None, self.config.n_class])
self.keep_prob1 = tf.placeholder(tf.float32)
self.keep_prob2 = tf.placeholder(tf.float32)
def add_embedding(self, domain):
if self.config.pre_trained:
self.word2id, w2v = load_word2vector(self.config.word2id_path, domain)
# self.word2id, self.id2word, w2v = load_w2v_mongo(domain)
else:
self.word2id = load_word2id(self.config.word2id_path, domain)
self.vocab_size = len(self.word2id)
w2v = tf.random_uniform([self.vocab_size, self.config.embedding_dim], -1.0, 1.0, trainable=True, seed=42)
if self.config.embedding_type == 'static':
self.embedding = tf.constant(w2v, dtype=tf.float32, name='word_embedding')
else:
self.embedding = tf.Variable(w2v, dtype=tf.float32, name='word_embedding')
inputs = tf.nn.embedding_lookup(self.embedding, self.sen_x_batch)
return inputs
def create_feed_dict(self, sen_x_batch, sent_len_batch, sen_y_batch, kp1=1.0, kp2=1.0):
holder_list = [self.sen_x_batch, self.sent_len_batch, self.sen_y_batch,
self.keep_prob1, self.keep_prob2]
feed_list = [sen_x_batch, sent_len_batch, sen_y_batch, kp1, kp2]
return dict(zip(holder_list, feed_list))
# cnn layer
def add_cnn_layer(self, inputs, inputs_dim, max_len, scope_name='cnn'):
inputs = tf.expand_dims(inputs, -1)
pooling_outputs = []
for i, filter_size in enumerate(self.filter_list):
ksize = [filter_size, inputs_dim]
conv = tf.contrib.layers.conv2d(inputs=inputs,
num_outputs=self.filter_num,
kernel_size=ksize,
stride=1,
padding='VALID',
activation_fn=tf.nn.relu,
scope='conv_' + scope_name + str(i))
ksize = [max_len - filter_size + 1, 1]
pooling = tf.contrib.layers.max_pool2d(inputs=conv,
kernel_size=ksize,
stride=1,
padding='VALID',
scope='pooling_' + scope_name)
pooling_outputs.append(pooling)
hiddens = tf.concat(pooling_outputs, 3)
hiddens = tf.reshape(hiddens, [-1, self.filter_num * len(self.filter_list)])
return hiddens
# cnn layer
def add_noisy_cnn_layer(self, inputs, inputs_dim, max_len, scope_name='cnn'):
inputs = tf.expand_dims(inputs, -1)
pooling_outputs = []
for i, filter_size in enumerate(self.filter_list):
ksize = [filter_size, inputs_dim]
conv = tf.contrib.layers.conv2d(inputs=inputs,
num_outputs=self.filter_num,
kernel_size=ksize,
stride=1,
padding='VALID',
activation_fn=tf.nn.relu,
scope='conv_' + scope_name + str(i))
ksize = [max_len - filter_size + 1, 1]
pooling = tf.contrib.layers.max_pool2d(inputs=conv,
kernel_size=ksize,
stride=1,
padding='VALID',
scope='pooling_' + scope_name)
pooling_outputs.append(pooling)
hiddens = tf.concat(pooling_outputs, 3)
hiddens = tf.reshape(hiddens, [-1, self.filter_num * len(self.filter_list)])
return hiddens
def netAb(self, inputs):
print('Running NetAb...')
inputs = tf.nn.dropout(inputs, keep_prob=self.keep_prob1, seed=42)
inputs = tf.reshape(inputs, [-1, self.config.max_sentence_len, self.config.embedding_dim])
# word-sentence: cnn
outputs_sen = self.add_cnn_layer(inputs, self.config.embedding_dim, self.config.max_sentence_len, 'h')
outputs_sen_dim = self.filter_num * len(self.filter_list)
outputs_sen = tf.reshape(outputs_sen, [-1, outputs_sen_dim])
noisy_cnn = self.add_noisy_cnn_layer(inputs, self.config.embedding_dim, self.config.max_sentence_len, 'u')
noisy_cnn = tf.reshape(noisy_cnn, [-1, outputs_sen_dim])
# fully-connection
clean_logits = softmax_layer(outputs_sen, outputs_sen_dim, self.config.random_base, self.keep_prob2,
self.config.l2_reg, self.config.n_class, 'sen_softmax')
p1 = transition_layer(noisy_cnn, outputs_sen_dim, self.config.l2_reg, self.config.random_base, 'p1')
p2 = transition_layer(noisy_cnn, outputs_sen_dim, self.config.l2_reg, self.config.random_base, 'p2')
p1 = tf.expand_dims(p1, 2)
p2 = tf.expand_dims(p2, 2)
prob = tf.concat([p1, p2], 2)
sen_logits = tf.expand_dims(clean_logits, 1)
noisy_logits = tf.squeeze(tf.matmul(sen_logits, prob))
return noisy_logits, clean_logits
def add_loss(self, sen_logits):
loss = tf.nn.softmax_cross_entropy_with_logits_v2(logits=sen_logits, labels=self.sen_y_batch)
self.sen_vars = [var for var in tf.global_variables()
if 'h' in var.name or 'u' in var.name or 'p1' in var.name or 'p2' in var.name]
# print(self.sen_vars)
reg_loss = tf.get_collection(tf.GraphKeys.REGULARIZATION_LOSSES, scope='sen_softmax')
# print(reg_loss)
loss = tf.reduce_mean(loss) # TODO+ self.config.l1_reg * tf.add_n(reg_loss)
return loss
def add_accuracy(self, scores):
correct_predicts = tf.equal(tf.argmax(scores, 1), tf.argmax(self.sen_y_batch, 1))
accuracy_num = tf.reduce_sum(tf.cast(correct_predicts, tf.int32)) # the number of correct predicting docs
accuracy = tf.reduce_mean(tf.cast(correct_predicts, tf.float32), name='accuracy') # accuracy metric result
return accuracy, accuracy_num
def add_train_op(self, doc_loss):
# new_learning_rate = current_learning_rate * decay_rate ^ (global_step / decay_steps)
global_step = tf.Variable(0, name='global_step', trainable=False) # record the current step (global step)
self.lr = tf.train.exponential_decay(self.config.lr, global_step, self.config.decay_steps,
self.config.decay_rate, staircase=True)
# the optimizer used in this work
# optimizer = tf.train.AdadeltaOptimizer(self.lr)
optimizer = tf.train.AdamOptimizer(self.lr)
grads, global_norm = tf.clip_by_global_norm(tf.gradients(doc_loss, self.sen_vars, gate_gradients=True), self.config.max_grad_norm)
train_op = optimizer.apply_gradients(zip(grads, self.sen_vars), name='train_op', global_step=global_step)
# train_op = optimizer.minimize(doc_loss, global_step=global_step, var_list=self.doc_vars)
return train_op
def run_op(self, sess, op, sen_x, sen_len, sen_y, kp1=1.0, kp2=1.0):
res_list = []
len_list = []
for indices in batch_index(len(sen_x), self.config.batch_size, n_iter=1, is_shuffle=False, is_train=False):
feed_dict = self.create_feed_dict(sen_x[indices], sen_len[indices], sen_y[indices], kp1, kp2)
res = sess.run(op, feed_dict=feed_dict)
res_list.append(res)
len_list.append(len(indices))
if type(res_list[0]) is list: # if op is a list
res = np.concatenate(res_list, axis=1)
elif op is self.accuracy_num or op is self.accuracy_num2:
res = sum(res_list) # sum all batches
elif op is self.sen_logits or op is self.sen_logits2:
res = np.concatenate(np.asarray(res_list), 0)
else: # for los, etc.
res = sum(res_list) * 1.0 / len(len_list)
return res
def run_cleaner(self, sess, feed_dict):
sess.run([self.train_op2], feed_dict=feed_dict)
def pre_run(self, sess, feed_dict):
sess.run([self.train_op2], feed_dict=feed_dict)
def run(self, sess, feed_dict):
logits = sess.run([self.sen_logits2], feed_dict=feed_dict)
_, loss, acc_num = sess.run([self.train_op, self.loss, self.accuracy_num], feed_dict=feed_dict)
return loss, acc_num, np.concatenate(np.asarray(logits), 0)
def test_case(sess, classifier, sen_x, sen_len, sen_y):
score = classifier.run_op(sess, classifier.sen_logits2, sen_x, sen_len, sen_y)
loss = classifier.run_op(sess, classifier.loss2, sen_x, sen_len, sen_y)
acc_num = classifier.run_op(sess, classifier.accuracy_num2, sen_x, sen_len, sen_y)
y_pred = np.argmax(score, axis=1)
y_true = np.argmax(sen_y, axis=1)
p_class, r_class, f_class, support_micro = precision_recall_fscore_support(y_true=y_true, y_pred=y_pred,
labels=[0, 1], average=None)
return acc_num * 1.0 / len(sen_y), loss, f_class[0]
def run_test(sess, classifier, domain, sen_x, sen_len, sen_y):
scores = classifier.run_op(sess, classifier.sen_logits2, sen_x, sen_len, sen_y)
acc_num = classifier.run_op(sess, classifier.accuracy_num2, sen_x, sen_len, sen_y)
y_pred = np.argmax(scores, axis=1)
y_true = np.argmax(sen_y, axis=1)
p_class, r_class, f_class, support_micro = precision_recall_fscore_support(y_true=y_true, y_pred=y_pred,
labels=[0, 1], average=None)
_, id2y = load_y2id_id2y('./data/y2id.txt')
result_save_path = classifier.config.result_path + classifier.config.model + '/'
if not os.path.exists(result_save_path):
os.makedirs(result_save_path)
with open(result_save_path + domain + '_test.txt', 'w', encoding='utf-8') as fin:
fin.write('ACC: ' + str(acc_num * 1.0 / len(sen_x)) + '\t')
fin.write('P: ' + str(p_class) + '\tR: ' + str(r_class) +
'\tF1: ' + str(f_class) + '\tF1_macro: ' + str(f_class.mean()) + '\n')
for id_y in y_pred:
fin.write(id2y[id_y] + '\n')
with open(result_save_path + domain + '_true.txt', 'w', encoding='utf-8') as fin:
for id_y in y_true:
fin.write(id2y[id_y] + '\n')
print('Test. Acc = {}, P = {}, R = {}, F1 = {}, F1_macro = {}'.
format(acc_num * 1.0 / len(sen_x), p_class, r_class, f_class, f_class.mean()))
def train_run(_):
flags_ = config.FLAGS
domain = flags_.dataset # movie, laptop, restaurant
print('{} Learning start: >>>\n'.format(domain))
tf.reset_default_graph()
# os.environ['CUDA_VISIBLE_DEVICES'] = flags_.gpu
classifier = NetAbModel(domain, flags_)
gpu_config = tf.ConfigProto(intra_op_parallelism_threads=1, inter_op_parallelism_threads=1) #arguments I added
# gpu_config.gpu_options.per_process_gpu_memory_fraction = 0.85
# gpu_config.gpu_options.allow_growth = True
# gpu_config.allow_soft_placement = True # If 'True': allow cpu, if no gpu
saver = tf.train.Saver(tf.global_variables())
save_path = classifier.config.ckpt_path + classifier.config.model + '/' + domain + '/' + domain + '_ckpt'
with tf.Session(config=gpu_config) as sess:
sess.run(tf.global_variables_initializer())
best_val_acc = 0
best_val_epoch = 0
# best_test_acc = 0
training_path = os.path.join(flags_.data_path, 'TrainingSens/')
train_sen_x, train_sen_len, train_sen_y = recover_data_from_files(
training_path, 'training', domain, flags_.max_sentence_len)
val_path = os.path.join(flags_.data_path, 'ValSens/')
val_sen_x, val_sen_len, val_sen_y = recover_data_from_files(
val_path, 'validation', domain, flags_.max_sentence_len)
test_path = os.path.join(flags_.data_path, 'TestSens/')
test_sen_x, test_sen_len, test_sen_y = recover_data_from_files(
test_path, 'test', domain, flags_.max_sentence_len)
# train_sen_x, train_sen_len, train_sen_y = load_inputs_document_mongo(
# domain, 'train_noisy', classifier.word2id, flags_.max_sentence_len, flags_.max_doc_len)
# val_sen_x, val_sen_len, val_sen_y = load_inputs_document_mongo(
# domain, 'dev', classifier.word2id, flags_.max_sentence_len, flags_.max_doc_len)
# test_sen_x, test_sen_len, test_sen_y = load_inputs_document_mongo(
# domain, 'test', classifier.word2id, flags_.max_sentence_len, flags_.max_doc_len)
if classifier.config.is_train:
for epoch_i in range(flags_.n_epoch):
print('=' * 20 + 'Epoch ', epoch_i, '=' * 20)
total_loss = []
total_acc_num = []
total_num = []
if epoch_i < classifier.config.initial_epochs: # initial epochs
for step, indices in enumerate(batch_index(len(train_sen_y), flags_.batch_size, n_iter=1, is_shuffle=False), 1):
indices = list(indices)
print(train_sen_x[indices], [train_sen_y]) #-------------------------------------- I added
feed_dict = classifier.create_feed_dict(train_sen_x[indices], train_sen_len[indices],
train_sen_y[indices],
flags_.keep_prob1, flags_.keep_prob2)
classifier.pre_run(sess, feed_dict=feed_dict)
continue
for step, indices in enumerate(batch_index(len(train_sen_y), flags_.batch_size, n_iter=1, is_shuffle=False), 1):
indices = list(indices)
# if epoch_i < 10:
print("indices", train_sen_x[indices], [train_sen_y]) #-------------------------------------- I added
feed_dict = classifier.create_feed_dict(train_sen_x[indices], train_sen_len[indices],
train_sen_y[indices],
flags_.keep_prob1, flags_.keep_prob2)
loss, acc_num, logits = classifier.run(sess, feed_dict=feed_dict)
y_pred_set = np.argmax(logits, axis=1)
y_true_set = np.argmax(train_sen_y[indices], axis=1)
f_indices = np.arange(0, len(indices))
valid_indices = f_indices[y_pred_set == y_true_set]
indices_new = list(np.array(indices)[valid_indices])
print("newindices", train_sen_x[indices], [train_sen_y]) #-------------------------------------- I added
if indices_new is None:
continue
# else:
# indices_new = indices
# indices_new = indices
feed_dict = classifier.create_feed_dict(train_sen_x[indices_new], train_sen_len[indices_new],
train_sen_y[indices_new],
flags_.keep_prob1, flags_.keep_prob2)
classifier.run_cleaner(sess, feed_dict=feed_dict)
total_loss.append(loss)
total_acc_num.append(acc_num)
total_num.append(len(indices))
verbose = flags_.display_step
if step % verbose == 0:
print('[INFO] Len {}, Epoch {} - Batch {} : loss = {}, acc = {}'.format(
len(indices_new), epoch_i, step, np.mean(total_loss[-verbose:]),
sum(total_acc_num[-verbose:]) * 1.0 / sum(total_num[-verbose:])))
loss = np.mean(total_loss)
acc = sum(total_acc_num) * 1.0 / sum(total_num)
print('\n[INFO] Epoch {} : mean loss = {}, mean acc = {}'.format(epoch_i, loss, acc))
if np.isnan(loss):
raise ValueError('[Error] loss is not a number!')
# validation
val_acc, val_loss, val_f1 = test_case(sess, classifier, val_sen_x, val_sen_len, val_sen_y)
print('[INFO] val loss: {}, val acc: {}, val f1: {}'.format(val_loss, val_acc, val_f1))
# test
test_acc, test_loss, test_f1 = test_case(sess, classifier, test_sen_x, test_sen_len, test_sen_y)
print('[INFO] test loss: {}, test acc: {}, test f1: {}'.format(test_loss, test_acc, test_f1))
print('=' * 25 + ' end', '=' * 25 + '\n')
if best_val_acc < val_acc:
best_val_acc = val_acc
best_val_epoch = epoch_i
# best_test_acc = test_acc
if not os.path.exists(classifier.config.ckpt_path + classifier.config.model + '/'):
os.makedirs(classifier.config.ckpt_path + classifier.config.model + '/')
saver.save(sess, save_path=save_path)
if epoch_i - best_val_epoch > classifier.config.early_stopping:
# here early_stopping is 5 :> 'the number of early stopping epoch'
print('Normal early stop at {}!'.format(best_val_epoch))
break
print('Best val acc = {}'.format(best_val_acc))
# print('Test acc = {}'.format(best_test_acc))
best_val_epoch_save_path = classifier.config.result_path + classifier.config.model + '/'
if not os.path.exists(best_val_epoch_save_path):
os.makedirs(best_val_epoch_save_path)
with open(best_val_epoch_save_path + domain + '_bestEpoch.txt', 'w', encoding='utf-8') as fin:
fin.write('Best epoch: ' + str(best_val_epoch) + '\n')
saver.restore(sess, save_path)
print('Model restored from %s' % save_path)
# # test now
run_test(sess, classifier, domain, test_sen_x, test_sen_len, test_sen_y)
else:
saver.restore(sess, save_path)
print('Model restored from %s' % save_path)
# # test now
run_test(sess, classifier, domain, test_sen_x, test_sen_len, test_sen_y)
print('Domain {} is done..'.format(domain))
print('\nTraining complete!\n')
if __name__ == '__main__':
tf.app.run(train_run)
``` | True | Getting non-deterministic results on TF1.9 - Getting different results on different attempts of execution.
I AM RUNNING IT ON CPU.
1. os.environ['TF_CUDNN_DETERMINISTIC']='1'
2. os.environ['HOROVOD_FUSION_THRESHOLD']='0'
The above 2 lines can be ignored for CPU execution.
Could anyone help me getting determinstic results?
```
import os
import numpy as np
import tensorflow as tf
from sklearn.metrics import precision_recall_fscore_support
import config
from data_helper import batch_index, load_word2id, load_y2id_id2y, load_word2vector, recover_data_from_files
from model.nn_layer import transition_layer, softmax_layer
import random
tf.set_random_seed(42)
np.random.seed(42)
os.environ['PYTHONHASHSEED']=str(42)
random.seed(42)
os.environ['TF_DETERMINISTIC_OPS'] = '1'
os.environ['TF_CUDNN_DETERMINISTIC']='1'
os.environ['HOROVOD_FUSION_THRESHOLD']='0'
class NetAbModel(object):
def __init__(self, domain, flags, filter_list=(3, 4, 5), filter_num=100):
self.config = flags
self.filter_list = filter_list
self.filter_num = filter_num
# placeholder
self.sen_x_batch = None
self.sent_len_batch = None
self.sen_y_batch = None
self.keep_prob1 = None
self.keep_prob2 = None
# embedding
self.add_placeholder()
self.word2id = None
# self.id2word = None
self.vocab_size = None
self.embedding = None
inputs = self.add_embedding(domain)
# model
self.sen_logits, self.sen_logits2 = self.netAb(inputs)
# noisy-loss
self.loss = self.add_loss(self.sen_logits)
self.accuracy, self.accuracy_num = self.add_accuracy(self.sen_logits)
self.train_op = self.add_train_op(self.loss)
# clean-loss
self.loss2 = self.add_loss(self.sen_logits2)
self.accuracy2, self.accuracy_num2 = self.add_accuracy(self.sen_logits2)
self.train_op2 = self.add_train_op(self.loss2)
def add_placeholder(self):
self.sen_x_batch = tf.placeholder(tf.int32, [None, self.config.max_sentence_len])
self.sent_len_batch = tf.placeholder(tf.int32, [None])
self.sen_y_batch = tf.placeholder(tf.float32, [None, self.config.n_class])
self.keep_prob1 = tf.placeholder(tf.float32)
self.keep_prob2 = tf.placeholder(tf.float32)
def add_embedding(self, domain):
if self.config.pre_trained:
self.word2id, w2v = load_word2vector(self.config.word2id_path, domain)
# self.word2id, self.id2word, w2v = load_w2v_mongo(domain)
else:
self.word2id = load_word2id(self.config.word2id_path, domain)
self.vocab_size = len(self.word2id)
w2v = tf.random_uniform([self.vocab_size, self.config.embedding_dim], -1.0, 1.0, trainable=True, seed=42)
if self.config.embedding_type == 'static':
self.embedding = tf.constant(w2v, dtype=tf.float32, name='word_embedding')
else:
self.embedding = tf.Variable(w2v, dtype=tf.float32, name='word_embedding')
inputs = tf.nn.embedding_lookup(self.embedding, self.sen_x_batch)
return inputs
def create_feed_dict(self, sen_x_batch, sent_len_batch, sen_y_batch, kp1=1.0, kp2=1.0):
holder_list = [self.sen_x_batch, self.sent_len_batch, self.sen_y_batch,
self.keep_prob1, self.keep_prob2]
feed_list = [sen_x_batch, sent_len_batch, sen_y_batch, kp1, kp2]
return dict(zip(holder_list, feed_list))
# cnn layer
def add_cnn_layer(self, inputs, inputs_dim, max_len, scope_name='cnn'):
inputs = tf.expand_dims(inputs, -1)
pooling_outputs = []
for i, filter_size in enumerate(self.filter_list):
ksize = [filter_size, inputs_dim]
conv = tf.contrib.layers.conv2d(inputs=inputs,
num_outputs=self.filter_num,
kernel_size=ksize,
stride=1,
padding='VALID',
activation_fn=tf.nn.relu,
scope='conv_' + scope_name + str(i))
ksize = [max_len - filter_size + 1, 1]
pooling = tf.contrib.layers.max_pool2d(inputs=conv,
kernel_size=ksize,
stride=1,
padding='VALID',
scope='pooling_' + scope_name)
pooling_outputs.append(pooling)
hiddens = tf.concat(pooling_outputs, 3)
hiddens = tf.reshape(hiddens, [-1, self.filter_num * len(self.filter_list)])
return hiddens
# cnn layer
def add_noisy_cnn_layer(self, inputs, inputs_dim, max_len, scope_name='cnn'):
inputs = tf.expand_dims(inputs, -1)
pooling_outputs = []
for i, filter_size in enumerate(self.filter_list):
ksize = [filter_size, inputs_dim]
conv = tf.contrib.layers.conv2d(inputs=inputs,
num_outputs=self.filter_num,
kernel_size=ksize,
stride=1,
padding='VALID',
activation_fn=tf.nn.relu,
scope='conv_' + scope_name + str(i))
ksize = [max_len - filter_size + 1, 1]
pooling = tf.contrib.layers.max_pool2d(inputs=conv,
kernel_size=ksize,
stride=1,
padding='VALID',
scope='pooling_' + scope_name)
pooling_outputs.append(pooling)
hiddens = tf.concat(pooling_outputs, 3)
hiddens = tf.reshape(hiddens, [-1, self.filter_num * len(self.filter_list)])
return hiddens
def netAb(self, inputs):
print('Running NetAb...')
inputs = tf.nn.dropout(inputs, keep_prob=self.keep_prob1, seed=42)
inputs = tf.reshape(inputs, [-1, self.config.max_sentence_len, self.config.embedding_dim])
# word-sentence: cnn
outputs_sen = self.add_cnn_layer(inputs, self.config.embedding_dim, self.config.max_sentence_len, 'h')
outputs_sen_dim = self.filter_num * len(self.filter_list)
outputs_sen = tf.reshape(outputs_sen, [-1, outputs_sen_dim])
noisy_cnn = self.add_noisy_cnn_layer(inputs, self.config.embedding_dim, self.config.max_sentence_len, 'u')
noisy_cnn = tf.reshape(noisy_cnn, [-1, outputs_sen_dim])
# fully-connection
clean_logits = softmax_layer(outputs_sen, outputs_sen_dim, self.config.random_base, self.keep_prob2,
self.config.l2_reg, self.config.n_class, 'sen_softmax')
p1 = transition_layer(noisy_cnn, outputs_sen_dim, self.config.l2_reg, self.config.random_base, 'p1')
p2 = transition_layer(noisy_cnn, outputs_sen_dim, self.config.l2_reg, self.config.random_base, 'p2')
p1 = tf.expand_dims(p1, 2)
p2 = tf.expand_dims(p2, 2)
prob = tf.concat([p1, p2], 2)
sen_logits = tf.expand_dims(clean_logits, 1)
noisy_logits = tf.squeeze(tf.matmul(sen_logits, prob))
return noisy_logits, clean_logits
def add_loss(self, sen_logits):
loss = tf.nn.softmax_cross_entropy_with_logits_v2(logits=sen_logits, labels=self.sen_y_batch)
self.sen_vars = [var for var in tf.global_variables()
if 'h' in var.name or 'u' in var.name or 'p1' in var.name or 'p2' in var.name]
# print(self.sen_vars)
reg_loss = tf.get_collection(tf.GraphKeys.REGULARIZATION_LOSSES, scope='sen_softmax')
# print(reg_loss)
loss = tf.reduce_mean(loss) # TODO+ self.config.l1_reg * tf.add_n(reg_loss)
return loss
def add_accuracy(self, scores):
correct_predicts = tf.equal(tf.argmax(scores, 1), tf.argmax(self.sen_y_batch, 1))
accuracy_num = tf.reduce_sum(tf.cast(correct_predicts, tf.int32)) # the number of correct predicting docs
accuracy = tf.reduce_mean(tf.cast(correct_predicts, tf.float32), name='accuracy') # accuracy metric result
return accuracy, accuracy_num
def add_train_op(self, doc_loss):
# new_learning_rate = current_learning_rate * decay_rate ^ (global_step / decay_steps)
global_step = tf.Variable(0, name='global_step', trainable=False) # record the current step (global step)
self.lr = tf.train.exponential_decay(self.config.lr, global_step, self.config.decay_steps,
self.config.decay_rate, staircase=True)
# the optimizer used in this work
# optimizer = tf.train.AdadeltaOptimizer(self.lr)
optimizer = tf.train.AdamOptimizer(self.lr)
grads, global_norm = tf.clip_by_global_norm(tf.gradients(doc_loss, self.sen_vars, gate_gradients=True), self.config.max_grad_norm)
train_op = optimizer.apply_gradients(zip(grads, self.sen_vars), name='train_op', global_step=global_step)
# train_op = optimizer.minimize(doc_loss, global_step=global_step, var_list=self.doc_vars)
return train_op
def run_op(self, sess, op, sen_x, sen_len, sen_y, kp1=1.0, kp2=1.0):
res_list = []
len_list = []
for indices in batch_index(len(sen_x), self.config.batch_size, n_iter=1, is_shuffle=False, is_train=False):
feed_dict = self.create_feed_dict(sen_x[indices], sen_len[indices], sen_y[indices], kp1, kp2)
res = sess.run(op, feed_dict=feed_dict)
res_list.append(res)
len_list.append(len(indices))
if type(res_list[0]) is list: # if op is a list
res = np.concatenate(res_list, axis=1)
elif op is self.accuracy_num or op is self.accuracy_num2:
res = sum(res_list) # sum all batches
elif op is self.sen_logits or op is self.sen_logits2:
res = np.concatenate(np.asarray(res_list), 0)
else: # for los, etc.
res = sum(res_list) * 1.0 / len(len_list)
return res
def run_cleaner(self, sess, feed_dict):
sess.run([self.train_op2], feed_dict=feed_dict)
def pre_run(self, sess, feed_dict):
sess.run([self.train_op2], feed_dict=feed_dict)
def run(self, sess, feed_dict):
logits = sess.run([self.sen_logits2], feed_dict=feed_dict)
_, loss, acc_num = sess.run([self.train_op, self.loss, self.accuracy_num], feed_dict=feed_dict)
return loss, acc_num, np.concatenate(np.asarray(logits), 0)
def test_case(sess, classifier, sen_x, sen_len, sen_y):
score = classifier.run_op(sess, classifier.sen_logits2, sen_x, sen_len, sen_y)
loss = classifier.run_op(sess, classifier.loss2, sen_x, sen_len, sen_y)
acc_num = classifier.run_op(sess, classifier.accuracy_num2, sen_x, sen_len, sen_y)
y_pred = np.argmax(score, axis=1)
y_true = np.argmax(sen_y, axis=1)
p_class, r_class, f_class, support_micro = precision_recall_fscore_support(y_true=y_true, y_pred=y_pred,
labels=[0, 1], average=None)
return acc_num * 1.0 / len(sen_y), loss, f_class[0]
def run_test(sess, classifier, domain, sen_x, sen_len, sen_y):
scores = classifier.run_op(sess, classifier.sen_logits2, sen_x, sen_len, sen_y)
acc_num = classifier.run_op(sess, classifier.accuracy_num2, sen_x, sen_len, sen_y)
y_pred = np.argmax(scores, axis=1)
y_true = np.argmax(sen_y, axis=1)
p_class, r_class, f_class, support_micro = precision_recall_fscore_support(y_true=y_true, y_pred=y_pred,
labels=[0, 1], average=None)
_, id2y = load_y2id_id2y('./data/y2id.txt')
result_save_path = classifier.config.result_path + classifier.config.model + '/'
if not os.path.exists(result_save_path):
os.makedirs(result_save_path)
with open(result_save_path + domain + '_test.txt', 'w', encoding='utf-8') as fin:
fin.write('ACC: ' + str(acc_num * 1.0 / len(sen_x)) + '\t')
fin.write('P: ' + str(p_class) + '\tR: ' + str(r_class) +
'\tF1: ' + str(f_class) + '\tF1_macro: ' + str(f_class.mean()) + '\n')
for id_y in y_pred:
fin.write(id2y[id_y] + '\n')
with open(result_save_path + domain + '_true.txt', 'w', encoding='utf-8') as fin:
for id_y in y_true:
fin.write(id2y[id_y] + '\n')
print('Test. Acc = {}, P = {}, R = {}, F1 = {}, F1_macro = {}'.
format(acc_num * 1.0 / len(sen_x), p_class, r_class, f_class, f_class.mean()))
def train_run(_):
flags_ = config.FLAGS
domain = flags_.dataset # movie, laptop, restaurant
print('{} Learning start: >>>\n'.format(domain))
tf.reset_default_graph()
# os.environ['CUDA_VISIBLE_DEVICES'] = flags_.gpu
classifier = NetAbModel(domain, flags_)
gpu_config = tf.ConfigProto(intra_op_parallelism_threads=1, inter_op_parallelism_threads=1) #arguments I added
# gpu_config.gpu_options.per_process_gpu_memory_fraction = 0.85
# gpu_config.gpu_options.allow_growth = True
# gpu_config.allow_soft_placement = True # If 'True': allow cpu, if no gpu
saver = tf.train.Saver(tf.global_variables())
save_path = classifier.config.ckpt_path + classifier.config.model + '/' + domain + '/' + domain + '_ckpt'
with tf.Session(config=gpu_config) as sess:
sess.run(tf.global_variables_initializer())
best_val_acc = 0
best_val_epoch = 0
# best_test_acc = 0
training_path = os.path.join(flags_.data_path, 'TrainingSens/')
train_sen_x, train_sen_len, train_sen_y = recover_data_from_files(
training_path, 'training', domain, flags_.max_sentence_len)
val_path = os.path.join(flags_.data_path, 'ValSens/')
val_sen_x, val_sen_len, val_sen_y = recover_data_from_files(
val_path, 'validation', domain, flags_.max_sentence_len)
test_path = os.path.join(flags_.data_path, 'TestSens/')
test_sen_x, test_sen_len, test_sen_y = recover_data_from_files(
test_path, 'test', domain, flags_.max_sentence_len)
# train_sen_x, train_sen_len, train_sen_y = load_inputs_document_mongo(
# domain, 'train_noisy', classifier.word2id, flags_.max_sentence_len, flags_.max_doc_len)
# val_sen_x, val_sen_len, val_sen_y = load_inputs_document_mongo(
# domain, 'dev', classifier.word2id, flags_.max_sentence_len, flags_.max_doc_len)
# test_sen_x, test_sen_len, test_sen_y = load_inputs_document_mongo(
# domain, 'test', classifier.word2id, flags_.max_sentence_len, flags_.max_doc_len)
if classifier.config.is_train:
for epoch_i in range(flags_.n_epoch):
print('=' * 20 + 'Epoch ', epoch_i, '=' * 20)
total_loss = []
total_acc_num = []
total_num = []
if epoch_i < classifier.config.initial_epochs: # initial epochs
for step, indices in enumerate(batch_index(len(train_sen_y), flags_.batch_size, n_iter=1, is_shuffle=False), 1):
indices = list(indices)
print(train_sen_x[indices], [train_sen_y]) #-------------------------------------- I added
feed_dict = classifier.create_feed_dict(train_sen_x[indices], train_sen_len[indices],
train_sen_y[indices],
flags_.keep_prob1, flags_.keep_prob2)
classifier.pre_run(sess, feed_dict=feed_dict)
continue
for step, indices in enumerate(batch_index(len(train_sen_y), flags_.batch_size, n_iter=1, is_shuffle=False), 1):
indices = list(indices)
# if epoch_i < 10:
print("indices", train_sen_x[indices], [train_sen_y]) #-------------------------------------- I added
feed_dict = classifier.create_feed_dict(train_sen_x[indices], train_sen_len[indices],
train_sen_y[indices],
flags_.keep_prob1, flags_.keep_prob2)
loss, acc_num, logits = classifier.run(sess, feed_dict=feed_dict)
y_pred_set = np.argmax(logits, axis=1)
y_true_set = np.argmax(train_sen_y[indices], axis=1)
f_indices = np.arange(0, len(indices))
valid_indices = f_indices[y_pred_set == y_true_set]
indices_new = list(np.array(indices)[valid_indices])
print("newindices", train_sen_x[indices], [train_sen_y]) #-------------------------------------- I added
if indices_new is None:
continue
# else:
# indices_new = indices
# indices_new = indices
feed_dict = classifier.create_feed_dict(train_sen_x[indices_new], train_sen_len[indices_new],
train_sen_y[indices_new],
flags_.keep_prob1, flags_.keep_prob2)
classifier.run_cleaner(sess, feed_dict=feed_dict)
total_loss.append(loss)
total_acc_num.append(acc_num)
total_num.append(len(indices))
verbose = flags_.display_step
if step % verbose == 0:
print('[INFO] Len {}, Epoch {} - Batch {} : loss = {}, acc = {}'.format(
len(indices_new), epoch_i, step, np.mean(total_loss[-verbose:]),
sum(total_acc_num[-verbose:]) * 1.0 / sum(total_num[-verbose:])))
loss = np.mean(total_loss)
acc = sum(total_acc_num) * 1.0 / sum(total_num)
print('\n[INFO] Epoch {} : mean loss = {}, mean acc = {}'.format(epoch_i, loss, acc))
if np.isnan(loss):
raise ValueError('[Error] loss is not a number!')
# validation
val_acc, val_loss, val_f1 = test_case(sess, classifier, val_sen_x, val_sen_len, val_sen_y)
print('[INFO] val loss: {}, val acc: {}, val f1: {}'.format(val_loss, val_acc, val_f1))
# test
test_acc, test_loss, test_f1 = test_case(sess, classifier, test_sen_x, test_sen_len, test_sen_y)
print('[INFO] test loss: {}, test acc: {}, test f1: {}'.format(test_loss, test_acc, test_f1))
print('=' * 25 + ' end', '=' * 25 + '\n')
if best_val_acc < val_acc:
best_val_acc = val_acc
best_val_epoch = epoch_i
# best_test_acc = test_acc
if not os.path.exists(classifier.config.ckpt_path + classifier.config.model + '/'):
os.makedirs(classifier.config.ckpt_path + classifier.config.model + '/')
saver.save(sess, save_path=save_path)
if epoch_i - best_val_epoch > classifier.config.early_stopping:
# here early_stopping is 5 :> 'the number of early stopping epoch'
print('Normal early stop at {}!'.format(best_val_epoch))
break
print('Best val acc = {}'.format(best_val_acc))
# print('Test acc = {}'.format(best_test_acc))
best_val_epoch_save_path = classifier.config.result_path + classifier.config.model + '/'
if not os.path.exists(best_val_epoch_save_path):
os.makedirs(best_val_epoch_save_path)
with open(best_val_epoch_save_path + domain + '_bestEpoch.txt', 'w', encoding='utf-8') as fin:
fin.write('Best epoch: ' + str(best_val_epoch) + '\n')
saver.restore(sess, save_path)
print('Model restored from %s' % save_path)
# # test now
run_test(sess, classifier, domain, test_sen_x, test_sen_len, test_sen_y)
else:
saver.restore(sess, save_path)
print('Model restored from %s' % save_path)
# # test now
run_test(sess, classifier, domain, test_sen_x, test_sen_len, test_sen_y)
print('Domain {} is done..'.format(domain))
print('\nTraining complete!\n')
if __name__ == '__main__':
tf.app.run(train_run)
``` | non_defect | getting non deterministic results on getting different results on different attempts of execution i am running it on cpu os environ os environ the above lines can be ignored for cpu execution could anyone help me getting determinstic results import os import numpy as np import tensorflow as tf from sklearn metrics import precision recall fscore support import config from data helper import batch index load load load recover data from files from model nn layer import transition layer softmax layer import random tf set random seed np random seed os environ str random seed os environ os environ os environ class netabmodel object def init self domain flags filter list filter num self config flags self filter list filter list self filter num filter num placeholder self sen x batch none self sent len batch none self sen y batch none self keep none self keep none embedding self add placeholder self none self none self vocab size none self embedding none inputs self add embedding domain model self sen logits self sen self netab inputs noisy loss self loss self add loss self sen logits self accuracy self accuracy num self add accuracy self sen logits self train op self add train op self loss clean loss self self add loss self sen self self accuracy self add accuracy self sen self train self add train op self def add placeholder self self sen x batch tf placeholder tf self sent len batch tf placeholder tf self sen y batch tf placeholder tf self keep tf placeholder tf self keep tf placeholder tf def add embedding self domain if self config pre trained self load self config path domain self self load mongo domain else self load self config path domain self vocab size len self tf random uniform trainable true seed if self config embedding type static self embedding tf constant dtype tf name word embedding else self embedding tf variable dtype tf name word embedding inputs tf nn embedding lookup self embedding self sen x batch return inputs def create feed dict self sen x batch sent len batch sen y batch holder list self sen x batch self sent len batch self sen y batch self keep self keep feed list return dict zip holder list feed list cnn layer def add cnn layer self inputs inputs dim max len scope name cnn inputs tf expand dims inputs pooling outputs for i filter size in enumerate self filter list ksize conv tf contrib layers inputs inputs num outputs self filter num kernel size ksize stride padding valid activation fn tf nn relu scope conv scope name str i ksize pooling tf contrib layers max inputs conv kernel size ksize stride padding valid scope pooling scope name pooling outputs append pooling hiddens tf concat pooling outputs hiddens tf reshape hiddens return hiddens cnn layer def add noisy cnn layer self inputs inputs dim max len scope name cnn inputs tf expand dims inputs pooling outputs for i filter size in enumerate self filter list ksize conv tf contrib layers inputs inputs num outputs self filter num kernel size ksize stride padding valid activation fn tf nn relu scope conv scope name str i ksize pooling tf contrib layers max inputs conv kernel size ksize stride padding valid scope pooling scope name pooling outputs append pooling hiddens tf concat pooling outputs hiddens tf reshape hiddens return hiddens def netab self inputs print running netab inputs tf nn dropout inputs keep prob self keep seed inputs tf reshape inputs word sentence cnn outputs sen self add cnn layer inputs self config embedding dim self config max sentence len h outputs sen dim self filter num len self filter list outputs sen tf reshape outputs sen noisy cnn self add noisy cnn layer inputs self config embedding dim self config max sentence len u noisy cnn tf reshape noisy cnn fully connection clean logits softmax layer outputs sen outputs sen dim self config random base self keep self config reg self config n class sen softmax transition layer noisy cnn outputs sen dim self config reg self config random base transition layer noisy cnn outputs sen dim self config reg self config random base tf expand dims tf expand dims prob tf concat sen logits tf expand dims clean logits noisy logits tf squeeze tf matmul sen logits prob return noisy logits clean logits def add loss self sen logits loss tf nn softmax cross entropy with logits logits sen logits labels self sen y batch self sen vars var for var in tf global variables if h in var name or u in var name or in var name or in var name print self sen vars reg loss tf get collection tf graphkeys regularization losses scope sen softmax print reg loss loss tf reduce mean loss todo self config reg tf add n reg loss return loss def add accuracy self scores correct predicts tf equal tf argmax scores tf argmax self sen y batch accuracy num tf reduce sum tf cast correct predicts tf the number of correct predicting docs accuracy tf reduce mean tf cast correct predicts tf name accuracy accuracy metric result return accuracy accuracy num def add train op self doc loss new learning rate current learning rate decay rate global step decay steps global step tf variable name global step trainable false record the current step global step self lr tf train exponential decay self config lr global step self config decay steps self config decay rate staircase true the optimizer used in this work optimizer tf train adadeltaoptimizer self lr optimizer tf train adamoptimizer self lr grads global norm tf clip by global norm tf gradients doc loss self sen vars gate gradients true self config max grad norm train op optimizer apply gradients zip grads self sen vars name train op global step global step train op optimizer minimize doc loss global step global step var list self doc vars return train op def run op self sess op sen x sen len sen y res list len list for indices in batch index len sen x self config batch size n iter is shuffle false is train false feed dict self create feed dict sen x sen len sen y res sess run op feed dict feed dict res list append res len list append len indices if type res list is list if op is a list res np concatenate res list axis elif op is self accuracy num or op is self accuracy res sum res list sum all batches elif op is self sen logits or op is self sen res np concatenate np asarray res list else for los etc res sum res list len len list return res def run cleaner self sess feed dict sess run feed dict feed dict def pre run self sess feed dict sess run feed dict feed dict def run self sess feed dict logits sess run feed dict feed dict loss acc num sess run feed dict feed dict return loss acc num np concatenate np asarray logits def test case sess classifier sen x sen len sen y score classifier run op sess classifier sen sen x sen len sen y loss classifier run op sess classifier sen x sen len sen y acc num classifier run op sess classifier accuracy sen x sen len sen y y pred np argmax score axis y true np argmax sen y axis p class r class f class support micro precision recall fscore support y true y true y pred y pred labels average none return acc num len sen y loss f class def run test sess classifier domain sen x sen len sen y scores classifier run op sess classifier sen sen x sen len sen y acc num classifier run op sess classifier accuracy sen x sen len sen y y pred np argmax scores axis y true np argmax sen y axis p class r class f class support micro precision recall fscore support y true y true y pred y pred labels average none load data txt result save path classifier config result path classifier config model if not os path exists result save path os makedirs result save path with open result save path domain test txt w encoding utf as fin fin write acc str acc num len sen x t fin write p str p class tr str r class str f class macro str f class mean n for id y in y pred fin write n with open result save path domain true txt w encoding utf as fin for id y in y true fin write n print test acc p r macro format acc num len sen x p class r class f class f class mean def train run flags config flags domain flags dataset movie laptop restaurant print learning start n format domain tf reset default graph os environ flags gpu classifier netabmodel domain flags gpu config tf configproto intra op parallelism threads inter op parallelism threads arguments i added gpu config gpu options per process gpu memory fraction gpu config gpu options allow growth true gpu config allow soft placement true if true allow cpu if no gpu saver tf train saver tf global variables save path classifier config ckpt path classifier config model domain domain ckpt with tf session config gpu config as sess sess run tf global variables initializer best val acc best val epoch best test acc training path os path join flags data path trainingsens train sen x train sen len train sen y recover data from files training path training domain flags max sentence len val path os path join flags data path valsens val sen x val sen len val sen y recover data from files val path validation domain flags max sentence len test path os path join flags data path testsens test sen x test sen len test sen y recover data from files test path test domain flags max sentence len train sen x train sen len train sen y load inputs document mongo domain train noisy classifier flags max sentence len flags max doc len val sen x val sen len val sen y load inputs document mongo domain dev classifier flags max sentence len flags max doc len test sen x test sen len test sen y load inputs document mongo domain test classifier flags max sentence len flags max doc len if classifier config is train for epoch i in range flags n epoch print epoch epoch i total loss total acc num total num if epoch i classifier config initial epochs initial epochs for step indices in enumerate batch index len train sen y flags batch size n iter is shuffle false indices list indices print train sen x i added feed dict classifier create feed dict train sen x train sen len train sen y flags keep flags keep classifier pre run sess feed dict feed dict continue for step indices in enumerate batch index len train sen y flags batch size n iter is shuffle false indices list indices if epoch i print indices train sen x i added feed dict classifier create feed dict train sen x train sen len train sen y flags keep flags keep loss acc num logits classifier run sess feed dict feed dict y pred set np argmax logits axis y true set np argmax train sen y axis f indices np arange len indices valid indices f indices indices new list np array indices print newindices train sen x i added if indices new is none continue else indices new indices indices new indices feed dict classifier create feed dict train sen x train sen len train sen y flags keep flags keep classifier run cleaner sess feed dict feed dict total loss append loss total acc num append acc num total num append len indices verbose flags display step if step verbose print len epoch batch loss acc format len indices new epoch i step np mean total loss sum total acc num sum total num loss np mean total loss acc sum total acc num sum total num print n epoch mean loss mean acc format epoch i loss acc if np isnan loss raise valueerror loss is not a number validation val acc val loss val test case sess classifier val sen x val sen len val sen y print val loss val acc val format val loss val acc val test test acc test loss test test case sess classifier test sen x test sen len test sen y print test loss test acc test format test loss test acc test print end n if best val acc val acc best val acc val acc best val epoch epoch i best test acc test acc if not os path exists classifier config ckpt path classifier config model os makedirs classifier config ckpt path classifier config model saver save sess save path save path if epoch i best val epoch classifier config early stopping here early stopping is the number of early stopping epoch print normal early stop at format best val epoch break print best val acc format best val acc print test acc format best test acc best val epoch save path classifier config result path classifier config model if not os path exists best val epoch save path os makedirs best val epoch save path with open best val epoch save path domain bestepoch txt w encoding utf as fin fin write best epoch str best val epoch n saver restore sess save path print model restored from s save path test now run test sess classifier domain test sen x test sen len test sen y else saver restore sess save path print model restored from s save path test now run test sess classifier domain test sen x test sen len test sen y print domain is done format domain print ntraining complete n if name main tf app run train run | 0 |
46,187 | 13,055,864,757 | IssuesEvent | 2020-07-30 02:57:54 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | CMake libarchive detection gives false positive (Trac #709) | Incomplete Migration Migrated from Trac cmake defect | Migrated from https://code.icecube.wisc.edu/ticket/709
```json
{
"status": "closed",
"changetime": "2013-05-01T18:29:02",
"description": "On a system where libarchive exists (as a library) but has no headers (and thus is not usable), but on which liblzma is fully installed (i.e. a Mac OS 10.8 system with lzma added by the user) cmake will believe that it has found a useable copy of libarchive, but the build will later fail because archive.h doe not actually exist. This seems to happen because cmake/tools/libarchive.cmake the find_path call to search for lzma.h uses the LIBARCHIVE_INCLUDE_DIR variable. ",
"reporter": "cweaver",
"cc": "",
"resolution": "fixed",
"_ts": "1367432942000000",
"component": "cmake",
"summary": "CMake libarchive detection gives false positive",
"priority": "normal",
"keywords": "libarchive",
"time": "2013-04-29T19:09:06",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| 1.0 | CMake libarchive detection gives false positive (Trac #709) - Migrated from https://code.icecube.wisc.edu/ticket/709
```json
{
"status": "closed",
"changetime": "2013-05-01T18:29:02",
"description": "On a system where libarchive exists (as a library) but has no headers (and thus is not usable), but on which liblzma is fully installed (i.e. a Mac OS 10.8 system with lzma added by the user) cmake will believe that it has found a useable copy of libarchive, but the build will later fail because archive.h doe not actually exist. This seems to happen because cmake/tools/libarchive.cmake the find_path call to search for lzma.h uses the LIBARCHIVE_INCLUDE_DIR variable. ",
"reporter": "cweaver",
"cc": "",
"resolution": "fixed",
"_ts": "1367432942000000",
"component": "cmake",
"summary": "CMake libarchive detection gives false positive",
"priority": "normal",
"keywords": "libarchive",
"time": "2013-04-29T19:09:06",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| defect | cmake libarchive detection gives false positive trac migrated from json status closed changetime description on a system where libarchive exists as a library but has no headers and thus is not usable but on which liblzma is fully installed i e a mac os system with lzma added by the user cmake will believe that it has found a useable copy of libarchive but the build will later fail because archive h doe not actually exist this seems to happen because cmake tools libarchive cmake the find path call to search for lzma h uses the libarchive include dir variable reporter cweaver cc resolution fixed ts component cmake summary cmake libarchive detection gives false positive priority normal keywords libarchive time milestone owner nega type defect | 1 |
10,080 | 2,618,934,959 | IssuesEvent | 2015-03-03 00:01:42 | chrsmith/open-ig | https://api.github.com/repos/chrsmith/open-ig | closed | NullPointerException when moving a fighter into a new fleet | auto-migrated Component-Logic Priority-Medium Type-Defect | ```
Game version: 0.95.141
Operating System: xp 32
Java runtime version: jdk1.7.0_04
Installed using the Launcher? yes
Game language de
In the level 2 start save from Issue 764, attempt to divide the main fleet by
moving one of the fighters into a new fleet. When klicking onto "part and
finish" the exception throws.
```
Original issue reported on code.google.com by `anonymis...@arcor.de` on 27 Aug 2013 at 3:23 | 1.0 | NullPointerException when moving a fighter into a new fleet - ```
Game version: 0.95.141
Operating System: xp 32
Java runtime version: jdk1.7.0_04
Installed using the Launcher? yes
Game language de
In the level 2 start save from Issue 764, attempt to divide the main fleet by
moving one of the fighters into a new fleet. When klicking onto "part and
finish" the exception throws.
```
Original issue reported on code.google.com by `anonymis...@arcor.de` on 27 Aug 2013 at 3:23 | defect | nullpointerexception when moving a fighter into a new fleet game version operating system xp java runtime version installed using the launcher yes game language de in the level start save from issue attempt to divide the main fleet by moving one of the fighters into a new fleet when klicking onto part and finish the exception throws original issue reported on code google com by anonymis arcor de on aug at | 1 |
68,363 | 7,094,046,537 | IssuesEvent | 2018-01-12 23:28:26 | rust-lang/rust | https://api.github.com/repos/rust-lang/rust | closed | ICE when passing `()` to a function generic over `Copy` | C-bug E-needstest T-compiler | ```rust
#![feature(core_intrinsics)]
extern crate core;
use core::intrinsics;
fn write_reg<T: Copy>(reg: u32, val: T) {
unsafe { intrinsics::volatile_store(reg as *mut T, val); }
}
fn main() {
write_reg(0x10000000, ());
}
```
I was met with an internal compiler error:
`error: internal compiler error: unexpected panic`
I expected some kind of user-comprehensible error, instead.
## Meta
Rust version: `rustc 1.19.0-nightly (01951a61a 2017-05-20)`
Associated Rust Playground: https://is.gd/ZHy4Av
Backtrace:
```
thread 'rustc' panicked at 'index out of bounds: the len is 1 but the index is 1', /checkout/src/librustc_trans/intrinsic.rs:248
note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace.
stack backtrace:
0: std::sys::imp::backtrace::tracing::imp::unwind_backtrace
at /checkout/src/libstd/sys/unix/backtrace/tracing/gcc_s.rs:49
1: std::sys_common::backtrace::_print
at /checkout/src/libstd/sys_common/backtrace.rs:71
2: std::panicking::default_hook::{{closure}}
at /checkout/src/libstd/sys_common/backtrace.rs:60
at /checkout/src/libstd/panicking.rs:355
3: std::panicking::default_hook
at /checkout/src/libstd/panicking.rs:365
4: std::panicking::rust_panic_with_hook
at /checkout/src/libstd/panicking.rs:549
5: std::panicking::begin_panic
at /checkout/src/libstd/panicking.rs:511
6: std::panicking::begin_panic_fmt
at /checkout/src/libstd/panicking.rs:495
7: rust_begin_unwind
at /checkout/src/libstd/panicking.rs:471
8: core::panicking::panic_fmt
at /checkout/src/libcore/panicking.rs:69
9: core::panicking::panic_bounds_check
at /checkout/src/libcore/panicking.rs:56
10: rustc_trans::intrinsic::trans_intrinsic_call
11: rustc_trans::mir::block::<impl rustc_trans::mir::MirContext<'a, 'tcx>>::trans_block
12: rustc_trans::mir::trans_mir
13: rustc_trans::trans_item::TransItem::define
14: rustc_trans::base::trans_crate
15: rustc_driver::driver::phase_4_translate_to_llvm
16: rustc_driver::driver::compile_input::{{closure}}
17: rustc_driver::driver::phase_3_run_analysis_passes::{{closure}}
18: rustc_driver::driver::phase_3_run_analysis_passes
19: rustc_driver::driver::compile_input
20: rustc_driver::run_compiler
``` | 1.0 | ICE when passing `()` to a function generic over `Copy` - ```rust
#![feature(core_intrinsics)]
extern crate core;
use core::intrinsics;
fn write_reg<T: Copy>(reg: u32, val: T) {
unsafe { intrinsics::volatile_store(reg as *mut T, val); }
}
fn main() {
write_reg(0x10000000, ());
}
```
I was met with an internal compiler error:
`error: internal compiler error: unexpected panic`
I expected some kind of user-comprehensible error, instead.
## Meta
Rust version: `rustc 1.19.0-nightly (01951a61a 2017-05-20)`
Associated Rust Playground: https://is.gd/ZHy4Av
Backtrace:
```
thread 'rustc' panicked at 'index out of bounds: the len is 1 but the index is 1', /checkout/src/librustc_trans/intrinsic.rs:248
note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace.
stack backtrace:
0: std::sys::imp::backtrace::tracing::imp::unwind_backtrace
at /checkout/src/libstd/sys/unix/backtrace/tracing/gcc_s.rs:49
1: std::sys_common::backtrace::_print
at /checkout/src/libstd/sys_common/backtrace.rs:71
2: std::panicking::default_hook::{{closure}}
at /checkout/src/libstd/sys_common/backtrace.rs:60
at /checkout/src/libstd/panicking.rs:355
3: std::panicking::default_hook
at /checkout/src/libstd/panicking.rs:365
4: std::panicking::rust_panic_with_hook
at /checkout/src/libstd/panicking.rs:549
5: std::panicking::begin_panic
at /checkout/src/libstd/panicking.rs:511
6: std::panicking::begin_panic_fmt
at /checkout/src/libstd/panicking.rs:495
7: rust_begin_unwind
at /checkout/src/libstd/panicking.rs:471
8: core::panicking::panic_fmt
at /checkout/src/libcore/panicking.rs:69
9: core::panicking::panic_bounds_check
at /checkout/src/libcore/panicking.rs:56
10: rustc_trans::intrinsic::trans_intrinsic_call
11: rustc_trans::mir::block::<impl rustc_trans::mir::MirContext<'a, 'tcx>>::trans_block
12: rustc_trans::mir::trans_mir
13: rustc_trans::trans_item::TransItem::define
14: rustc_trans::base::trans_crate
15: rustc_driver::driver::phase_4_translate_to_llvm
16: rustc_driver::driver::compile_input::{{closure}}
17: rustc_driver::driver::phase_3_run_analysis_passes::{{closure}}
18: rustc_driver::driver::phase_3_run_analysis_passes
19: rustc_driver::driver::compile_input
20: rustc_driver::run_compiler
``` | non_defect | ice when passing to a function generic over copy rust extern crate core use core intrinsics fn write reg reg val t unsafe intrinsics volatile store reg as mut t val fn main write reg i was met with an internal compiler error error internal compiler error unexpected panic i expected some kind of user comprehensible error instead meta rust version rustc nightly associated rust playground backtrace thread rustc panicked at index out of bounds the len is but the index is checkout src librustc trans intrinsic rs note some details are omitted run with rust backtrace full for a verbose backtrace stack backtrace std sys imp backtrace tracing imp unwind backtrace at checkout src libstd sys unix backtrace tracing gcc s rs std sys common backtrace print at checkout src libstd sys common backtrace rs std panicking default hook closure at checkout src libstd sys common backtrace rs at checkout src libstd panicking rs std panicking default hook at checkout src libstd panicking rs std panicking rust panic with hook at checkout src libstd panicking rs std panicking begin panic at checkout src libstd panicking rs std panicking begin panic fmt at checkout src libstd panicking rs rust begin unwind at checkout src libstd panicking rs core panicking panic fmt at checkout src libcore panicking rs core panicking panic bounds check at checkout src libcore panicking rs rustc trans intrinsic trans intrinsic call rustc trans mir block trans block rustc trans mir trans mir rustc trans trans item transitem define rustc trans base trans crate rustc driver driver phase translate to llvm rustc driver driver compile input closure rustc driver driver phase run analysis passes closure rustc driver driver phase run analysis passes rustc driver driver compile input rustc driver run compiler | 0 |
123,773 | 16,536,250,305 | IssuesEvent | 2021-05-27 12:13:01 | VNG-Realisatie/API-Kennisbank | https://api.github.com/repos/VNG-Realisatie/API-Kennisbank | opened | Technische definities van properties alleen opnemen voor zover dat noodzakelijk is voor gebruik | Design Rule | Meestal volstaat het om in de definitie van properties van de response de propertynaam, type, format, description en example op te nemen. We gebruiken daar niet de technische definities pattern, minimum, maximum, minLength, maxLength, minItems en required, tenzij....
De technische definitie title gebruiken we alleen wanneer de propertynaam afwijkt van de naam in het informatiemodel (en het informatiemodel bekend is bij veel gebruikers) zie #83.
De technische definitie enum gebruiken we wanneer (sommige/mogelijke) gebruikers de mogelijke waarden gebruiken in code/algoritmes en daarom moeten weten welke mogelijke waarden er zijn. | 1.0 | Technische definities van properties alleen opnemen voor zover dat noodzakelijk is voor gebruik - Meestal volstaat het om in de definitie van properties van de response de propertynaam, type, format, description en example op te nemen. We gebruiken daar niet de technische definities pattern, minimum, maximum, minLength, maxLength, minItems en required, tenzij....
De technische definitie title gebruiken we alleen wanneer de propertynaam afwijkt van de naam in het informatiemodel (en het informatiemodel bekend is bij veel gebruikers) zie #83.
De technische definitie enum gebruiken we wanneer (sommige/mogelijke) gebruikers de mogelijke waarden gebruiken in code/algoritmes en daarom moeten weten welke mogelijke waarden er zijn. | non_defect | technische definities van properties alleen opnemen voor zover dat noodzakelijk is voor gebruik meestal volstaat het om in de definitie van properties van de response de propertynaam type format description en example op te nemen we gebruiken daar niet de technische definities pattern minimum maximum minlength maxlength minitems en required tenzij de technische definitie title gebruiken we alleen wanneer de propertynaam afwijkt van de naam in het informatiemodel en het informatiemodel bekend is bij veel gebruikers zie de technische definitie enum gebruiken we wanneer sommige mogelijke gebruikers de mogelijke waarden gebruiken in code algoritmes en daarom moeten weten welke mogelijke waarden er zijn | 0 |
54,929 | 14,046,180,287 | IssuesEvent | 2020-11-02 03:54:19 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | closed | disk partition information disappeared after power accident | Status: Triage Needed Type: Defect | ### System information
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | 18.04.5
Linux Kernel | 4.15.0-122-generic
Architecture |
ZFS Version | 0.7.5-1ubuntu16.10
SPL Version | 0.7.5-1ubuntu2.2
### Describe the problem you're observing
I have 5 machines installed with zfs, after an power incident, I found the pool in 2 machine disappeared, even the partition table disappeared
### Describe how to reproduce the problem
I did't try to reproduce this issue
### Include any warning/errors/backtraces from the system logs
message from zfs-import-cache service:
```
Oct 30 02:38:34 K-GXGK247 systemd[1]: Starting Import ZFS pools by cache file...
Oct 30 02:38:36 K-GXGK247 zpool[1627]: cannot import 'data': no such pool or dataset
Oct 30 02:38:36 K-GXGK247 zpool[1627]: Destroy and re-create the pool from
Oct 30 02:38:36 K-GXGK247 zpool[1627]: a backup source.
Oct 30 02:38:36 K-GXGK247 systemd[1]: zfs-import-cache.service: Main process exited, code=exited, status=1/FAILURE
Oct 30 02:38:36 K-GXGK247 systemd[1]: zfs-import-cache.service: Failed with result 'exit-code'.
Oct 30 02:38:36 K-GXGK247 systemd[1]: Failed to start Import ZFS pools by cache file.
this is an example how log text should be marked (wrap it with ```)
```
output of lsblk:
```
root@K-GXGK247:~# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 7.3T 0 disk
sdb 8:16 0 7.3T 0 disk
sdc 8:32 0 7.3T 0 disk
sdd 8:48 0 7.3T 0 disk
sde 8:64 0 7.3T 0 disk
sdf 8:80 0 7.3T 0 disk
sdg 8:96 0 7.3T 0 disk
sdh 8:112 0 7.3T 0 disk
sdi 8:128 0 7.3T 0 disk
sdj 8:144 0 7.3T 0 disk
sdk 8:160 0 7.3T 0 disk
sdl 8:176 0 7.3T 0 disk
sdm 8:192 0 7.3T 0 disk
sdn 8:208 0 7.3T 0 disk
sdo 8:224 0 7.3T 0 disk
sdp 8:240 0 7.3T 0 disk
sdq 65:0 0 7.3T 0 disk
sdr 65:16 0 7.3T 0 disk
sds 65:32 0 7.3T 0 disk
sdt 65:48 0 7.3T 0 disk
sdu 65:64 0 7.3T 0 disk
sdv 65:80 0 7.3T 0 disk
sdw 65:96 0 7.3T 0 disk
sdx 65:112 0 7.3T 0 disk
nvme0n1 259:0 0 1.9T 0 disk
nvme0n1p1 259:1 0 512M 0 part /boot/efi
nvme0n1p2 259:2 0 1.9T 0 part /
```
| 1.0 | disk partition information disappeared after power accident - ### System information
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | 18.04.5
Linux Kernel | 4.15.0-122-generic
Architecture |
ZFS Version | 0.7.5-1ubuntu16.10
SPL Version | 0.7.5-1ubuntu2.2
### Describe the problem you're observing
I have 5 machines installed with zfs, after an power incident, I found the pool in 2 machine disappeared, even the partition table disappeared
### Describe how to reproduce the problem
I did't try to reproduce this issue
### Include any warning/errors/backtraces from the system logs
message from zfs-import-cache service:
```
Oct 30 02:38:34 K-GXGK247 systemd[1]: Starting Import ZFS pools by cache file...
Oct 30 02:38:36 K-GXGK247 zpool[1627]: cannot import 'data': no such pool or dataset
Oct 30 02:38:36 K-GXGK247 zpool[1627]: Destroy and re-create the pool from
Oct 30 02:38:36 K-GXGK247 zpool[1627]: a backup source.
Oct 30 02:38:36 K-GXGK247 systemd[1]: zfs-import-cache.service: Main process exited, code=exited, status=1/FAILURE
Oct 30 02:38:36 K-GXGK247 systemd[1]: zfs-import-cache.service: Failed with result 'exit-code'.
Oct 30 02:38:36 K-GXGK247 systemd[1]: Failed to start Import ZFS pools by cache file.
this is an example how log text should be marked (wrap it with ```)
```
output of lsblk:
```
root@K-GXGK247:~# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 7.3T 0 disk
sdb 8:16 0 7.3T 0 disk
sdc 8:32 0 7.3T 0 disk
sdd 8:48 0 7.3T 0 disk
sde 8:64 0 7.3T 0 disk
sdf 8:80 0 7.3T 0 disk
sdg 8:96 0 7.3T 0 disk
sdh 8:112 0 7.3T 0 disk
sdi 8:128 0 7.3T 0 disk
sdj 8:144 0 7.3T 0 disk
sdk 8:160 0 7.3T 0 disk
sdl 8:176 0 7.3T 0 disk
sdm 8:192 0 7.3T 0 disk
sdn 8:208 0 7.3T 0 disk
sdo 8:224 0 7.3T 0 disk
sdp 8:240 0 7.3T 0 disk
sdq 65:0 0 7.3T 0 disk
sdr 65:16 0 7.3T 0 disk
sds 65:32 0 7.3T 0 disk
sdt 65:48 0 7.3T 0 disk
sdu 65:64 0 7.3T 0 disk
sdv 65:80 0 7.3T 0 disk
sdw 65:96 0 7.3T 0 disk
sdx 65:112 0 7.3T 0 disk
nvme0n1 259:0 0 1.9T 0 disk
nvme0n1p1 259:1 0 512M 0 part /boot/efi
nvme0n1p2 259:2 0 1.9T 0 part /
```
| defect | disk partition information disappeared after power accident system information type version name distribution name ubuntu distribution version linux kernel generic architecture zfs version spl version describe the problem you re observing i have machines installed with zfs after an power incident i found the pool in machine disappeared even the partition table disappeared describe how to reproduce the problem i did t try to reproduce this issue include any warning errors backtraces from the system logs message from zfs import cache service oct k systemd starting import zfs pools by cache file oct k zpool cannot import data no such pool or dataset oct k zpool destroy and re create the pool from oct k zpool a backup source oct k systemd zfs import cache service main process exited code exited status failure oct k systemd zfs import cache service failed with result exit code oct k systemd failed to start import zfs pools by cache file this is an example how log text should be marked wrap it with output of lsblk root k lsblk name maj min rm size ro type mountpoint sda disk sdb disk sdc disk sdd disk sde disk sdf disk sdg disk sdh disk sdi disk sdj disk sdk disk sdl disk sdm disk sdn disk sdo disk sdp disk sdq disk sdr disk sds disk sdt disk sdu disk sdv disk sdw disk sdx disk disk part boot efi part | 1 |
29,300 | 5,638,508,420 | IssuesEvent | 2017-04-06 12:10:10 | BOINC/boinc | https://api.github.com/repos/BOINC/boinc | closed | Increase in CPU consumption of BOINC core client when network (or project) is unavailable | C: Client - Scheduler Policy P: Major T: Defect | **Reported by Raistmer on 26 May 38695851 20:26 UTC**
When whole network or single project down and many results waiting to be sent BOINC make retry to send for each result separately that leads to almost constant atempts to send one of results. This leads to increase in CPU consumption by core client.
BOINC could check once per some period defined by settings if network/particular project site available and only after success check retry all results sending for this project/all projects.
Migrated-From: http://boinc.berkeley.edu/trac/ticket/734
| 1.0 | Increase in CPU consumption of BOINC core client when network (or project) is unavailable - **Reported by Raistmer on 26 May 38695851 20:26 UTC**
When whole network or single project down and many results waiting to be sent BOINC make retry to send for each result separately that leads to almost constant atempts to send one of results. This leads to increase in CPU consumption by core client.
BOINC could check once per some period defined by settings if network/particular project site available and only after success check retry all results sending for this project/all projects.
Migrated-From: http://boinc.berkeley.edu/trac/ticket/734
| defect | increase in cpu consumption of boinc core client when network or project is unavailable reported by raistmer on may utc when whole network or single project down and many results waiting to be sent boinc make retry to send for each result separately that leads to almost constant atempts to send one of results this leads to increase in cpu consumption by core client boinc could check once per some period defined by settings if network particular project site available and only after success check retry all results sending for this project all projects migrated from | 1 |
399,141 | 11,743,529,826 | IssuesEvent | 2020-03-12 04:51:56 | AY1920S2-CS2103-T09-2/main | https://api.github.com/repos/AY1920S2-CS2103-T09-2/main | closed | As a user, I can view all transactions in the system. | priority.High status.Ongoing type.Story | As a user, I can view all transactions in the system, so I can keep track of how my business is doing. | 1.0 | As a user, I can view all transactions in the system. - As a user, I can view all transactions in the system, so I can keep track of how my business is doing. | non_defect | as a user i can view all transactions in the system as a user i can view all transactions in the system so i can keep track of how my business is doing | 0 |
53,754 | 6,756,695,780 | IssuesEvent | 2017-10-24 08:09:57 | perryyui/SOFT20091 | https://api.github.com/repos/perryyui/SOFT20091 | opened | Suppose that queue is an object of type queueType<string> of size 50, queue contains 20 elements, and the index of the first element queue is 35. | Coding Design | a. What is the index of the last element queue?
b. After removing the next element from queue, what is the index of the
first element?
c. Write the expression that returns true if queue is nonempty, false
otherwise.
d. Write the statement that inserts “programming” into queue. What is
the index of the last element after the insertion operation. | 1.0 | Suppose that queue is an object of type queueType<string> of size 50, queue contains 20 elements, and the index of the first element queue is 35. - a. What is the index of the last element queue?
b. After removing the next element from queue, what is the index of the
first element?
c. Write the expression that returns true if queue is nonempty, false
otherwise.
d. Write the statement that inserts “programming” into queue. What is
the index of the last element after the insertion operation. | non_defect | suppose that queue is an object of type queuetype of size queue contains elements and the index of the first element queue is a what is the index of the last element queue b after removing the next element from queue what is the index of the first element c write the expression that returns true if queue is nonempty false otherwise d write the statement that inserts “programming” into queue what is the index of the last element after the insertion operation | 0 |
511,547 | 14,876,153,927 | IssuesEvent | 2021-01-20 00:14:52 | courselab/ttsnake | https://api.github.com/repos/courselab/ttsnake | closed | Increment snake size when head meets energy blocks. | +priority enhance | This depends on #5 and #3.
The snake grows by adding a body block right after the head (how about that?). | 1.0 | Increment snake size when head meets energy blocks. - This depends on #5 and #3.
The snake grows by adding a body block right after the head (how about that?). | non_defect | increment snake size when head meets energy blocks this depends on and the snake grows by adding a body block right after the head how about that | 0 |
91,713 | 15,856,579,816 | IssuesEvent | 2021-04-08 02:41:05 | cluesblues/git-travis | https://api.github.com/repos/cluesblues/git-travis | opened | CVE-2015-9251 (Medium) detected in jquery-1.3.2.min.js | security vulnerability | ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.3.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.3.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.3.2/jquery.min.js</a></p>
<p>Path to dependency file: git-travis/node_modules/underscore.string/test/test_underscore/temp_tests.html</p>
<p>Path to vulnerable library: git-travis/node_modules/underscore.string/test/test_underscore/vendor/jquery.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.3.2.min.js** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2015-9251 (Medium) detected in jquery-1.3.2.min.js - ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.3.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.3.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.3.2/jquery.min.js</a></p>
<p>Path to dependency file: git-travis/node_modules/underscore.string/test/test_underscore/temp_tests.html</p>
<p>Path to vulnerable library: git-travis/node_modules/underscore.string/test/test_underscore/vendor/jquery.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.3.2.min.js** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file git travis node modules underscore string test test underscore temp tests html path to vulnerable library git travis node modules underscore string test test underscore vendor jquery js dependency hierarchy x jquery min js vulnerable library found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource | 0 |
78,765 | 27,750,719,721 | IssuesEvent | 2023-03-15 20:29:59 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | Datepicker: ClassCastException when using a converter | :lady_beetle: defect | ### Describe the bug
The issue happens when:
- the value is not a Date but something else (eg. an Integer) and an appropriate converter is defined
- an invalid value is submitted
### Reproducer
[primefaces-test.zip](https://github.com/primefaces/primefaces/files/10932589/primefaces-test.zip)
1. `mvn install`
2. `mvn jetty:run`
3. navigate to http://localhost:8080/primefaces-test/test.xhtml
4. change something in the datepicker to make it invalid (eg replace a number with a letter)
5. click the button
Result:
```
java.lang.ClassCastException: class java.time.LocalDate cannot be cast to class java.util.Optional (java.time.LocalDate and java.util.Optional are in module java.base of loader 'bootstrap')
at org.primefaces.test.OptionalConverter.getAsString(OptionalConverter.java:16)
at org.primefaces.util.CalendarUtils.getValue(CalendarUtils.java:290)
at org.primefaces.util.CalendarUtils.getValueAsString(CalendarUtils.java:281)
at org.primefaces.util.CalendarUtils.getValueAsString(CalendarUtils.java:248)
at org.primefaces.component.datepicker.DatePickerRenderer.encodeScript(DatePickerRenderer.java:167)
at org.primefaces.component.calendar.BaseCalendarRenderer.encodeEnd(BaseCalendarRenderer.java:87)
at org.primefaces.component.datepicker.DatePickerRenderer.encodeEnd(DatePickerRenderer.java:92)
...
```
### Expected behavior
No exception
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.3.19
### Java version
Eclipse Adoptium jdk-11.0.17.8-hotspot
### Browser(s)
_No response_ | 1.0 | Datepicker: ClassCastException when using a converter - ### Describe the bug
The issue happens when:
- the value is not a Date but something else (eg. an Integer) and an appropriate converter is defined
- an invalid value is submitted
### Reproducer
[primefaces-test.zip](https://github.com/primefaces/primefaces/files/10932589/primefaces-test.zip)
1. `mvn install`
2. `mvn jetty:run`
3. navigate to http://localhost:8080/primefaces-test/test.xhtml
4. change something in the datepicker to make it invalid (eg replace a number with a letter)
5. click the button
Result:
```
java.lang.ClassCastException: class java.time.LocalDate cannot be cast to class java.util.Optional (java.time.LocalDate and java.util.Optional are in module java.base of loader 'bootstrap')
at org.primefaces.test.OptionalConverter.getAsString(OptionalConverter.java:16)
at org.primefaces.util.CalendarUtils.getValue(CalendarUtils.java:290)
at org.primefaces.util.CalendarUtils.getValueAsString(CalendarUtils.java:281)
at org.primefaces.util.CalendarUtils.getValueAsString(CalendarUtils.java:248)
at org.primefaces.component.datepicker.DatePickerRenderer.encodeScript(DatePickerRenderer.java:167)
at org.primefaces.component.calendar.BaseCalendarRenderer.encodeEnd(BaseCalendarRenderer.java:87)
at org.primefaces.component.datepicker.DatePickerRenderer.encodeEnd(DatePickerRenderer.java:92)
...
```
### Expected behavior
No exception
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.3.19
### Java version
Eclipse Adoptium jdk-11.0.17.8-hotspot
### Browser(s)
_No response_ | defect | datepicker classcastexception when using a converter describe the bug the issue happens when the value is not a date but something else eg an integer and an appropriate converter is defined an invalid value is submitted reproducer mvn install mvn jetty run navigate to change something in the datepicker to make it invalid eg replace a number with a letter click the button result java lang classcastexception class java time localdate cannot be cast to class java util optional java time localdate and java util optional are in module java base of loader bootstrap at org primefaces test optionalconverter getasstring optionalconverter java at org primefaces util calendarutils getvalue calendarutils java at org primefaces util calendarutils getvalueasstring calendarutils java at org primefaces util calendarutils getvalueasstring calendarutils java at org primefaces component datepicker datepickerrenderer encodescript datepickerrenderer java at org primefaces component calendar basecalendarrenderer encodeend basecalendarrenderer java at org primefaces component datepicker datepickerrenderer encodeend datepickerrenderer java expected behavior no exception primefaces edition community primefaces version theme no response jsf implementation mojarra jsf version java version eclipse adoptium jdk hotspot browser s no response | 1 |
115,880 | 17,346,988,172 | IssuesEvent | 2021-07-29 01:16:07 | Pr0methean/godot | https://api.github.com/repos/Pr0methean/godot | opened | CVE-2019-16910 (Medium) detected in godot3.1-stable | security vulnerability | ## CVE-2019-16910 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>godot3.1-stable</b></p></summary>
<p>
<p>Godot Engine ��� Multi-platform 2D and 3D game engine</p>
<p>Library home page: <a href=https://github.com/godotengine/godot.git>https://github.com/godotengine/godot.git</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (5)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>godot/thirdparty/mbedtls/library/hmac_drbg.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>godot/thirdparty/mbedtls/library/hmac_drbg.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>godot/thirdparty/mbedtls/library/des.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>godot/thirdparty/mbedtls/library/sha512.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>godot/thirdparty/mbedtls/library/des.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Arm Mbed TLS before 2.19.0 and Arm Mbed Crypto before 2.0.0, when deterministic ECDSA is enabled, use an RNG with insufficient entropy for blinding, which might allow an attacker to recover a private key via side-channel attacks if a victim signs the same message many times. (For Mbed TLS, the fix is also available in versions 2.7.12 and 2.16.3.)
<p>Publish Date: 2019-09-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16910>CVE-2019-16910</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16910">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16910</a></p>
<p>Release Date: 2019-09-26</p>
<p>Fix Resolution: Mbed TLS-2.7.12,Mbed-2.16.3,Mbed TLS-2.19.0;Mbed Crypto-2.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-16910 (Medium) detected in godot3.1-stable - ## CVE-2019-16910 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>godot3.1-stable</b></p></summary>
<p>
<p>Godot Engine ��� Multi-platform 2D and 3D game engine</p>
<p>Library home page: <a href=https://github.com/godotengine/godot.git>https://github.com/godotengine/godot.git</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (5)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>godot/thirdparty/mbedtls/library/hmac_drbg.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>godot/thirdparty/mbedtls/library/hmac_drbg.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>godot/thirdparty/mbedtls/library/des.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>godot/thirdparty/mbedtls/library/sha512.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>godot/thirdparty/mbedtls/library/des.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Arm Mbed TLS before 2.19.0 and Arm Mbed Crypto before 2.0.0, when deterministic ECDSA is enabled, use an RNG with insufficient entropy for blinding, which might allow an attacker to recover a private key via side-channel attacks if a victim signs the same message many times. (For Mbed TLS, the fix is also available in versions 2.7.12 and 2.16.3.)
<p>Publish Date: 2019-09-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16910>CVE-2019-16910</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16910">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16910</a></p>
<p>Release Date: 2019-09-26</p>
<p>Fix Resolution: Mbed TLS-2.7.12,Mbed-2.16.3,Mbed TLS-2.19.0;Mbed Crypto-2.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in stable cve medium severity vulnerability vulnerable library stable godot engine ��� multi platform and game engine library home page a href vulnerable source files godot thirdparty mbedtls library hmac drbg c godot thirdparty mbedtls library hmac drbg c godot thirdparty mbedtls library des c godot thirdparty mbedtls library c godot thirdparty mbedtls library des c vulnerability details arm mbed tls before and arm mbed crypto before when deterministic ecdsa is enabled use an rng with insufficient entropy for blinding which might allow an attacker to recover a private key via side channel attacks if a victim signs the same message many times for mbed tls the fix is also available in versions and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution mbed tls mbed mbed tls mbed crypto step up your open source security game with whitesource | 0 |
778,276 | 27,309,510,024 | IssuesEvent | 2023-02-24 11:00:03 | farooqkz/chooj | https://api.github.com/repos/farooqkz/chooj | closed | Reduce bundle size | bug high priority Medium complexity level | Currently, parcel is the build system chooj uses. When targeting FF48, parcel tries to create a bundle which works on FF48 newer browsers. FF48 does not support JS modules but most newer browsers do. So parcel creates two sets of JS assets for module and nomodule. And the module JS assets never get used in KaiOS 2.x.
Related issue:
https://github.com/parcel-bundler/parcel/issues/8783 | 1.0 | Reduce bundle size - Currently, parcel is the build system chooj uses. When targeting FF48, parcel tries to create a bundle which works on FF48 newer browsers. FF48 does not support JS modules but most newer browsers do. So parcel creates two sets of JS assets for module and nomodule. And the module JS assets never get used in KaiOS 2.x.
Related issue:
https://github.com/parcel-bundler/parcel/issues/8783 | non_defect | reduce bundle size currently parcel is the build system chooj uses when targeting parcel tries to create a bundle which works on newer browsers does not support js modules but most newer browsers do so parcel creates two sets of js assets for module and nomodule and the module js assets never get used in kaios x related issue | 0 |
13,638 | 2,773,971,239 | IssuesEvent | 2015-05-04 02:31:22 | yvrkarthik/UsageStats | https://api.github.com/repos/yvrkarthik/UsageStats | closed | If the Month is changed back to an existing value, the records are over written | priority: normal status: active type: defect value: functionality | If the month value gets changed to an old value the old month values are being over written. | 1.0 | If the Month is changed back to an existing value, the records are over written - If the month value gets changed to an old value the old month values are being over written. | defect | if the month is changed back to an existing value the records are over written if the month value gets changed to an old value the old month values are being over written | 1 |
43,394 | 11,697,216,282 | IssuesEvent | 2020-03-06 11:19:47 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | returning() Method generates wrong statement | T: Defect | ### Expected behavior and actual behavior:
Hi,
if we use the returning() Method after declaring an insert the generated sql statment ist wrong.
### Steps to reproduce the problem (if possible, create an MCVE: https://github.com/jOOQ/jOOQ-mcve):
This is our code.
```java
IncentiveGutscheincodes igc = new IncentiveGutscheincodes().as("igc");
final InsertSetMoreStep<IncentiveGutscheincodesRecord> insert =
create.insertInto(igc)
.set(igc.INGC_ID, Sequences.SEQ_INGC.nextval().coerce(Long.class))
.set(igc.INGC_ERSTELLT_VON, DSL.currentUser())
.set(igc.INGC_ERSTELLUNGSZEITPUNKT, LocalDateTime.now());
LOG.info("correct insert statement: {} ", insert.getSQL(ParamType.INLINED));
final InsertResultStep<IncentiveGutscheincodesRecord> insertWithReturn = insert.returning();
LOG.info("broken insertWithReturn statement: {} ", insertWithReturn.getSQL(ParamType.INLINED));
//-> java.sql.SQLSyntaxErrorException: ORA-04043: Objekt TEST."IGC" ist nicht vorhanden
final IncentiveGutscheincodesRecord incentiveGutscheincodesRecord = insertWithReturn.fetchOne();
```
These are the generated statments
`correct insert statement: insert into "TEST"."INCENTIVE_GUTSCHEINCODES" "igc" ("INGC_ID", "INGC_ERSTELLT_VON", "INGC_ERSTELLUNGSZEITPUNKT") values ("TEST"."SEQ_INGC".nextval, user, cast(timestamp '2020-03-06 12:17:15.1303232' as date)) `
`broken insertWithReturn statement: insert into TEST."igc" ("INGC_ID", "INGC_ERSTELLT_VON", "INGC_ERSTELLUNGSZEITPUNKT") values ("TEST"."SEQ_INGC".nextval, user, cast(timestamp '2020-03-06 12:17:15.1303232' as date)) `
As you can see the table name is replaced with only the alias so the statment fails..
### Versions:
- jOOQ: 3.12.x & 3.13.x
- Java: 8 & 11
- Database (include vendor): Oracle 18
- OS : Windows / Linux
- JDBC Driver (include name if inofficial driver):
| 1.0 | returning() Method generates wrong statement - ### Expected behavior and actual behavior:
Hi,
if we use the returning() Method after declaring an insert the generated sql statment ist wrong.
### Steps to reproduce the problem (if possible, create an MCVE: https://github.com/jOOQ/jOOQ-mcve):
This is our code.
```java
IncentiveGutscheincodes igc = new IncentiveGutscheincodes().as("igc");
final InsertSetMoreStep<IncentiveGutscheincodesRecord> insert =
create.insertInto(igc)
.set(igc.INGC_ID, Sequences.SEQ_INGC.nextval().coerce(Long.class))
.set(igc.INGC_ERSTELLT_VON, DSL.currentUser())
.set(igc.INGC_ERSTELLUNGSZEITPUNKT, LocalDateTime.now());
LOG.info("correct insert statement: {} ", insert.getSQL(ParamType.INLINED));
final InsertResultStep<IncentiveGutscheincodesRecord> insertWithReturn = insert.returning();
LOG.info("broken insertWithReturn statement: {} ", insertWithReturn.getSQL(ParamType.INLINED));
//-> java.sql.SQLSyntaxErrorException: ORA-04043: Objekt TEST."IGC" ist nicht vorhanden
final IncentiveGutscheincodesRecord incentiveGutscheincodesRecord = insertWithReturn.fetchOne();
```
These are the generated statments
`correct insert statement: insert into "TEST"."INCENTIVE_GUTSCHEINCODES" "igc" ("INGC_ID", "INGC_ERSTELLT_VON", "INGC_ERSTELLUNGSZEITPUNKT") values ("TEST"."SEQ_INGC".nextval, user, cast(timestamp '2020-03-06 12:17:15.1303232' as date)) `
`broken insertWithReturn statement: insert into TEST."igc" ("INGC_ID", "INGC_ERSTELLT_VON", "INGC_ERSTELLUNGSZEITPUNKT") values ("TEST"."SEQ_INGC".nextval, user, cast(timestamp '2020-03-06 12:17:15.1303232' as date)) `
As you can see the table name is replaced with only the alias so the statment fails..
### Versions:
- jOOQ: 3.12.x & 3.13.x
- Java: 8 & 11
- Database (include vendor): Oracle 18
- OS : Windows / Linux
- JDBC Driver (include name if inofficial driver):
| defect | returning method generates wrong statement expected behavior and actual behavior hi if we use the returning method after declaring an insert the generated sql statment ist wrong steps to reproduce the problem if possible create an mcve this is our code java incentivegutscheincodes igc new incentivegutscheincodes as igc final insertsetmorestep insert create insertinto igc set igc ingc id sequences seq ingc nextval coerce long class set igc ingc erstellt von dsl currentuser set igc ingc erstellungszeitpunkt localdatetime now log info correct insert statement insert getsql paramtype inlined final insertresultstep insertwithreturn insert returning log info broken insertwithreturn statement insertwithreturn getsql paramtype inlined java sql sqlsyntaxerrorexception ora objekt test igc ist nicht vorhanden final incentivegutscheincodesrecord incentivegutscheincodesrecord insertwithreturn fetchone these are the generated statments correct insert statement insert into test incentive gutscheincodes igc ingc id ingc erstellt von ingc erstellungszeitpunkt values test seq ingc nextval user cast timestamp as date broken insertwithreturn statement insert into test igc ingc id ingc erstellt von ingc erstellungszeitpunkt values test seq ingc nextval user cast timestamp as date as you can see the table name is replaced with only the alias so the statment fails versions jooq x x java database include vendor oracle os windows linux jdbc driver include name if inofficial driver | 1 |
71,124 | 23,459,079,743 | IssuesEvent | 2022-08-16 11:34:47 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Incorrect handling of blocking calls in ProcessorSupplier.init() [HZ-1204] | Type: Defect Team: Core Source: Internal Module: Jet to-jira | According to its contract, the `init()` and `close()` methods in `Processor`, `ProcessorSupplier` and `ProcessorMetaSupplier` are allowed to block, even for cooperative processors. It's explicitly written in `Processor` javadoc, but it's assumed in the docs, as the methods are used for resource initialization and cleanup. We also use it this way in our code, e.g. for opening the HZ client for remote map reader.
However, if the methods are blocking for a long time, they can easily starve thread pools:
- the coordinator thread pool, with four hard-coded threads, on master member. When all threads are blocked, most operations in the `Job` class are blocked too
- the generic operation thread pool, with configurable size. When all threads in this pool are blocked, it blocks many non-Jet operations, such as `IMap.values()`.
There are 2 possible solutions:
1. we will offload these methods to their own pool (shared with "cached" threads) so that they don't block anything else other than the initialization of other jobs. The handling of `InitExeutionOperation` will need to become async. This will create extra overhead for most processors that don't do any blocking calls in `init` or `close` and will slow down the job deploy+cleanup time. We can guard with methods akin to `Processor.closeIsCooperative()`, but that's an extra burden for the developer and it's easy to omit.
2. we can change the contract of these methods to specify that they can block only for a short time. If they need to block longer, they should move the initialization later and do it lazily in `process()`. This includes fixing all our processors to comply.
Below are tests that can be added to `ExecutionLifecycleTest`, should we implement the 1st approach. Today these tests fail.
```java
@Test
public void when_pmsInitBlocks_then_otherJobsNotBlocked() throws ExecutionException, InterruptedException {
// Given
DAG dagBlocking = new DAG().vertex(new Vertex("test",
new MockPMS(() -> new MockPS(MockP::new, MEMBER_COUNT)).initBlocks()));
DAG dagNormal = new DAG().vertex(new Vertex("test",
new MockPS(MockP::new, MEMBER_COUNT)));
List<Future<Job>> submitFutures = new ArrayList<>();
// When
int numJobs = 16;
for (int i = 0; i < numJobs; i++) {
submitFutures.add(spawn(() -> newJob(dagBlocking)));
}
// Then
// JetService.getJobs()
if (useLightJob) {
assertEquals(0, instance().getJet().getJobs().size());
} else {
assertTrueEventually(() -> assertEquals(numJobs, instance().getJet().getJobs().size()), 5);
}
// newJob()
instance().getJet().newJob(dagNormal).join();
// newLightJob()
instance().getJet().newLightJob(dagNormal).join();
// generic API operation - generic API threads should not be starved
for (Object m : instance().getMap("m").values()) {
System.out.println(m);
}
for (int i = 0; i < submitFutures.size(); i++) {
MockPS.unblock();
}
for (Future<Job> f : submitFutures) {
f.get().join();
}
}
@Test
public void when_psInitBlocks_then_otherJobsNotBlocked() throws Exception {
// Given
DAG dagBlocking = new DAG().vertex(new Vertex("test",
new MockPS(MockP::new, MEMBER_COUNT).initBlocks()));
DAG dagNormal = new DAG().vertex(new Vertex("test",
new MockPS(MockP::new, MEMBER_COUNT)));
List<Future<Job>> submitFutures = new ArrayList<>();
// When
int numJobs = 16;
for (int i = 0; i < numJobs; i++) {
submitFutures.add(spawn(() -> newJob(dagBlocking)));
}
// Then
// JetService.getJobs()
if (useLightJob) {
assertEquals(0, instance().getJet().getJobs().size());
} else {
assertTrueEventually(() -> assertEquals(numJobs, instance().getJet().getJobs().size()), 5);
}
// newJob()
instance().getJet().newJob(dagNormal).join();
// newLightJob()
instance().getJet().newLightJob(dagNormal).join();
// generic API operation - generic API threads should not be starved
for (Object m : instance().getMap("m").values()) {
System.out.println(m);
}
for (int i = 0; i < submitFutures.size() * MEMBER_COUNT; i++) {
MockPS.unblock();
}
for (Future<Job> f : submitFutures) {
f.get().join();
}
}
@Test
public void when_processorInitBlocks_then_otherJobsNotBlocked() throws Exception {
// Given
DAG dagBlocking = new DAG().vertex(new Vertex("test",
() -> new MockP().initBlocks()));
DAG dagNormal = new DAG().vertex(new Vertex("test",
MockP::new));
List<Future<Job>> submitFutures = new ArrayList<>();
// When
int numJobs = 16;
for (int i = 0; i < numJobs; i++) {
submitFutures.add(spawn(() -> newJob(dagBlocking)));
}
// Then
// JetService.getJobs()
if (useLightJob) {
assertEquals(0, instance().getJet().getJobs().size());
} else {
assertTrueEventually(() -> assertEquals(numJobs, instance().getJet().getJobs().size()), 5);
}
// newJob()
instance().getJet().newJob(dagNormal).join();
// newLightJob()
instance().getJet().newLightJob(dagNormal).join();
// generic API operation - generic API threads should not be starved
for (Object m : instance().getMap("m").values()) {
System.out.println(m);
}
for (int i = 0; i < submitFutures.size() * MEMBER_COUNT * parallelism; i++) {
MockP.unblock();
}
for (Future<Job> f : submitFutures) {
f.get().join();
}
}
```
For these tests, the `TestProcessors` class also needs to be modified according to the [TestProcessors.txt](https://github.com/hazelcast/hazelcast/files/8800214/TestProcessors.txt) file. | 1.0 | Incorrect handling of blocking calls in ProcessorSupplier.init() [HZ-1204] - According to its contract, the `init()` and `close()` methods in `Processor`, `ProcessorSupplier` and `ProcessorMetaSupplier` are allowed to block, even for cooperative processors. It's explicitly written in `Processor` javadoc, but it's assumed in the docs, as the methods are used for resource initialization and cleanup. We also use it this way in our code, e.g. for opening the HZ client for remote map reader.
However, if the methods are blocking for a long time, they can easily starve thread pools:
- the coordinator thread pool, with four hard-coded threads, on master member. When all threads are blocked, most operations in the `Job` class are blocked too
- the generic operation thread pool, with configurable size. When all threads in this pool are blocked, it blocks many non-Jet operations, such as `IMap.values()`.
There are 2 possible solutions:
1. we will offload these methods to their own pool (shared with "cached" threads) so that they don't block anything else other than the initialization of other jobs. The handling of `InitExeutionOperation` will need to become async. This will create extra overhead for most processors that don't do any blocking calls in `init` or `close` and will slow down the job deploy+cleanup time. We can guard with methods akin to `Processor.closeIsCooperative()`, but that's an extra burden for the developer and it's easy to omit.
2. we can change the contract of these methods to specify that they can block only for a short time. If they need to block longer, they should move the initialization later and do it lazily in `process()`. This includes fixing all our processors to comply.
Below are tests that can be added to `ExecutionLifecycleTest`, should we implement the 1st approach. Today these tests fail.
```java
@Test
public void when_pmsInitBlocks_then_otherJobsNotBlocked() throws ExecutionException, InterruptedException {
// Given
DAG dagBlocking = new DAG().vertex(new Vertex("test",
new MockPMS(() -> new MockPS(MockP::new, MEMBER_COUNT)).initBlocks()));
DAG dagNormal = new DAG().vertex(new Vertex("test",
new MockPS(MockP::new, MEMBER_COUNT)));
List<Future<Job>> submitFutures = new ArrayList<>();
// When
int numJobs = 16;
for (int i = 0; i < numJobs; i++) {
submitFutures.add(spawn(() -> newJob(dagBlocking)));
}
// Then
// JetService.getJobs()
if (useLightJob) {
assertEquals(0, instance().getJet().getJobs().size());
} else {
assertTrueEventually(() -> assertEquals(numJobs, instance().getJet().getJobs().size()), 5);
}
// newJob()
instance().getJet().newJob(dagNormal).join();
// newLightJob()
instance().getJet().newLightJob(dagNormal).join();
// generic API operation - generic API threads should not be starved
for (Object m : instance().getMap("m").values()) {
System.out.println(m);
}
for (int i = 0; i < submitFutures.size(); i++) {
MockPS.unblock();
}
for (Future<Job> f : submitFutures) {
f.get().join();
}
}
@Test
public void when_psInitBlocks_then_otherJobsNotBlocked() throws Exception {
// Given
DAG dagBlocking = new DAG().vertex(new Vertex("test",
new MockPS(MockP::new, MEMBER_COUNT).initBlocks()));
DAG dagNormal = new DAG().vertex(new Vertex("test",
new MockPS(MockP::new, MEMBER_COUNT)));
List<Future<Job>> submitFutures = new ArrayList<>();
// When
int numJobs = 16;
for (int i = 0; i < numJobs; i++) {
submitFutures.add(spawn(() -> newJob(dagBlocking)));
}
// Then
// JetService.getJobs()
if (useLightJob) {
assertEquals(0, instance().getJet().getJobs().size());
} else {
assertTrueEventually(() -> assertEquals(numJobs, instance().getJet().getJobs().size()), 5);
}
// newJob()
instance().getJet().newJob(dagNormal).join();
// newLightJob()
instance().getJet().newLightJob(dagNormal).join();
// generic API operation - generic API threads should not be starved
for (Object m : instance().getMap("m").values()) {
System.out.println(m);
}
for (int i = 0; i < submitFutures.size() * MEMBER_COUNT; i++) {
MockPS.unblock();
}
for (Future<Job> f : submitFutures) {
f.get().join();
}
}
@Test
public void when_processorInitBlocks_then_otherJobsNotBlocked() throws Exception {
// Given
DAG dagBlocking = new DAG().vertex(new Vertex("test",
() -> new MockP().initBlocks()));
DAG dagNormal = new DAG().vertex(new Vertex("test",
MockP::new));
List<Future<Job>> submitFutures = new ArrayList<>();
// When
int numJobs = 16;
for (int i = 0; i < numJobs; i++) {
submitFutures.add(spawn(() -> newJob(dagBlocking)));
}
// Then
// JetService.getJobs()
if (useLightJob) {
assertEquals(0, instance().getJet().getJobs().size());
} else {
assertTrueEventually(() -> assertEquals(numJobs, instance().getJet().getJobs().size()), 5);
}
// newJob()
instance().getJet().newJob(dagNormal).join();
// newLightJob()
instance().getJet().newLightJob(dagNormal).join();
// generic API operation - generic API threads should not be starved
for (Object m : instance().getMap("m").values()) {
System.out.println(m);
}
for (int i = 0; i < submitFutures.size() * MEMBER_COUNT * parallelism; i++) {
MockP.unblock();
}
for (Future<Job> f : submitFutures) {
f.get().join();
}
}
```
For these tests, the `TestProcessors` class also needs to be modified according to the [TestProcessors.txt](https://github.com/hazelcast/hazelcast/files/8800214/TestProcessors.txt) file. | defect | incorrect handling of blocking calls in processorsupplier init according to its contract the init and close methods in processor processorsupplier and processormetasupplier are allowed to block even for cooperative processors it s explicitly written in processor javadoc but it s assumed in the docs as the methods are used for resource initialization and cleanup we also use it this way in our code e g for opening the hz client for remote map reader however if the methods are blocking for a long time they can easily starve thread pools the coordinator thread pool with four hard coded threads on master member when all threads are blocked most operations in the job class are blocked too the generic operation thread pool with configurable size when all threads in this pool are blocked it blocks many non jet operations such as imap values there are possible solutions we will offload these methods to their own pool shared with cached threads so that they don t block anything else other than the initialization of other jobs the handling of initexeutionoperation will need to become async this will create extra overhead for most processors that don t do any blocking calls in init or close and will slow down the job deploy cleanup time we can guard with methods akin to processor closeiscooperative but that s an extra burden for the developer and it s easy to omit we can change the contract of these methods to specify that they can block only for a short time if they need to block longer they should move the initialization later and do it lazily in process this includes fixing all our processors to comply below are tests that can be added to executionlifecycletest should we implement the approach today these tests fail java test public void when pmsinitblocks then otherjobsnotblocked throws executionexception interruptedexception given dag dagblocking new dag vertex new vertex test new mockpms new mockps mockp new member count initblocks dag dagnormal new dag vertex new vertex test new mockps mockp new member count list submitfutures new arraylist when int numjobs for int i i numjobs i submitfutures add spawn newjob dagblocking then jetservice getjobs if uselightjob assertequals instance getjet getjobs size else asserttrueeventually assertequals numjobs instance getjet getjobs size newjob instance getjet newjob dagnormal join newlightjob instance getjet newlightjob dagnormal join generic api operation generic api threads should not be starved for object m instance getmap m values system out println m for int i i submitfutures size i mockps unblock for future f submitfutures f get join test public void when psinitblocks then otherjobsnotblocked throws exception given dag dagblocking new dag vertex new vertex test new mockps mockp new member count initblocks dag dagnormal new dag vertex new vertex test new mockps mockp new member count list submitfutures new arraylist when int numjobs for int i i numjobs i submitfutures add spawn newjob dagblocking then jetservice getjobs if uselightjob assertequals instance getjet getjobs size else asserttrueeventually assertequals numjobs instance getjet getjobs size newjob instance getjet newjob dagnormal join newlightjob instance getjet newlightjob dagnormal join generic api operation generic api threads should not be starved for object m instance getmap m values system out println m for int i i submitfutures size member count i mockps unblock for future f submitfutures f get join test public void when processorinitblocks then otherjobsnotblocked throws exception given dag dagblocking new dag vertex new vertex test new mockp initblocks dag dagnormal new dag vertex new vertex test mockp new list submitfutures new arraylist when int numjobs for int i i numjobs i submitfutures add spawn newjob dagblocking then jetservice getjobs if uselightjob assertequals instance getjet getjobs size else asserttrueeventually assertequals numjobs instance getjet getjobs size newjob instance getjet newjob dagnormal join newlightjob instance getjet newlightjob dagnormal join generic api operation generic api threads should not be starved for object m instance getmap m values system out println m for int i i submitfutures size member count parallelism i mockp unblock for future f submitfutures f get join for these tests the testprocessors class also needs to be modified according to the file | 1 |
27,272 | 5,328,274,949 | IssuesEvent | 2017-02-15 11:33:33 | coala/coala | https://api.github.com/repos/coala/coala | closed | Developers/Writing_Good_Commits.rst: It should describe only one change and not more. (Reword) | area/documentation difficulty/newcomer | _From @Udayan12167 on October 6, 2016 13:49_
Drop Only.
_Copied from original issue: coala/documentation#150_ | 1.0 | Developers/Writing_Good_Commits.rst: It should describe only one change and not more. (Reword) - _From @Udayan12167 on October 6, 2016 13:49_
Drop Only.
_Copied from original issue: coala/documentation#150_ | non_defect | developers writing good commits rst it should describe only one change and not more reword from on october drop only copied from original issue coala documentation | 0 |
75,156 | 25,562,379,069 | IssuesEvent | 2022-11-30 11:48:11 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | Informix DDL statements generate invalid SQL for timestamp columns | T: Defect C: Functionality C: DB: Informix P: High R: Fixed E: Enterprise Edition | A new integration test query fails on Informix:
```sql
CREATE TABLE T_COMP_CLIENT_SIDE_DEP (
ID INTEGER NOT NULL,
-- ...
X_MODIFIED_AT DATETIME,
PRIMARY KEY (ID) CONSTRAINT PK_T_COMP_CLIENT_SIDE_DEP
);
```
We should use the cast type for `DATETIME`, e.g. `DATETIME YEAR TO FRACTION (5)` | 1.0 | Informix DDL statements generate invalid SQL for timestamp columns - A new integration test query fails on Informix:
```sql
CREATE TABLE T_COMP_CLIENT_SIDE_DEP (
ID INTEGER NOT NULL,
-- ...
X_MODIFIED_AT DATETIME,
PRIMARY KEY (ID) CONSTRAINT PK_T_COMP_CLIENT_SIDE_DEP
);
```
We should use the cast type for `DATETIME`, e.g. `DATETIME YEAR TO FRACTION (5)` | defect | informix ddl statements generate invalid sql for timestamp columns a new integration test query fails on informix sql create table t comp client side dep id integer not null x modified at datetime primary key id constraint pk t comp client side dep we should use the cast type for datetime e g datetime year to fraction | 1 |
281,913 | 21,315,448,040 | IssuesEvent | 2022-04-16 07:29:49 | sharpstorm/pe | https://api.github.com/repos/sharpstorm/pe | opened | Inconsistent Parser Diagrams | severity.Low type.DocumentationBug | The Parser class diagrams on page and page 8 do not match, in particular the arrow from `CliSyntax` to `Prefix`. One of it must be correct, and the other wrong. Since this difference changes the meaning of the diagram, this cannot be considered a cosmetic bug.
## Screenshots of diagrams
**Page 7:**

**Page 8:**

<!--session: 1650089504746-5589ba0c-52cc-49e8-b8cd-9e6904a786dd-->
<!--Version: Web v3.4.2--> | 1.0 | Inconsistent Parser Diagrams - The Parser class diagrams on page and page 8 do not match, in particular the arrow from `CliSyntax` to `Prefix`. One of it must be correct, and the other wrong. Since this difference changes the meaning of the diagram, this cannot be considered a cosmetic bug.
## Screenshots of diagrams
**Page 7:**

**Page 8:**

<!--session: 1650089504746-5589ba0c-52cc-49e8-b8cd-9e6904a786dd-->
<!--Version: Web v3.4.2--> | non_defect | inconsistent parser diagrams the parser class diagrams on page and page do not match in particular the arrow from clisyntax to prefix one of it must be correct and the other wrong since this difference changes the meaning of the diagram this cannot be considered a cosmetic bug screenshots of diagrams page page | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.