Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
276,214
| 20,974,221,263
|
IssuesEvent
|
2022-03-28 13:59:53
|
DavidT3/XGA
|
https://api.github.com/repos/DavidT3/XGA
|
opened
|
Should make it clear that coordinates are currently expected to be J2000
|
documentation enhancement
|
Or swap out the coordinate quantities to be actual astropy coordinate objects as they should have been from the beginning
|
1.0
|
Should make it clear that coordinates are currently expected to be J2000 - Or swap out the coordinate quantities to be actual astropy coordinate objects as they should have been from the beginning
|
non_defect
|
should make it clear that coordinates are currently expected to be or swap out the coordinate quantities to be actual astropy coordinate objects as they should have been from the beginning
| 0
|
3,805
| 2,610,069,349
|
IssuesEvent
|
2015-02-26 18:20:20
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
台州割包茎哪个医院最好
|
auto-migrated Priority-Medium Type-Defect
|
```
台州割包茎哪个医院最好【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 11:43
|
1.0
|
台州割包茎哪个医院最好 - ```
台州割包茎哪个医院最好【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 11:43
|
defect
|
台州割包茎哪个医院最好 台州割包茎哪个医院最好【台州五洲生殖医院】 询热线 微信号tzwzszyy 医院地址 台州 (枫南大转盘旁)乘车线路 、 � �� 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 1
|
195,890
| 14,787,311,142
|
IssuesEvent
|
2021-01-12 07:22:58
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: backupTPCC failed
|
C-test-failure O-roachtest O-robot branch-release-20.1 release-blocker
|
[(roachtest).backupTPCC failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2574842&tab=buildLog) on [release-20.1@e395c0c7c48a279334f0e94dfb7030a3eafa093f](https://github.com/cockroachdb/cockroach/commits/e395c0c7c48a279334f0e94dfb7030a3eafa093f):
```
The test failed on branch=release-20.1, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/backupTPCC/run_1
cluster.go:2198,backup.go:108,test_runner.go:749: output in run_072240.275_n1_workload_init_tpcc: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2574842-1610435721-21-n3cpu4:1 -- ./workload init tpcc --warehouses=10 {pgurl:1-3} --deprecated-fk-indexes returned: exit status 20
(1) attached stack trace
| main.(*cluster).RunE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2276
| main.(*cluster).Run
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2196
| main.registerBackup.func3
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/backup.go:108
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:749
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (2) 2 safe details enclosed
Wraps: (3) output in run_072240.275_n1_workload_init_tpcc
Wraps: (4) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2574842-1610435721-21-n3cpu4:1 -- ./workload init tpcc --warehouses=10 {pgurl:1-3} --deprecated-fk-indexes returned
| stderr:
| bash: line 1: 4884 Illegal instruction (core dumped) bash -c "./workload init tpcc --warehouses=10 'postgres://root@10.128.0.9:26257?sslmode=disable' 'postgres://root@10.128.0.130:26257?sslmode=disable' 'postgres://root@10.128.0.129:26257?sslmode=disable' --deprecated-fk-indexes"
| Error: COMMAND_PROBLEM: exit status 132
| (1) COMMAND_PROBLEM
| Wraps: (2) Node 1. Command with error:
| | ```
| | ./workload init tpcc --warehouses=10 {pgurl:1-3} --deprecated-fk-indexes
| | ```
| Wraps: (3) exit status 132
| Error types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError
|
| stdout:
Wraps: (5) exit status 20
Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError
```
<details><summary>More</summary><p>
Artifacts: [/backupTPCC](https://teamcity.cockroachdb.com/viewLog.html?buildId=2574842&tab=artifacts#/backupTPCC)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2AbackupTPCC.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
2.0
|
roachtest: backupTPCC failed - [(roachtest).backupTPCC failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2574842&tab=buildLog) on [release-20.1@e395c0c7c48a279334f0e94dfb7030a3eafa093f](https://github.com/cockroachdb/cockroach/commits/e395c0c7c48a279334f0e94dfb7030a3eafa093f):
```
The test failed on branch=release-20.1, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/backupTPCC/run_1
cluster.go:2198,backup.go:108,test_runner.go:749: output in run_072240.275_n1_workload_init_tpcc: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2574842-1610435721-21-n3cpu4:1 -- ./workload init tpcc --warehouses=10 {pgurl:1-3} --deprecated-fk-indexes returned: exit status 20
(1) attached stack trace
| main.(*cluster).RunE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2276
| main.(*cluster).Run
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2196
| main.registerBackup.func3
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/backup.go:108
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:749
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (2) 2 safe details enclosed
Wraps: (3) output in run_072240.275_n1_workload_init_tpcc
Wraps: (4) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2574842-1610435721-21-n3cpu4:1 -- ./workload init tpcc --warehouses=10 {pgurl:1-3} --deprecated-fk-indexes returned
| stderr:
| bash: line 1: 4884 Illegal instruction (core dumped) bash -c "./workload init tpcc --warehouses=10 'postgres://root@10.128.0.9:26257?sslmode=disable' 'postgres://root@10.128.0.130:26257?sslmode=disable' 'postgres://root@10.128.0.129:26257?sslmode=disable' --deprecated-fk-indexes"
| Error: COMMAND_PROBLEM: exit status 132
| (1) COMMAND_PROBLEM
| Wraps: (2) Node 1. Command with error:
| | ```
| | ./workload init tpcc --warehouses=10 {pgurl:1-3} --deprecated-fk-indexes
| | ```
| Wraps: (3) exit status 132
| Error types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError
|
| stdout:
Wraps: (5) exit status 20
Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError
```
<details><summary>More</summary><p>
Artifacts: [/backupTPCC](https://teamcity.cockroachdb.com/viewLog.html?buildId=2574842&tab=artifacts#/backupTPCC)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2AbackupTPCC.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
non_defect
|
roachtest backuptpcc failed on the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts backuptpcc run cluster go backup go test runner go output in run workload init tpcc home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload init tpcc warehouses pgurl deprecated fk indexes returned exit status attached stack trace main cluster rune home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main cluster run home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main registerbackup home agent work go src github com cockroachdb cockroach pkg cmd roachtest backup go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go runtime goexit usr local go src runtime asm s wraps safe details enclosed wraps output in run workload init tpcc wraps home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload init tpcc warehouses pgurl deprecated fk indexes returned stderr bash line illegal instruction core dumped bash c workload init tpcc warehouses postgres root sslmode disable postgres root sslmode disable postgres root sslmode disable deprecated fk indexes error command problem exit status command problem wraps node command with error workload init tpcc warehouses pgurl deprecated fk indexes wraps exit status error types errors cmd hintdetail withdetail exec exiterror stdout wraps exit status error types withstack withstack safedetails withsafedetails errutil withmessage main withcommanddetails exec exiterror more artifacts powered by
| 0
|
313,817
| 26,956,656,702
|
IssuesEvent
|
2023-02-08 15:20:14
|
vegaprotocol/frontend-monorepo
|
https://api.github.com/repos/vegaprotocol/frontend-monorepo
|
closed
|
Trading test refactor
|
Trading Testing 🧪
|
## The Chore
We have highly increased Trading app test coverage, and that will continue in the nearest future. There are also new capsule tests for trading. We need to clean up a little for better performance and to ease possible parallelisation.
## Tasks
- [x] Analyse existing test scenarios and improve their structure. There are tests, that execute multiple validations in single `it` block, which 1. decrease visibility, 2. prevent validations to be done if one of them is failed. We need to rework it into `describe`s with many `it`s, where possible.
- [x] Minimise situations where app is refreshed, as it takes much time to reload the app. If it's possible to perform many tests without refreshing, let's do that instead.
- [x] We have bigger and bigget test files with many contexts and describes. Break it down to multiple spec files, where possible, which increase readability and allow such tests to be run in parallel in the future.
|
1.0
|
Trading test refactor - ## The Chore
We have highly increased Trading app test coverage, and that will continue in the nearest future. There are also new capsule tests for trading. We need to clean up a little for better performance and to ease possible parallelisation.
## Tasks
- [x] Analyse existing test scenarios and improve their structure. There are tests, that execute multiple validations in single `it` block, which 1. decrease visibility, 2. prevent validations to be done if one of them is failed. We need to rework it into `describe`s with many `it`s, where possible.
- [x] Minimise situations where app is refreshed, as it takes much time to reload the app. If it's possible to perform many tests without refreshing, let's do that instead.
- [x] We have bigger and bigget test files with many contexts and describes. Break it down to multiple spec files, where possible, which increase readability and allow such tests to be run in parallel in the future.
|
non_defect
|
trading test refactor the chore we have highly increased trading app test coverage and that will continue in the nearest future there are also new capsule tests for trading we need to clean up a little for better performance and to ease possible parallelisation tasks analyse existing test scenarios and improve their structure there are tests that execute multiple validations in single it block which decrease visibility prevent validations to be done if one of them is failed we need to rework it into describe s with many it s where possible minimise situations where app is refreshed as it takes much time to reload the app if it s possible to perform many tests without refreshing let s do that instead we have bigger and bigget test files with many contexts and describes break it down to multiple spec files where possible which increase readability and allow such tests to be run in parallel in the future
| 0
|
43,079
| 11,462,021,369
|
IssuesEvent
|
2020-02-07 13:18:35
|
libarchive/libarchive
|
https://api.github.com/repos/libarchive/libarchive
|
closed
|
Hang observed with xz/7z
|
OpSys-All Priority-Medium Type-Defect
|
Original [issue 315](https://code.google.com/p/libarchive/issues/detail?id=315) created by Google Code user `Ztatik.Light` on 2013-04-23T08:45:33.000Z:
```
Concerning libarchive v3.1.2... This may or may not be known already, but more testing is required for the xz and 7zip formats because with small files, they work... But with larger files that contain several files, it either doesn't work or hangs.
What I'm testing with is about 100 files that total about 25 MB compressed as either .tar.xz or .7z
```
|
1.0
|
Hang observed with xz/7z - Original [issue 315](https://code.google.com/p/libarchive/issues/detail?id=315) created by Google Code user `Ztatik.Light` on 2013-04-23T08:45:33.000Z:
```
Concerning libarchive v3.1.2... This may or may not be known already, but more testing is required for the xz and 7zip formats because with small files, they work... But with larger files that contain several files, it either doesn't work or hangs.
What I'm testing with is about 100 files that total about 25 MB compressed as either .tar.xz or .7z
```
|
defect
|
hang observed with xz original created by google code user ztatik light on concerning libarchive this may or may not be known already but more testing is required for the xz and formats because with small files they work but with larger files that contain several files it either doesn t work or hangs what i m testing with is about files that total about mb compressed as either tar xz or
| 1
|
67,205
| 20,961,583,609
|
IssuesEvent
|
2022-03-27 21:43:54
|
abedmaatalla/foursquared
|
https://api.github.com/repos/abedmaatalla/foursquared
|
closed
|
Location not updated when checking in
|
Priority-Medium Type-Defect auto-migrated
|
```
I would expect a location based app to recalculate a user's location when they
(attempt to) check in - this means that they will never get a "We don't think
you are actually at that location" error, which they are, and GPS will confirm.
This error has occurred for me several times recently - mainly when I use
high-speed transport to get somewhere (ie - train) - if I don't wait long
enough before checking in, I am given this error, and score no points for the
location even after the position is refreshed and I retry checking in.
```
Original issue reported on code.google.com by `nat.abbo...@wavewatchers.org` on 6 Jan 2011 at 7:06
|
1.0
|
Location not updated when checking in - ```
I would expect a location based app to recalculate a user's location when they
(attempt to) check in - this means that they will never get a "We don't think
you are actually at that location" error, which they are, and GPS will confirm.
This error has occurred for me several times recently - mainly when I use
high-speed transport to get somewhere (ie - train) - if I don't wait long
enough before checking in, I am given this error, and score no points for the
location even after the position is refreshed and I retry checking in.
```
Original issue reported on code.google.com by `nat.abbo...@wavewatchers.org` on 6 Jan 2011 at 7:06
|
defect
|
location not updated when checking in i would expect a location based app to recalculate a user s location when they attempt to check in this means that they will never get a we don t think you are actually at that location error which they are and gps will confirm this error has occurred for me several times recently mainly when i use high speed transport to get somewhere ie train if i don t wait long enough before checking in i am given this error and score no points for the location even after the position is refreshed and i retry checking in original issue reported on code google com by nat abbo wavewatchers org on jan at
| 1
|
38,661
| 8,951,779,276
|
IssuesEvent
|
2019-01-25 14:53:26
|
svigerske/ipopt-donotuse
|
https://api.github.com/repos/svigerske/ipopt-donotuse
|
closed
|
trunk version fails on infeasible problem with all variables fixed
|
Ipopt defect
|
Issue created by migration from Trac.
Original creator: @svigerske
Original creation time: 2008-03-13 18:44:04
Assignee: ipopt-team
Version: 3.3
Hi,
if I run Ipopt on a problem where all variables are fixed and the equations are infeasible within the bounds, then Ipopt runs into a seg. fault at IpIpoptApplication.cpp:658, p2ip_data does not get initialized completely and then p2ip_cq->curr_f() tries to read the primal values of an IteratesVector that has not become allocated.
Would be nice if this could be fixed for trunk in a way similar to as it is working for stable already. There IpTNLPAdapter.cpp:586 does not throw LOCALLY_INFEASIBLE, but NO_FREE_VARIABLES_AND_INFEASIBLE.
My workaround is [here](https://projects.coin-or.org/GAMSlinks/browser/trunk/ipopt.patch), but it is not really nice.
Best,
Stefan
|
1.0
|
trunk version fails on infeasible problem with all variables fixed - Issue created by migration from Trac.
Original creator: @svigerske
Original creation time: 2008-03-13 18:44:04
Assignee: ipopt-team
Version: 3.3
Hi,
if I run Ipopt on a problem where all variables are fixed and the equations are infeasible within the bounds, then Ipopt runs into a seg. fault at IpIpoptApplication.cpp:658, p2ip_data does not get initialized completely and then p2ip_cq->curr_f() tries to read the primal values of an IteratesVector that has not become allocated.
Would be nice if this could be fixed for trunk in a way similar to as it is working for stable already. There IpTNLPAdapter.cpp:586 does not throw LOCALLY_INFEASIBLE, but NO_FREE_VARIABLES_AND_INFEASIBLE.
My workaround is [here](https://projects.coin-or.org/GAMSlinks/browser/trunk/ipopt.patch), but it is not really nice.
Best,
Stefan
|
defect
|
trunk version fails on infeasible problem with all variables fixed issue created by migration from trac original creator svigerske original creation time assignee ipopt team version hi if i run ipopt on a problem where all variables are fixed and the equations are infeasible within the bounds then ipopt runs into a seg fault at ipipoptapplication cpp data does not get initialized completely and then cq curr f tries to read the primal values of an iteratesvector that has not become allocated would be nice if this could be fixed for trunk in a way similar to as it is working for stable already there iptnlpadapter cpp does not throw locally infeasible but no free variables and infeasible my workaround is but it is not really nice best stefan
| 1
|
59,784
| 17,023,244,558
|
IssuesEvent
|
2021-07-03 01:01:53
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
potlatch created doubled-up way
|
Component: potlatch (flash editor) Priority: major Resolution: worksforme Type: defect
|
**[Submitted to the original trac issue database at 8.09pm, Sunday, 4th May 2008]**
Not sure exactly how this happened, but check out revision 2008-05-04T20:38:04+01:00 of way 15242140 at http://api.openstreetmap.org/api/0.5/way/15242140/history .
I was refining an existing way (call it A) in potlatch by adding a few nodes; I also merged it with another way (B) to the south. It started out with 29 nodes, the fixed (current) version has 56 nodes, the one I mention above has 82 nodes. In this broken version, the way starts with B, then the original A forwards, then the refined A backwards. (Or perhaps the other way round.)
I believe that A was already doubled up before I merged it with B, so the merging has nothing to do with it, but I'm not sure. Also, at some point, I was able to select the unmodified A as a different way by clicking the unrefined segments...
|
1.0
|
potlatch created doubled-up way - **[Submitted to the original trac issue database at 8.09pm, Sunday, 4th May 2008]**
Not sure exactly how this happened, but check out revision 2008-05-04T20:38:04+01:00 of way 15242140 at http://api.openstreetmap.org/api/0.5/way/15242140/history .
I was refining an existing way (call it A) in potlatch by adding a few nodes; I also merged it with another way (B) to the south. It started out with 29 nodes, the fixed (current) version has 56 nodes, the one I mention above has 82 nodes. In this broken version, the way starts with B, then the original A forwards, then the refined A backwards. (Or perhaps the other way round.)
I believe that A was already doubled up before I merged it with B, so the merging has nothing to do with it, but I'm not sure. Also, at some point, I was able to select the unmodified A as a different way by clicking the unrefined segments...
|
defect
|
potlatch created doubled up way not sure exactly how this happened but check out revision of way at i was refining an existing way call it a in potlatch by adding a few nodes i also merged it with another way b to the south it started out with nodes the fixed current version has nodes the one i mention above has nodes in this broken version the way starts with b then the original a forwards then the refined a backwards or perhaps the other way round i believe that a was already doubled up before i merged it with b so the merging has nothing to do with it but i m not sure also at some point i was able to select the unmodified a as a different way by clicking the unrefined segments
| 1
|
2,980
| 2,607,968,339
|
IssuesEvent
|
2015-02-26 00:43:23
|
chrsmithdemos/leveldb
|
https://api.github.com/repos/chrsmithdemos/leveldb
|
opened
|
oen advice about snapshot
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1.all code adn information showed in the additional information
2.
3.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
leveldb 1.9 fedora17
Please provide any additional information below.
code begin=================================================
const leveldb::Snapshot* snap_shot = db->GetSnapshot();
leveldb::ReadOptions read_options;
// change "key0"'s value from "value0" to "new"
db->Put(leveldb::WriteOptions(),"key0","new");
string val;
db->Get(read_options,"key0",&val);
cout<<"no snapshot value: "<<val<<endl;//output " new" .
read_options.snapshot = snap_shot;
db->Get(read_options,"key0",&val);
cout<<"snapshot value: "<<val<<endl;//output "value0" .
// release snapshot
db->ReleaseSnapshot(read_options.snapshot);
db->Get(read_options,"key0",&val);
cout<<"after release snapshot: "<<val<<endl;
//output "value0" !!! It's wrong
code end===================================================
Advice:
If i release the snapshot ,i should get the newest infomation.
But only i add the "read_options.snapshot = NULL; " after
db->ReleaseSnapshot(...)manually,can i get the right result---output "new".
I deep into the code in db_imp.cc and know it will judge
"options.snapshot!=NULL" when call NewIterator(...) and Get(...) .
SO I think set the read_options.snapshot=NULL in DB::ReleaseSnapshot() is a
good idea. Or you should modify the document clearly and add
"read_options.snapshot = NULL;" to the example in the Snapshots part of
http://leveldb.googlecode.com/svn/trunk/doc/index.html
```
-----
Original issue reported on code.google.com by `wangteng...@gmail.com` on 15 May 2013 at 11:37
|
1.0
|
oen advice about snapshot - ```
What steps will reproduce the problem?
1.all code adn information showed in the additional information
2.
3.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
leveldb 1.9 fedora17
Please provide any additional information below.
code begin=================================================
const leveldb::Snapshot* snap_shot = db->GetSnapshot();
leveldb::ReadOptions read_options;
// change "key0"'s value from "value0" to "new"
db->Put(leveldb::WriteOptions(),"key0","new");
string val;
db->Get(read_options,"key0",&val);
cout<<"no snapshot value: "<<val<<endl;//output " new" .
read_options.snapshot = snap_shot;
db->Get(read_options,"key0",&val);
cout<<"snapshot value: "<<val<<endl;//output "value0" .
// release snapshot
db->ReleaseSnapshot(read_options.snapshot);
db->Get(read_options,"key0",&val);
cout<<"after release snapshot: "<<val<<endl;
//output "value0" !!! It's wrong
code end===================================================
Advice:
If i release the snapshot ,i should get the newest infomation.
But only i add the "read_options.snapshot = NULL; " after
db->ReleaseSnapshot(...)manually,can i get the right result---output "new".
I deep into the code in db_imp.cc and know it will judge
"options.snapshot!=NULL" when call NewIterator(...) and Get(...) .
SO I think set the read_options.snapshot=NULL in DB::ReleaseSnapshot() is a
good idea. Or you should modify the document clearly and add
"read_options.snapshot = NULL;" to the example in the Snapshots part of
http://leveldb.googlecode.com/svn/trunk/doc/index.html
```
-----
Original issue reported on code.google.com by `wangteng...@gmail.com` on 15 May 2013 at 11:37
|
defect
|
oen advice about snapshot what steps will reproduce the problem all code adn information showed in the additional information what is the expected output what do you see instead what version of the product are you using on what operating system leveldb please provide any additional information below code begin const leveldb snapshot snap shot db getsnapshot leveldb readoptions read options change s value from to new db put leveldb writeoptions new string val db get read options val cout no snapshot value val endl output new read options snapshot snap shot db get read options val cout snapshot value val endl output release snapshot db releasesnapshot read options snapshot db get read options val cout after release snapshot val endl output it s wrong code end advice if i release the snapshot i should get the newest infomation but only i add the read options snapshot null after db releasesnapshot manually can i get the right result output new i deep into the code in db imp cc and know it will judge options snapshot null when call newiterator and get so i think set the read options snapshot null in db releasesnapshot is a good idea or you should modify the document clearly and add read options snapshot null to the example in the snapshots part of original issue reported on code google com by wangteng gmail com on may at
| 1
|
16,978
| 4,110,800,211
|
IssuesEvent
|
2016-06-07 01:28:49
|
brave/browser-laptop
|
https://api.github.com/repos/brave/browser-laptop
|
opened
|
General Settings: "Browser data import" setting
|
documentation settings
|
Strings:
Item label:
Browser data import
Button string:
"Import now..."
Functionality:
Choosing "Import now..." opens the Import Browser Data dialog which floats above the General Settings panel. (This dialog is modal to all other functionality within the browser settings panels.)
The pull-down menu list is populated by polling the OS for installed browsers that have at least one data type that can be imported. Each browser offers a different set of importable data types. The available toggle switches change dynamically based on the selected browser, becoming disabled to indicate lack of availability.
Safari: Preferences, Bookmarks
Chrome: Cookies, Browsing History, Bookmarks
Firefox: Browsing History, Bookmarks, Saved Passwords, Autofill form data
Superset: Browsing History, Bookmarks, Cookies, Preferences, Saved Passwords, Autofill form data
Spec image: (issue item shown in blue)

Spec image with import dialog open:

|
1.0
|
General Settings: "Browser data import" setting - Strings:
Item label:
Browser data import
Button string:
"Import now..."
Functionality:
Choosing "Import now..." opens the Import Browser Data dialog which floats above the General Settings panel. (This dialog is modal to all other functionality within the browser settings panels.)
The pull-down menu list is populated by polling the OS for installed browsers that have at least one data type that can be imported. Each browser offers a different set of importable data types. The available toggle switches change dynamically based on the selected browser, becoming disabled to indicate lack of availability.
Safari: Preferences, Bookmarks
Chrome: Cookies, Browsing History, Bookmarks
Firefox: Browsing History, Bookmarks, Saved Passwords, Autofill form data
Superset: Browsing History, Bookmarks, Cookies, Preferences, Saved Passwords, Autofill form data
Spec image: (issue item shown in blue)

Spec image with import dialog open:

|
non_defect
|
general settings browser data import setting strings item label browser data import button string import now functionality choosing import now opens the import browser data dialog which floats above the general settings panel this dialog is modal to all other functionality within the browser settings panels the pull down menu list is populated by polling the os for installed browsers that have at least one data type that can be imported each browser offers a different set of importable data types the available toggle switches change dynamically based on the selected browser becoming disabled to indicate lack of availability safari preferences bookmarks chrome cookies browsing history bookmarks firefox browsing history bookmarks saved passwords autofill form data superset browsing history bookmarks cookies preferences saved passwords autofill form data spec image issue item shown in blue spec image with import dialog open
| 0
|
462,026
| 13,239,873,648
|
IssuesEvent
|
2020-08-19 04:50:00
|
larshp/abapGit
|
https://api.github.com/repos/larshp/abapGit
|
closed
|
Error when using "Advanced > Force Stage"
|
bug low priority
|
If there are no diffs, then using "Advanced > Force Stage" leads to the following error:

|
1.0
|
Error when using "Advanced > Force Stage" - If there are no diffs, then using "Advanced > Force Stage" leads to the following error:

|
non_defect
|
error when using advanced force stage if there are no diffs then using advanced force stage leads to the following error
| 0
|
50,784
| 13,187,742,662
|
IssuesEvent
|
2020-08-13 04:25:47
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
Missing Factor 2 in python implementation of paraboloid (Trac #1364)
|
Migrated from Trac combo reconstruction defect
|
In paraboloid/releases/V01-07-00/python/fit_paraboloid.py
Nice docu in code. I followed the calculation of the formulars.
There is an coefficient comparison between L. 30 and L. 38 which results in
relation between coefficients in L. 41-43.
Here line 41 & 43 are missing a factor of 2.
L. 41 should be
a = 2(A cosfi**2 + B sinfi**2)
and L 43 should be
c = 2(A sinfi**2 + B cosfi**2)
these factors are missing in all following calculations.
I have done the exercise. The result should be:
case b=0 and a=c:
A=B=a/2=c/2
case b=0 and a!=c:
a=2A
c=2B
case b!=0:
tanfi_1/2 = 1/(2b) * [c-a +- sqrt( (a-c)**2 +4b**2)]
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1364">https://code.icecube.wisc.edu/ticket/1364</a>, reported by reimann and owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "In paraboloid/releases/V01-07-00/python/fit_paraboloid.py\n\nNice docu in code. I followed the calculation of the formulars. \nThere is an coefficient comparison between L. 30 and L. 38 which results in \nrelation between coefficients in L. 41-43.\nHere line 41 & 43 are missing a factor of 2.\nL. 41 should be \na = 2(A cosfi**2 + B sinfi**2)\nand L 43 should be\nc = 2(A sinfi**2 + B cosfi**2)\n\nthese factors are missing in all following calculations. \n\nI have done the exercise. The result should be:\ncase b=0 and a=c:\n A=B=a/2=c/2\ncase b=0 and a!=c:\n a=2A\n c=2B\ncase b!=0:\n tanfi_1/2 = 1/(2b) * [c-a +- sqrt( (a-c)**2 +4b**2)] ",
"reporter": "reimann",
"cc": "kmeagher@ulb.ac.be",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "Missing Factor 2 in python implementation of paraboloid",
"priority": "normal",
"keywords": "paraboloid",
"time": "2015-09-22T15:19:30",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Missing Factor 2 in python implementation of paraboloid (Trac #1364) - In paraboloid/releases/V01-07-00/python/fit_paraboloid.py
Nice docu in code. I followed the calculation of the formulars.
There is an coefficient comparison between L. 30 and L. 38 which results in
relation between coefficients in L. 41-43.
Here line 41 & 43 are missing a factor of 2.
L. 41 should be
a = 2(A cosfi**2 + B sinfi**2)
and L 43 should be
c = 2(A sinfi**2 + B cosfi**2)
these factors are missing in all following calculations.
I have done the exercise. The result should be:
case b=0 and a=c:
A=B=a/2=c/2
case b=0 and a!=c:
a=2A
c=2B
case b!=0:
tanfi_1/2 = 1/(2b) * [c-a +- sqrt( (a-c)**2 +4b**2)]
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1364">https://code.icecube.wisc.edu/ticket/1364</a>, reported by reimann and owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "In paraboloid/releases/V01-07-00/python/fit_paraboloid.py\n\nNice docu in code. I followed the calculation of the formulars. \nThere is an coefficient comparison between L. 30 and L. 38 which results in \nrelation between coefficients in L. 41-43.\nHere line 41 & 43 are missing a factor of 2.\nL. 41 should be \na = 2(A cosfi**2 + B sinfi**2)\nand L 43 should be\nc = 2(A sinfi**2 + B cosfi**2)\n\nthese factors are missing in all following calculations. \n\nI have done the exercise. The result should be:\ncase b=0 and a=c:\n A=B=a/2=c/2\ncase b=0 and a!=c:\n a=2A\n c=2B\ncase b!=0:\n tanfi_1/2 = 1/(2b) * [c-a +- sqrt( (a-c)**2 +4b**2)] ",
"reporter": "reimann",
"cc": "kmeagher@ulb.ac.be",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "Missing Factor 2 in python implementation of paraboloid",
"priority": "normal",
"keywords": "paraboloid",
"time": "2015-09-22T15:19:30",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
|
defect
|
missing factor in python implementation of paraboloid trac in paraboloid releases python fit paraboloid py nice docu in code i followed the calculation of the formulars there is an coefficient comparison between l and l which results in relation between coefficients in l here line are missing a factor of l should be a a cosfi b sinfi and l should be c a sinfi b cosfi these factors are missing in all following calculations i have done the exercise the result should be case b and a c a b a c case b and a c a c case b tanfi migrated from json status closed changetime description in paraboloid releases python fit paraboloid py n nnice docu in code i followed the calculation of the formulars nthere is an coefficient comparison between l and l which results in nrelation between coefficients in l nhere line are missing a factor of nl should be na a cosfi b sinfi nand l should be nc a sinfi b cosfi n nthese factors are missing in all following calculations n ni have done the exercise the result should be ncase b and a c n a b a c ncase b and a c n a n c ncase b n tanfi reporter reimann cc kmeagher ulb ac be resolution fixed ts component combo reconstruction summary missing factor in python implementation of paraboloid priority normal keywords paraboloid time milestone owner kjmeagher type defect
| 1
|
68,729
| 21,803,050,904
|
IssuesEvent
|
2022-05-16 07:47:22
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Image metadata tooltip is shown on reply previews
|
T-Defect S-Minor A-Timeline A-Media A-Replies O-Occasional
|
### Steps to reproduce
1. Reply to an image
2. Hover over the reply preview
### Outcome
#### What did you expect?
The image metadata tooltip shouldn't appear, since it's a tiny preview that can't fit that much information.
#### What happened instead?
It's shown, and duplicates the information that's already shown to the right of the image:

### Operating system
NixOS unstable
### Browser information
Firefox 100.0
### URL for webapp
develop.element.io
### Application version
Element version: f080b1fb273e-react-b54c7d8fafe3-js-3e4f02b41e7f Olm version: 3.2.8
### Homeserver
Synapse 1.58.0
### Will you send logs?
No
|
1.0
|
Image metadata tooltip is shown on reply previews - ### Steps to reproduce
1. Reply to an image
2. Hover over the reply preview
### Outcome
#### What did you expect?
The image metadata tooltip shouldn't appear, since it's a tiny preview that can't fit that much information.
#### What happened instead?
It's shown, and duplicates the information that's already shown to the right of the image:

### Operating system
NixOS unstable
### Browser information
Firefox 100.0
### URL for webapp
develop.element.io
### Application version
Element version: f080b1fb273e-react-b54c7d8fafe3-js-3e4f02b41e7f Olm version: 3.2.8
### Homeserver
Synapse 1.58.0
### Will you send logs?
No
|
defect
|
image metadata tooltip is shown on reply previews steps to reproduce reply to an image hover over the reply preview outcome what did you expect the image metadata tooltip shouldn t appear since it s a tiny preview that can t fit that much information what happened instead it s shown and duplicates the information that s already shown to the right of the image operating system nixos unstable browser information firefox url for webapp develop element io application version element version react js olm version homeserver synapse will you send logs no
| 1
|
137,727
| 11,157,620,649
|
IssuesEvent
|
2019-12-25 13:59:10
|
appirio-tech/connect-app
|
https://api.github.com/repos/appirio-tech/connect-app
|
closed
|
If there are multiple copilot join requests, accepting one request shows processing symbol for others too
|
P3 QA Pass in Dev Test Environment
|
1. Login as Manager user and invite 2 Copilots into a project
2. Login as Copilot Manager and approve one of them and check
Actual:
Clicking approve for one-request, shows approval-processing symbol for both the requests.
[false processing.zip](https://github.com/appirio-tech/connect-app/files/3947871/false.processing.zip)
|
1.0
|
If there are multiple copilot join requests, accepting one request shows processing symbol for others too - 1. Login as Manager user and invite 2 Copilots into a project
2. Login as Copilot Manager and approve one of them and check
Actual:
Clicking approve for one-request, shows approval-processing symbol for both the requests.
[false processing.zip](https://github.com/appirio-tech/connect-app/files/3947871/false.processing.zip)
|
non_defect
|
if there are multiple copilot join requests accepting one request shows processing symbol for others too login as manager user and invite copilots into a project login as copilot manager and approve one of them and check actual clicking approve for one request shows approval processing symbol for both the requests
| 0
|
177,062
| 6,573,838,990
|
IssuesEvent
|
2017-09-11 10:19:46
|
LDMW/cms
|
https://api.github.com/repos/LDMW/cms
|
opened
|
Copy change: I am, or know someone who is, feeling - on homepage
|
priority-3
|
Adding in comma.
Currently says:
I am, or know someone who is feeling
Needs to say:
I am, or know someone who is, feeling
|
1.0
|
Copy change: I am, or know someone who is, feeling - on homepage - Adding in comma.
Currently says:
I am, or know someone who is feeling
Needs to say:
I am, or know someone who is, feeling
|
non_defect
|
copy change i am or know someone who is feeling on homepage adding in comma currently says i am or know someone who is feeling needs to say i am or know someone who is feeling
| 0
|
6,446
| 2,610,243,353
|
IssuesEvent
|
2015-02-26 19:17:23
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
台州割包皮包茎手术需要多少钱
|
auto-migrated Priority-Medium Type-Defect
|
```
台州割包皮包茎手术需要多少钱【台州五洲生殖医院】24小时
健康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地�
��:台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐1
04、108、118、198及椒江一金清公交车直达枫南小区,乘坐107、
105、109、112、901、
902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 31 May 2014 at 1:46
|
1.0
|
台州割包皮包茎手术需要多少钱 - ```
台州割包皮包茎手术需要多少钱【台州五洲生殖医院】24小时
健康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地�
��:台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐1
04、108、118、198及椒江一金清公交车直达枫南小区,乘坐107、
105、109、112、901、
902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 31 May 2014 at 1:46
|
defect
|
台州割包皮包茎手术需要多少钱 台州割包皮包茎手术需要多少钱【台州五洲生殖医院】 健康咨询热线 微信号tzwzszyy 医院地� �� (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 1
|
113,744
| 4,568,235,333
|
IssuesEvent
|
2016-09-15 13:55:20
|
PowerlineApp/powerline-mobile
|
https://api.github.com/repos/PowerlineApp/powerline-mobile
|
closed
|
Comment: rate up + rate delete end up with rates_count:1
|
bug P2 - Medium Priority
|
I have a comment under a Post which has 0 rates.
I up rate the comment:
```
Request URL:https://api-dev.powerli.ne/api/v2/post-comments/579/rate
Request Method:POST
Status Code:200 OK
{"rate_value":"up"}
```
and the rate numbers are as expected:
```
"rate_sum":"1","rates_count":"1"
```
Now I change my mind and want to undo my rate (i.e. delete it). I call:
```
Request URL:https://api-dev.powerli.ne/api/v2/post-comments/579/rate
Request Method:POST
Status Code:200 OK
{"rate_value":"delete"}
```
but in response I receive following numbers:
```
,"rate_sum":"0","rates_count":"1"
```
However I would expect the `rates_count` to be `0`.
The consequence of `"rate_sum":"0","rates_count":"1"` is that the computed number of upvotes and downvotes is `0.5`:
<img width="744" alt="screen shot 2016-09-15 at 09 15 07" src="https://cloud.githubusercontent.com/assets/225506/18541364/17c451ae-7b25-11e6-8240-04873f142fef.png">
[here is business logic](https://github.com/PowerlineApp/powerline-mobile/blob/develop/www/js/services/discussion.js#L130) to compute number of upvotes and downvotes. Actually, it would be much better if the backend will just provide number of `upvotes` and `downvotes` for the comment, so that frontend is not required to compute it on its own.
the same applies for comment under every other entity type (poll/user petition)
|
1.0
|
Comment: rate up + rate delete end up with rates_count:1 - I have a comment under a Post which has 0 rates.
I up rate the comment:
```
Request URL:https://api-dev.powerli.ne/api/v2/post-comments/579/rate
Request Method:POST
Status Code:200 OK
{"rate_value":"up"}
```
and the rate numbers are as expected:
```
"rate_sum":"1","rates_count":"1"
```
Now I change my mind and want to undo my rate (i.e. delete it). I call:
```
Request URL:https://api-dev.powerli.ne/api/v2/post-comments/579/rate
Request Method:POST
Status Code:200 OK
{"rate_value":"delete"}
```
but in response I receive following numbers:
```
,"rate_sum":"0","rates_count":"1"
```
However I would expect the `rates_count` to be `0`.
The consequence of `"rate_sum":"0","rates_count":"1"` is that the computed number of upvotes and downvotes is `0.5`:
<img width="744" alt="screen shot 2016-09-15 at 09 15 07" src="https://cloud.githubusercontent.com/assets/225506/18541364/17c451ae-7b25-11e6-8240-04873f142fef.png">
[here is business logic](https://github.com/PowerlineApp/powerline-mobile/blob/develop/www/js/services/discussion.js#L130) to compute number of upvotes and downvotes. Actually, it would be much better if the backend will just provide number of `upvotes` and `downvotes` for the comment, so that frontend is not required to compute it on its own.
the same applies for comment under every other entity type (poll/user petition)
|
non_defect
|
comment rate up rate delete end up with rates count i have a comment under a post which has rates i up rate the comment request url request method post status code ok rate value up and the rate numbers are as expected rate sum rates count now i change my mind and want to undo my rate i e delete it i call request url request method post status code ok rate value delete but in response i receive following numbers rate sum rates count however i would expect the rates count to be the consequence of rate sum rates count is that the computed number of upvotes and downvotes is img width alt screen shot at src to compute number of upvotes and downvotes actually it would be much better if the backend will just provide number of upvotes and downvotes for the comment so that frontend is not required to compute it on its own the same applies for comment under every other entity type poll user petition
| 0
|
31,006
| 8,638,759,756
|
IssuesEvent
|
2018-11-23 15:50:28
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
Ubuntu package with GeoIP missing
|
build change debian-repo
|
Hi,
I installed the goaccess package on Ubuntu 18.04, but this does not seem to have GeoIP enabled:
goaccess: unrecognized option '--geoip-database'
Is there a version with GeoIP avaliable?
Thx in advance
Jan
|
1.0
|
Ubuntu package with GeoIP missing - Hi,
I installed the goaccess package on Ubuntu 18.04, but this does not seem to have GeoIP enabled:
goaccess: unrecognized option '--geoip-database'
Is there a version with GeoIP avaliable?
Thx in advance
Jan
|
non_defect
|
ubuntu package with geoip missing hi i installed the goaccess package on ubuntu but this does not seem to have geoip enabled goaccess unrecognized option geoip database is there a version with geoip avaliable thx in advance jan
| 0
|
15,978
| 2,870,249,021
|
IssuesEvent
|
2015-06-07 00:32:02
|
pdelia/away3d
|
https://api.github.com/repos/pdelia/away3d
|
closed
|
Cube is borken.
|
auto-migrated Priority-Medium Type-Defect
|
#28 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:13Z
```
What steps will reproduce the problem?
1. var cube:Cube = new Cube();
2.
3.
What is the expected output? What do you see instead?
TypeError: Error #1009: Cannot access a property or method of a null object
reference.
at
away3d.primitives::Cube/buildCube()[/Users/Li/Development/Actionscript/Projects/
ITM3
Phase 2/Workspace/away3d_secure/src/away3d/primitives/Cube.as:100]
Please use labels and text to provide additional information.
The error occurs at the buildCube method of cube, line 100:
uv0 = makeUV(1-(face.uv0.u+offU), face.uv0.v+offV);
```
Original issue reported on code.google.com by `paleblue...@gmail.com` on 27 Jan 2009 at 4:33
|
1.0
|
Cube is borken. - #28 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:13Z
```
What steps will reproduce the problem?
1. var cube:Cube = new Cube();
2.
3.
What is the expected output? What do you see instead?
TypeError: Error #1009: Cannot access a property or method of a null object
reference.
at
away3d.primitives::Cube/buildCube()[/Users/Li/Development/Actionscript/Projects/
ITM3
Phase 2/Workspace/away3d_secure/src/away3d/primitives/Cube.as:100]
Please use labels and text to provide additional information.
The error occurs at the buildCube method of cube, line 100:
uv0 = makeUV(1-(face.uv0.u+offU), face.uv0.v+offV);
```
Original issue reported on code.google.com by `paleblue...@gmail.com` on 27 Jan 2009 at 4:33
|
defect
|
cube is borken issue by googlecodeexporter created on what steps will reproduce the problem var cube cube new cube what is the expected output what do you see instead typeerror error cannot access a property or method of a null object reference at primitives cube buildcube users li development actionscript projects phase workspace secure src primitives cube as please use labels and text to provide additional information the error occurs at the buildcube method of cube line makeuv face u offu face v offv original issue reported on code google com by paleblue gmail com on jan at
| 1
|
262,296
| 27,944,429,558
|
IssuesEvent
|
2023-03-24 01:02:33
|
LalithK90/jnSuper
|
https://api.github.com/repos/LalithK90/jnSuper
|
opened
|
CVE-2023-1370 (High) detected in json-smart-2.3.jar
|
Mend: dependency security vulnerability
|
## CVE-2023-1370 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-smart-2.3.jar</b></p></summary>
<p>JSON (JavaScript Object Notation) is a lightweight data-interchange format. It is easy for humans to read and write. It is easy for machines to parse and generate. It is based on a subset of the JavaScript Programming Language, Standard ECMA-262 3rd Edition - December 1999. JSON is a text format that is completely language independent but uses conventions that are familiar to programmers of the C-family of languages, including C, C++, C#, Java, JavaScript, Perl, Python, and many others. These properties make JSON an ideal data-interchange language.</p>
<p>Library home page: <a href="http://www.minidev.net/">http://www.minidev.net/</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /tmp/ws-ua_20210506065133_NUNARB/downloadResource_CPLGLC/20210506065213/json-smart-2.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-test-2.2.4.RELEASE.jar (Root Library)
- json-path-2.4.0.jar
- :x: **json-smart-2.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LalithK90/jnSuper/commit/cd2c79cdbfafc19f783a2a48afe1dda8165cd067">cd2c79cdbfafc19f783a2a48afe1dda8165cd067</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
[Json-smart](https://netplex.github.io/json-smart/) is a performance focused, JSON processor lib. When reaching a ‘[‘ or ‘{‘ character in the JSON input, the code parses an array or an object respectively. It was discovered that the code does not have any limit to the nesting of such arrays or objects. Since the parsing of nested arrays and objects is done recursively, nesting too many of them can cause a stack exhaustion (stack overflow) and crash the software.
<p>Publish Date: 2023-03-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1370>CVE-2023-1370</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://research.jfrog.com/vulnerabilities/stack-exhaustion-in-json-smart-leads-to-denial-of-service-when-parsing-malformed-json-xray-427633/">https://research.jfrog.com/vulnerabilities/stack-exhaustion-in-json-smart-leads-to-denial-of-service-when-parsing-malformed-json-xray-427633/</a></p>
<p>Release Date: 2023-03-22</p>
<p>Fix Resolution (net.minidev:json-smart): 2.4.9</p>
<p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-test): 2.6.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-1370 (High) detected in json-smart-2.3.jar - ## CVE-2023-1370 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-smart-2.3.jar</b></p></summary>
<p>JSON (JavaScript Object Notation) is a lightweight data-interchange format. It is easy for humans to read and write. It is easy for machines to parse and generate. It is based on a subset of the JavaScript Programming Language, Standard ECMA-262 3rd Edition - December 1999. JSON is a text format that is completely language independent but uses conventions that are familiar to programmers of the C-family of languages, including C, C++, C#, Java, JavaScript, Perl, Python, and many others. These properties make JSON an ideal data-interchange language.</p>
<p>Library home page: <a href="http://www.minidev.net/">http://www.minidev.net/</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /tmp/ws-ua_20210506065133_NUNARB/downloadResource_CPLGLC/20210506065213/json-smart-2.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-test-2.2.4.RELEASE.jar (Root Library)
- json-path-2.4.0.jar
- :x: **json-smart-2.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LalithK90/jnSuper/commit/cd2c79cdbfafc19f783a2a48afe1dda8165cd067">cd2c79cdbfafc19f783a2a48afe1dda8165cd067</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
[Json-smart](https://netplex.github.io/json-smart/) is a performance focused, JSON processor lib. When reaching a ‘[‘ or ‘{‘ character in the JSON input, the code parses an array or an object respectively. It was discovered that the code does not have any limit to the nesting of such arrays or objects. Since the parsing of nested arrays and objects is done recursively, nesting too many of them can cause a stack exhaustion (stack overflow) and crash the software.
<p>Publish Date: 2023-03-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1370>CVE-2023-1370</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://research.jfrog.com/vulnerabilities/stack-exhaustion-in-json-smart-leads-to-denial-of-service-when-parsing-malformed-json-xray-427633/">https://research.jfrog.com/vulnerabilities/stack-exhaustion-in-json-smart-leads-to-denial-of-service-when-parsing-malformed-json-xray-427633/</a></p>
<p>Release Date: 2023-03-22</p>
<p>Fix Resolution (net.minidev:json-smart): 2.4.9</p>
<p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter-test): 2.6.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in json smart jar cve high severity vulnerability vulnerable library json smart jar json javascript object notation is a lightweight data interchange format it is easy for humans to read and write it is easy for machines to parse and generate it is based on a subset of the javascript programming language standard ecma edition december json is a text format that is completely language independent but uses conventions that are familiar to programmers of the c family of languages including c c c java javascript perl python and many others these properties make json an ideal data interchange language library home page a href path to dependency file build gradle path to vulnerable library tmp ws ua nunarb downloadresource cplglc json smart jar dependency hierarchy spring boot starter test release jar root library json path jar x json smart jar vulnerable library found in head commit a href found in base branch master vulnerability details is a performance focused json processor lib when reaching a ‘ ‘ or ‘ ‘ character in the json input the code parses an array or an object respectively it was discovered that the code does not have any limit to the nesting of such arrays or objects since the parsing of nested arrays and objects is done recursively nesting too many of them can cause a stack exhaustion stack overflow and crash the software publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution net minidev json smart direct dependency fix resolution org springframework boot spring boot starter test step up your open source security game with mend
| 0
|
25,351
| 4,300,751,185
|
IssuesEvent
|
2016-07-20 03:12:39
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
opened
|
METHOD / METHODS in update_and_rebuild_libmesh
|
C: MOOSE P: normal T: defect
|
### Description of the enhancement or error report
What exactly is going on in `update_and_rebuild_libmesh.sh` with regards to `METHOD` and `METHODS`?
Specifically, what is this?
https://github.com/idaholab/moose/blame/devel/scripts/update_and_rebuild_libmesh.sh#L39
Of course I have `METHOD` and `METHODS` both set... and of course they are set to different things (setting `METHOD="opt dbg oprof"`) doesn't make any sense at all.
Normally you would have `METHODS` set to all the ways you might want to build and then you would set `METHOD` to the one you are currently using right now. This is the way I work 100% of the time (but I usually don't use `update_and_rebuild_libmesh.sh`). But this currently generates an error from `update_and_rebuild_libmesh.sh`!
In fact, I actually noticed a bit of what I just thought was brain damage in a post to moose-users the other day that is because of this. Check out this post from Hans:
https://groups.google.com/d/msg/moose-users/0C-kOiUiTQk/lWGdxkFsAQAJ
When I saw it I was like "Why the heck is he unsetting METHOD before running `update_and_rebuild_libmesh.sh`??? Must just be some leftover brain damage"... but it's actually not! It's actually required by the brain damage that's in `update_and_rebuild_libmesh.sh`!
I'm sure there's already been a discussion about this that I wasn't around for. It looks like @jwpeterson changed this back in April. What is the rationale for the way it's currently coded? If there isn't one then this check needs to be removed as it's not right!
### Rationale for the enhancement or information for reproducing the error
Set both `METHOD` and `METHODS` in your environment and make them not the same.
### Identified impact
Removal of brain damage 😉
|
1.0
|
METHOD / METHODS in update_and_rebuild_libmesh - ### Description of the enhancement or error report
What exactly is going on in `update_and_rebuild_libmesh.sh` with regards to `METHOD` and `METHODS`?
Specifically, what is this?
https://github.com/idaholab/moose/blame/devel/scripts/update_and_rebuild_libmesh.sh#L39
Of course I have `METHOD` and `METHODS` both set... and of course they are set to different things (setting `METHOD="opt dbg oprof"`) doesn't make any sense at all.
Normally you would have `METHODS` set to all the ways you might want to build and then you would set `METHOD` to the one you are currently using right now. This is the way I work 100% of the time (but I usually don't use `update_and_rebuild_libmesh.sh`). But this currently generates an error from `update_and_rebuild_libmesh.sh`!
In fact, I actually noticed a bit of what I just thought was brain damage in a post to moose-users the other day that is because of this. Check out this post from Hans:
https://groups.google.com/d/msg/moose-users/0C-kOiUiTQk/lWGdxkFsAQAJ
When I saw it I was like "Why the heck is he unsetting METHOD before running `update_and_rebuild_libmesh.sh`??? Must just be some leftover brain damage"... but it's actually not! It's actually required by the brain damage that's in `update_and_rebuild_libmesh.sh`!
I'm sure there's already been a discussion about this that I wasn't around for. It looks like @jwpeterson changed this back in April. What is the rationale for the way it's currently coded? If there isn't one then this check needs to be removed as it's not right!
### Rationale for the enhancement or information for reproducing the error
Set both `METHOD` and `METHODS` in your environment and make them not the same.
### Identified impact
Removal of brain damage 😉
|
defect
|
method methods in update and rebuild libmesh description of the enhancement or error report what exactly is going on in update and rebuild libmesh sh with regards to method and methods specifically what is this of course i have method and methods both set and of course they are set to different things setting method opt dbg oprof doesn t make any sense at all normally you would have methods set to all the ways you might want to build and then you would set method to the one you are currently using right now this is the way i work of the time but i usually don t use update and rebuild libmesh sh but this currently generates an error from update and rebuild libmesh sh in fact i actually noticed a bit of what i just thought was brain damage in a post to moose users the other day that is because of this check out this post from hans when i saw it i was like why the heck is he unsetting method before running update and rebuild libmesh sh must just be some leftover brain damage but it s actually not it s actually required by the brain damage that s in update and rebuild libmesh sh i m sure there s already been a discussion about this that i wasn t around for it looks like jwpeterson changed this back in april what is the rationale for the way it s currently coded if there isn t one then this check needs to be removed as it s not right rationale for the enhancement or information for reproducing the error set both method and methods in your environment and make them not the same identified impact removal of brain damage 😉
| 1
|
279,224
| 30,702,475,516
|
IssuesEvent
|
2023-07-27 01:33:20
|
nidhi7598/linux-3.0.35_CVE-2018-13405
|
https://api.github.com/repos/nidhi7598/linux-3.0.35_CVE-2018-13405
|
closed
|
CVE-2022-1462 (Medium) detected in linux-stable-rtv3.8.6, linux-stable-rtv3.8.6 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2022-1462 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv3.8.6</b>, <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An out-of-bounds read flaw was found in the Linux kernel’s TeleTYpe subsystem. The issue occurs in how a user triggers a race condition using ioctls TIOCSPTLCK and TIOCGPTPEER and TIOCSTI and TCXONC with leakage of memory in the flush_to_ldisc function. This flaw allows a local user to crash the system or read unauthorized random data from memory.
<p>Publish Date: 2022-06-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1462>CVE-2022-1462</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1462">https://www.linuxkernelcves.com/cves/CVE-2022-1462</a></p>
<p>Release Date: 2022-06-02</p>
<p>Fix Resolution: v4.9.325,v4.14.290,v4.19.254,v5.4.208,v5.10.134,v5.15.58,v5.18.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-1462 (Medium) detected in linux-stable-rtv3.8.6, linux-stable-rtv3.8.6 - autoclosed - ## CVE-2022-1462 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv3.8.6</b>, <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An out-of-bounds read flaw was found in the Linux kernel’s TeleTYpe subsystem. The issue occurs in how a user triggers a race condition using ioctls TIOCSPTLCK and TIOCGPTPEER and TIOCSTI and TCXONC with leakage of memory in the flush_to_ldisc function. This flaw allows a local user to crash the system or read unauthorized random data from memory.
<p>Publish Date: 2022-06-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1462>CVE-2022-1462</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1462">https://www.linuxkernelcves.com/cves/CVE-2022-1462</a></p>
<p>Release Date: 2022-06-02</p>
<p>Fix Resolution: v4.9.325,v4.14.290,v4.19.254,v5.4.208,v5.10.134,v5.15.58,v5.18.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in linux stable linux stable autoclosed cve medium severity vulnerability vulnerable libraries linux stable linux stable vulnerability details an out of bounds read flaw was found in the linux kernel’s teletype subsystem the issue occurs in how a user triggers a race condition using ioctls tiocsptlck and tiocgptpeer and tiocsti and tcxonc with leakage of memory in the flush to ldisc function this flaw allows a local user to crash the system or read unauthorized random data from memory publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
154,299
| 12,199,133,872
|
IssuesEvent
|
2020-04-30 00:43:44
|
Princeton-CDH/mep-django
|
https://api.github.com/repos/Princeton-CDH/mep-django
|
closed
|
As a user I want items automatically sorted by relevance if I have a keyword search term active and otherwise by title (by default), so that I see best matches first for keyword searches.
|
awaiting testing
|
# testing notes
this needs to be tested on the books search page. check that:
- if you haven't typed anything the keyword search box, you shouldn't be allowed to select the "relevance" option
- as soon as you type in a keyword search, the sort should switch to relevance automatically
- after that, if you want, you can pick any of the search options and they should work
- if you have relevance selected and you delete everything in the search box, you should be switched off of relevance and it should become disabled again.
## dev notes
- [x] disable relevance sort option if no keyword search (django form)
- [x] default to relevance sort if there is a keyword search term (django view)
- [x] switch to relevance sort after user types keyword search term (js reactive search)
- [x] disable relevance sort and switch to default if keyword search is cleared (reactive search)
|
1.0
|
As a user I want items automatically sorted by relevance if I have a keyword search term active and otherwise by title (by default), so that I see best matches first for keyword searches. - # testing notes
this needs to be tested on the books search page. check that:
- if you haven't typed anything the keyword search box, you shouldn't be allowed to select the "relevance" option
- as soon as you type in a keyword search, the sort should switch to relevance automatically
- after that, if you want, you can pick any of the search options and they should work
- if you have relevance selected and you delete everything in the search box, you should be switched off of relevance and it should become disabled again.
## dev notes
- [x] disable relevance sort option if no keyword search (django form)
- [x] default to relevance sort if there is a keyword search term (django view)
- [x] switch to relevance sort after user types keyword search term (js reactive search)
- [x] disable relevance sort and switch to default if keyword search is cleared (reactive search)
|
non_defect
|
as a user i want items automatically sorted by relevance if i have a keyword search term active and otherwise by title by default so that i see best matches first for keyword searches testing notes this needs to be tested on the books search page check that if you haven t typed anything the keyword search box you shouldn t be allowed to select the relevance option as soon as you type in a keyword search the sort should switch to relevance automatically after that if you want you can pick any of the search options and they should work if you have relevance selected and you delete everything in the search box you should be switched off of relevance and it should become disabled again dev notes disable relevance sort option if no keyword search django form default to relevance sort if there is a keyword search term django view switch to relevance sort after user types keyword search term js reactive search disable relevance sort and switch to default if keyword search is cleared reactive search
| 0
|
32,701
| 6,896,973,432
|
IssuesEvent
|
2017-11-23 21:44:38
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
Decode request data
|
Defect http On hold
|
This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.5.5
* Platform and Target: AWS
### What you did
Hi there.
On my cakephp App, when a user creates an account / logins, the password is encoded on clients side using public/private key method.
On the App, we decode it, and create the password/check the login information. I've used the following command:
$this->request->data['password'] = $this->RsaDecode($this->request->data['password']); and it worked nicely.
With the version 3.5.5, I've changed every $this->request->data['password'] field with $this->request->getData['password']. Yet, when I call the identify, it returns false.
If this is no way to make it like it was, is there any workaround that I can do it?
### What happened
Password gets decoded but identify fails.
### What you expected to happen
Password to be decoded, saved into the request array again to the user be identified
|
1.0
|
Decode request data - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.5.5
* Platform and Target: AWS
### What you did
Hi there.
On my cakephp App, when a user creates an account / logins, the password is encoded on clients side using public/private key method.
On the App, we decode it, and create the password/check the login information. I've used the following command:
$this->request->data['password'] = $this->RsaDecode($this->request->data['password']); and it worked nicely.
With the version 3.5.5, I've changed every $this->request->data['password'] field with $this->request->getData['password']. Yet, when I call the identify, it returns false.
If this is no way to make it like it was, is there any workaround that I can do it?
### What happened
Password gets decoded but identify fails.
### What you expected to happen
Password to be decoded, saved into the request array again to the user be identified
|
defect
|
decode request data this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target aws what you did hi there on my cakephp app when a user creates an account logins the password is encoded on clients side using public private key method on the app we decode it and create the password check the login information i ve used the following command this request data this rsadecode this request data and it worked nicely with the version i ve changed every this request data field with this request getdata yet when i call the identify it returns false if this is no way to make it like it was is there any workaround that i can do it what happened password gets decoded but identify fails what you expected to happen password to be decoded saved into the request array again to the user be identified
| 1
|
227,988
| 18,145,241,711
|
IssuesEvent
|
2021-09-25 10:01:24
|
stylelint/stylelint
|
https://api.github.com/repos/stylelint/stylelint
|
closed
|
Add tests for Markdown
|
status: ready to implement type: tests good first issue
|
Looks like we don't have tests for Markdown. In https://github.com/stylelint/stylelint/pull/4239 markdown parser was removed and tests didn't fail, but they should have. Because we support Markdown out of the box.
|
1.0
|
Add tests for Markdown - Looks like we don't have tests for Markdown. In https://github.com/stylelint/stylelint/pull/4239 markdown parser was removed and tests didn't fail, but they should have. Because we support Markdown out of the box.
|
non_defect
|
add tests for markdown looks like we don t have tests for markdown in markdown parser was removed and tests didn t fail but they should have because we support markdown out of the box
| 0
|
24,963
| 4,155,727,118
|
IssuesEvent
|
2016-06-16 15:43:42
|
extnet/Ext.NET
|
https://api.github.com/repos/extnet/Ext.NET
|
closed
|
Direct Method call translates null value into a string containing "null"
|
3.x 4.x defect
|
Reported here: [Calling a DirectMethod with a string parameter with null as an argument results in the string "null" on the server](http://forums.ext.net/showthread.php?61191)
The server sent HTTP POST payload gets translated into `ExtraParams{"src":null}` without quotes in the 'null' parameter, but until it gets to the server, the `null` keyword becomes the `"null"` string.
|
1.0
|
Direct Method call translates null value into a string containing "null" - Reported here: [Calling a DirectMethod with a string parameter with null as an argument results in the string "null" on the server](http://forums.ext.net/showthread.php?61191)
The server sent HTTP POST payload gets translated into `ExtraParams{"src":null}` without quotes in the 'null' parameter, but until it gets to the server, the `null` keyword becomes the `"null"` string.
|
defect
|
direct method call translates null value into a string containing null reported here the server sent http post payload gets translated into extraparams src null without quotes in the null parameter but until it gets to the server the null keyword becomes the null string
| 1
|
70,186
| 23,041,360,056
|
IssuesEvent
|
2022-07-23 07:23:58
|
unascribed/Fabrication
|
https://api.github.com/repos/unascribed/Fabrication
|
opened
|
FabInjector failed to find injection point
|
k: Defect n: Fabric s: New
|
[23/07/2022 14:03:02 PM] [Render thread/ERROR] [Fabrication] FabInjector failed to find injection point for com/unascribed/fabrication/logic/CommandSourceInterfaceHack;method_9268(Ljava/lang/Iterable;Ljava/lang/String;Ljava/util/function/Function;Ljava/util/function/Consumer;)V Ljava/lang/String;equals(Ljava/lang/Object;)Z
[23/07/2022 14:03:02 PM] [Render thread/INFO] [EssentialCommands]: Mod Load Complete.
[23/07/2022 14:03:02 PM] [Render thread/INFO] Trying to read config file...
[23/07/2022 14:03:02 PM] [Render thread/INFO] A config file was found, loading it..
[23/07/2022 14:03:02 PM] [Render thread/INFO] Successfully loaded config file.
[23/07/2022 14:03:02 PM] [Render thread/INFO] Successfully read config file **USERSDIR**\instances\Fabric\config\extshape.nbt.
[23/07/2022 14:03:02 PM] [Render thread/INFO] Successfully wrote config file **USERSDIR**\instances\Fabric\config\extshape.nbt.
[23/07/2022 14:03:03 PM] [Render thread/INFO] Generating server data for Extended Block Shapes mod!
[23/07/2022 14:03:04 PM] [Render thread/INFO] Generating client resources for Extended Block Shapes mod!
[23/07/2022 14:03:04 PM] [ForkJoinPool.commonPool-worker-1/ERROR] Couldn't get latest version!
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of dimensional_tools.ini done in 4.868 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of block_logo.png done in 13.41 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of block_logo.ini done in 15.30 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of classic_block_drops.ini done in 2.609 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of fscript.ini done in 2.092 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of gear_components.ini done in 13.86 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of item_despawn.ini done in 699.8 μs with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of taggable_players.ini done in 5.954 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of yeet_recipes.ini done in 584.1 μs with 0 warnings
Not Crash but maybe can be useful information :
Full Log : https://paste.atlauncher.com/view/3dfea829-a191-4dae-9093-d96197b48ef2
|
1.0
|
FabInjector failed to find injection point - [23/07/2022 14:03:02 PM] [Render thread/ERROR] [Fabrication] FabInjector failed to find injection point for com/unascribed/fabrication/logic/CommandSourceInterfaceHack;method_9268(Ljava/lang/Iterable;Ljava/lang/String;Ljava/util/function/Function;Ljava/util/function/Consumer;)V Ljava/lang/String;equals(Ljava/lang/Object;)Z
[23/07/2022 14:03:02 PM] [Render thread/INFO] [EssentialCommands]: Mod Load Complete.
[23/07/2022 14:03:02 PM] [Render thread/INFO] Trying to read config file...
[23/07/2022 14:03:02 PM] [Render thread/INFO] A config file was found, loading it..
[23/07/2022 14:03:02 PM] [Render thread/INFO] Successfully loaded config file.
[23/07/2022 14:03:02 PM] [Render thread/INFO] Successfully read config file **USERSDIR**\instances\Fabric\config\extshape.nbt.
[23/07/2022 14:03:02 PM] [Render thread/INFO] Successfully wrote config file **USERSDIR**\instances\Fabric\config\extshape.nbt.
[23/07/2022 14:03:03 PM] [Render thread/INFO] Generating server data for Extended Block Shapes mod!
[23/07/2022 14:03:04 PM] [Render thread/INFO] Generating client resources for Extended Block Shapes mod!
[23/07/2022 14:03:04 PM] [ForkJoinPool.commonPool-worker-1/ERROR] Couldn't get latest version!
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of dimensional_tools.ini done in 4.868 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of block_logo.png done in 13.41 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of block_logo.ini done in 15.30 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of classic_block_drops.ini done in 2.609 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of fscript.ini done in 2.092 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of gear_components.ini done in 13.86 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of item_despawn.ini done in 699.8 μs with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of taggable_players.ini done in 5.954 ms with 0 warnings
[23/07/2022 14:03:04 PM] [Render thread/INFO] [Fabrication] Loading of yeet_recipes.ini done in 584.1 μs with 0 warnings
Not Crash but maybe can be useful information :
Full Log : https://paste.atlauncher.com/view/3dfea829-a191-4dae-9093-d96197b48ef2
|
defect
|
fabinjector failed to find injection point fabinjector failed to find injection point for com unascribed fabrication logic commandsourceinterfacehack method ljava lang iterable ljava lang string ljava util function function ljava util function consumer v ljava lang string equals ljava lang object z mod load complete trying to read config file a config file was found loading it successfully loaded config file successfully read config file usersdir instances fabric config extshape nbt successfully wrote config file usersdir instances fabric config extshape nbt generating server data for extended block shapes mod generating client resources for extended block shapes mod couldn t get latest version loading of dimensional tools ini done in ms with warnings loading of block logo png done in ms with warnings loading of block logo ini done in ms with warnings loading of classic block drops ini done in ms with warnings loading of fscript ini done in ms with warnings loading of gear components ini done in ms with warnings loading of item despawn ini done in î¼s with warnings loading of taggable players ini done in ms with warnings loading of yeet recipes ini done in î¼s with warnings not crash but maybe can be useful information full log
| 1
|
42,017
| 2,869,093,770
|
IssuesEvent
|
2015-06-05 23:17:01
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
add Mapping.urls property to sourcemaps
|
Area-Pkg Pkg-SourceMaps Priority-Medium Triaged Type-Enhancement
|
Bug reported by @skybrian in the old github repo (https://github.com/dart-lang/source-maps/issues/3)
Original bug description:
------------
It seems like you should be able to say parse(json).urls to get all the source URL's, regardless of whether it's a SingleMapping or a MultiSectionMapping. To support this, Mapping should have an abstract "urls" getter which is implemented by MultiSectionMapping to combine all the url lists from the individual mappings.
Workaround: implement this yourself; it's not that hard.
------------
|
1.0
|
add Mapping.urls property to sourcemaps - Bug reported by @skybrian in the old github repo (https://github.com/dart-lang/source-maps/issues/3)
Original bug description:
------------
It seems like you should be able to say parse(json).urls to get all the source URL's, regardless of whether it's a SingleMapping or a MultiSectionMapping. To support this, Mapping should have an abstract "urls" getter which is implemented by MultiSectionMapping to combine all the url lists from the individual mappings.
Workaround: implement this yourself; it's not that hard.
------------
|
non_defect
|
add mapping urls property to sourcemaps bug reported by skybrian in the old github repo original bug description it seems like you should be able to say parse json urls to get all the source url s regardless of whether it s a singlemapping or a multisectionmapping to support this mapping should have an abstract quot urls quot getter which is implemented by multisectionmapping to combine all the url lists from the individual mappings workaround implement this yourself it s not that hard
| 0
|
27,904
| 30,656,052,015
|
IssuesEvent
|
2023-07-25 12:14:04
|
inmanta/inmanta-core
|
https://api.github.com/repos/inmanta/inmanta-core
|
closed
|
Environment creation results in response code 500
|
bug usability tiny
|
The Inmanta server returns the response code 500 when the user tries to create an environment in a certain project with a name that already exists in that project. The endpoint to update an environment probably suffers from the same bug.
- [x] Fix the bug
- [x] Mention in the API documentation that the names of all environments in a certain project must have unique names.
|
True
|
Environment creation results in response code 500 - The Inmanta server returns the response code 500 when the user tries to create an environment in a certain project with a name that already exists in that project. The endpoint to update an environment probably suffers from the same bug.
- [x] Fix the bug
- [x] Mention in the API documentation that the names of all environments in a certain project must have unique names.
|
non_defect
|
environment creation results in response code the inmanta server returns the response code when the user tries to create an environment in a certain project with a name that already exists in that project the endpoint to update an environment probably suffers from the same bug fix the bug mention in the api documentation that the names of all environments in a certain project must have unique names
| 0
|
35,473
| 7,749,113,513
|
IssuesEvent
|
2018-05-30 10:21:43
|
contactlab/contacthub-sdk-java
|
https://api.github.com/repos/contactlab/contacthub-sdk-java
|
closed
|
Make fields of contextInfo optional
|
defect waiting for merge
|
[Trello card](https://trello.com/c/5b0e76983d0a04b49fc62d3f/)
Fields of contextInfo and its subclasses should be optional, in order to align them with the API
|
1.0
|
Make fields of contextInfo optional - [Trello card](https://trello.com/c/5b0e76983d0a04b49fc62d3f/)
Fields of contextInfo and its subclasses should be optional, in order to align them with the API
|
defect
|
make fields of contextinfo optional fields of contextinfo and its subclasses should be optional in order to align them with the api
| 1
|
445,232
| 31,224,709,034
|
IssuesEvent
|
2023-08-19 00:54:07
|
marcosdosea/PetLife
|
https://api.github.com/repos/marcosdosea/PetLife
|
opened
|
Realizar atualização das propriedades dos Contexts no Design Model
|
documentation
|
Observe DB_Context e PersistenciaEditora
Exemplo de como é após a atualização das propriedades

Exemplo de como é antes da atualização das propriedades

|
1.0
|
Realizar atualização das propriedades dos Contexts no Design Model - Observe DB_Context e PersistenciaEditora
Exemplo de como é após a atualização das propriedades

Exemplo de como é antes da atualização das propriedades

|
non_defect
|
realizar atualização das propriedades dos contexts no design model observe db context e persistenciaeditora exemplo de como é após a atualização das propriedades exemplo de como é antes da atualização das propriedades
| 0
|
713,654
| 24,534,621,695
|
IssuesEvent
|
2022-10-11 19:31:53
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
[extension/sigv4authextension] Auth failed without `sts_region` field
|
bug priority:p2 extension/sigv4auth
|
### What happened?
## Description
Sigv4 auth extension fails initialization inside an EKS-Anywhere cluster without `sts_region` value. The cluster is setup with IRSA and the service account is attached.
## Steps to Reproduce
1. Setup cluster
2. [Setup IRSA](https://anywhere.eks.amazonaws.com/docs/reference/clusterspec/optional/irsa/)
3. Add service account
4. Ensure service account is attached to deployment manifest
5. Run collector with attached config
## Expected Result
collector to start
## Actual Result
Collector fails to start due to sigv4 error
### Collector version
ADOT Collector v0.21.1-6b7244d. Uses upstream collector version `v0.60.0`
### Environment information
EKS-Anywhere cluster using vSphere provider, os Bottlerocket, Kubernetes 1.21
### OpenTelemetry Collector configuration
```yaml
exporters:
logging:
logLevel: info
prometheusremotewrite:
auth:
authenticator: sigv4auth
endpoint: https://aps-workspaces.us-west-2.amazonaws.com/workspaces/<redacted-namespace>/v1/api/remote_write
extensions:
health_check: {}
memory_ballast:
size_mib: "819"
sigv4auth:
region: us-west-2
service: aps
processors:
batch: {}
memory_limiter:
check_interval: 5s
limit_mib: 1638
spike_limit_mib: 512
receivers:
otlp:
protocols:
grpc:
endpoint: 0.0.0.0:4317
http:
endpoint: 0.0.0.0:4318
prometheus:
config:
scrape_configs:
- job_name: opentelemetry-collector
scrape_interval: 10s
static_configs:
- targets:
- ${MY_POD_IP}:8888
service:
extensions:
- health_check
- memory_ballast
- sigv4auth
pipelines:
metrics:
exporters:
- logging
- prometheusremotewrite
processors:
- batch
receivers:
- prometheus
telemetry:
metrics:
address: 0.0.0.0:8888
```
### Log output
```shell
2022/09/28 16:37:19 ADOT Collector version: v0.21.1
2022/09/28 16:37:19 found no extra config, skip it, err: open /opt/aws/aws-otel-collector/etc/extracfg.txt: no such file or directory
Error: failed to get config: invalid configuration: extension "sigv4auth" has invalid configuration: could not retrieve credential provider: failed to refresh cached credentials, failed to retrieve credentials, operation error STS: AssumeRoleWithWebIdentity, failed to resolve service endpoint, an AWS region is required, but was not found
2022/09/28 16:37:19 application run finished with error: failed to get config: invalid configuration: extension "sigv4auth" has invalid configuration: could not retrieve credential provider: failed to refresh cached credentials, failed to retrieve credentials, operation error STS: AssumeRoleWithWebIdentity, failed to resolve service endpoint, an AWS region is required, but was not found
```
### Additional context
Ideally, if `sts_region` isn't provided then the default `region` or `aws_region` value is used.
|
1.0
|
[extension/sigv4authextension] Auth failed without `sts_region` field - ### What happened?
## Description
Sigv4 auth extension fails initialization inside an EKS-Anywhere cluster without `sts_region` value. The cluster is setup with IRSA and the service account is attached.
## Steps to Reproduce
1. Setup cluster
2. [Setup IRSA](https://anywhere.eks.amazonaws.com/docs/reference/clusterspec/optional/irsa/)
3. Add service account
4. Ensure service account is attached to deployment manifest
5. Run collector with attached config
## Expected Result
collector to start
## Actual Result
Collector fails to start due to sigv4 error
### Collector version
ADOT Collector v0.21.1-6b7244d. Uses upstream collector version `v0.60.0`
### Environment information
EKS-Anywhere cluster using vSphere provider, os Bottlerocket, Kubernetes 1.21
### OpenTelemetry Collector configuration
```yaml
exporters:
logging:
logLevel: info
prometheusremotewrite:
auth:
authenticator: sigv4auth
endpoint: https://aps-workspaces.us-west-2.amazonaws.com/workspaces/<redacted-namespace>/v1/api/remote_write
extensions:
health_check: {}
memory_ballast:
size_mib: "819"
sigv4auth:
region: us-west-2
service: aps
processors:
batch: {}
memory_limiter:
check_interval: 5s
limit_mib: 1638
spike_limit_mib: 512
receivers:
otlp:
protocols:
grpc:
endpoint: 0.0.0.0:4317
http:
endpoint: 0.0.0.0:4318
prometheus:
config:
scrape_configs:
- job_name: opentelemetry-collector
scrape_interval: 10s
static_configs:
- targets:
- ${MY_POD_IP}:8888
service:
extensions:
- health_check
- memory_ballast
- sigv4auth
pipelines:
metrics:
exporters:
- logging
- prometheusremotewrite
processors:
- batch
receivers:
- prometheus
telemetry:
metrics:
address: 0.0.0.0:8888
```
### Log output
```shell
2022/09/28 16:37:19 ADOT Collector version: v0.21.1
2022/09/28 16:37:19 found no extra config, skip it, err: open /opt/aws/aws-otel-collector/etc/extracfg.txt: no such file or directory
Error: failed to get config: invalid configuration: extension "sigv4auth" has invalid configuration: could not retrieve credential provider: failed to refresh cached credentials, failed to retrieve credentials, operation error STS: AssumeRoleWithWebIdentity, failed to resolve service endpoint, an AWS region is required, but was not found
2022/09/28 16:37:19 application run finished with error: failed to get config: invalid configuration: extension "sigv4auth" has invalid configuration: could not retrieve credential provider: failed to refresh cached credentials, failed to retrieve credentials, operation error STS: AssumeRoleWithWebIdentity, failed to resolve service endpoint, an AWS region is required, but was not found
```
### Additional context
Ideally, if `sts_region` isn't provided then the default `region` or `aws_region` value is used.
|
non_defect
|
auth failed without sts region field what happened description auth extension fails initialization inside an eks anywhere cluster without sts region value the cluster is setup with irsa and the service account is attached steps to reproduce setup cluster add service account ensure service account is attached to deployment manifest run collector with attached config expected result collector to start actual result collector fails to start due to error collector version adot collector uses upstream collector version environment information eks anywhere cluster using vsphere provider os bottlerocket kubernetes opentelemetry collector configuration yaml exporters logging loglevel info prometheusremotewrite auth authenticator endpoint extensions health check memory ballast size mib region us west service aps processors batch memory limiter check interval limit mib spike limit mib receivers otlp protocols grpc endpoint http endpoint prometheus config scrape configs job name opentelemetry collector scrape interval static configs targets my pod ip service extensions health check memory ballast pipelines metrics exporters logging prometheusremotewrite processors batch receivers prometheus telemetry metrics address log output shell adot collector version found no extra config skip it err open opt aws aws otel collector etc extracfg txt no such file or directory error failed to get config invalid configuration extension has invalid configuration could not retrieve credential provider failed to refresh cached credentials failed to retrieve credentials operation error sts assumerolewithwebidentity failed to resolve service endpoint an aws region is required but was not found application run finished with error failed to get config invalid configuration extension has invalid configuration could not retrieve credential provider failed to refresh cached credentials failed to retrieve credentials operation error sts assumerolewithwebidentity failed to resolve service endpoint an aws region is required but was not found additional context ideally if sts region isn t provided then the default region or aws region value is used
| 0
|
247,230
| 26,688,883,693
|
IssuesEvent
|
2023-01-27 01:32:16
|
Nidhi77777/linux-4.1.15
|
https://api.github.com/repos/Nidhi77777/linux-4.1.15
|
reopened
|
CVE-2015-8962 (High) detected in linuxlinux-4.6
|
security vulnerability
|
## CVE-2015-8962 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Nidhi77777/linux-4.1.15/commit/9098330f8b0515526a6a4fb9c57f6a522a73231f">9098330f8b0515526a6a4fb9c57f6a522a73231f</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/sg.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/sg.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Double free vulnerability in the sg_common_write function in drivers/scsi/sg.c in the Linux kernel before 4.4 allows local users to gain privileges or cause a denial of service (memory corruption and system crash) by detaching a device during an SG_IO ioctl call.
<p>Publish Date: 2016-11-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-8962>CVE-2015-8962</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-8962">https://nvd.nist.gov/vuln/detail/CVE-2015-8962</a></p>
<p>Release Date: 2016-11-16</p>
<p>Fix Resolution: 4.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2015-8962 (High) detected in linuxlinux-4.6 - ## CVE-2015-8962 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Nidhi77777/linux-4.1.15/commit/9098330f8b0515526a6a4fb9c57f6a522a73231f">9098330f8b0515526a6a4fb9c57f6a522a73231f</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/sg.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/sg.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Double free vulnerability in the sg_common_write function in drivers/scsi/sg.c in the Linux kernel before 4.4 allows local users to gain privileges or cause a denial of service (memory corruption and system crash) by detaching a device during an SG_IO ioctl call.
<p>Publish Date: 2016-11-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-8962>CVE-2015-8962</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-8962">https://nvd.nist.gov/vuln/detail/CVE-2015-8962</a></p>
<p>Release Date: 2016-11-16</p>
<p>Fix Resolution: 4.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers scsi sg c drivers scsi sg c vulnerability details double free vulnerability in the sg common write function in drivers scsi sg c in the linux kernel before allows local users to gain privileges or cause a denial of service memory corruption and system crash by detaching a device during an sg io ioctl call publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
13,079
| 2,732,892,870
|
IssuesEvent
|
2015-04-17 10:02:20
|
tiku01/oryx-editor
|
https://api.github.com/repos/tiku01/oryx-editor
|
closed
|
Perfomance Issue: Batch Publishing models
|
auto-migrated Priority-High Type-Defect
|
```
What steps will reproduce the problem?
1. publish (at least) 24 models at once (repo->List View->Multiselect and
publish)
2.
3.
What is the expected output?
- ui feedback about the status
What do you see instead?
- blocking ui & server communication failed.
Please provide any additional information below.
I had this before but now (19:57 16.12.2008) the server did not respond for
minutes and even
afterwards (browser restart + relogin) the repository does not respond
properly. Models are not
loaded. It seems that the server is wracked. uups.
```
Original issue reported on code.google.com by `alexande...@googlemail.com` on 16 Dec 2008 at 7:07
Attachments:
* [performance.tiff](https://storage.googleapis.com/google-code-attachments/oryx-editor/issue-342/comment-0/performance.tiff)
* [killed.tiff](https://storage.googleapis.com/google-code-attachments/oryx-editor/issue-342/comment-0/killed.tiff)
|
1.0
|
Perfomance Issue: Batch Publishing models - ```
What steps will reproduce the problem?
1. publish (at least) 24 models at once (repo->List View->Multiselect and
publish)
2.
3.
What is the expected output?
- ui feedback about the status
What do you see instead?
- blocking ui & server communication failed.
Please provide any additional information below.
I had this before but now (19:57 16.12.2008) the server did not respond for
minutes and even
afterwards (browser restart + relogin) the repository does not respond
properly. Models are not
loaded. It seems that the server is wracked. uups.
```
Original issue reported on code.google.com by `alexande...@googlemail.com` on 16 Dec 2008 at 7:07
Attachments:
* [performance.tiff](https://storage.googleapis.com/google-code-attachments/oryx-editor/issue-342/comment-0/performance.tiff)
* [killed.tiff](https://storage.googleapis.com/google-code-attachments/oryx-editor/issue-342/comment-0/killed.tiff)
|
defect
|
perfomance issue batch publishing models what steps will reproduce the problem publish at least models at once repo list view multiselect and publish what is the expected output ui feedback about the status what do you see instead blocking ui server communication failed please provide any additional information below i had this before but now the server did not respond for minutes and even afterwards browser restart relogin the repository does not respond properly models are not loaded it seems that the server is wracked uups original issue reported on code google com by alexande googlemail com on dec at attachments
| 1
|
20,463
| 3,599,750,337
|
IssuesEvent
|
2016-02-03 00:47:56
|
flutter/flutter
|
https://api.github.com/repos/flutter/flutter
|
opened
|
Need a way to specify an appBar "detail view" whose size is based on the screen's width
|
affects: material design
|
The figure on the right side of the "Detail View" section of
https://www.google.com/design/spec/layout/metrics-keylines.html#metrics-keylines-keylines-spacing says that the size of the area below the toolbar/statusbar should have an aspect ration of 3:2. The current API really needs a fixed appBar height.
See https://github.com/flutter/flutter/pull/1540
|
1.0
|
Need a way to specify an appBar "detail view" whose size is based on the screen's width - The figure on the right side of the "Detail View" section of
https://www.google.com/design/spec/layout/metrics-keylines.html#metrics-keylines-keylines-spacing says that the size of the area below the toolbar/statusbar should have an aspect ration of 3:2. The current API really needs a fixed appBar height.
See https://github.com/flutter/flutter/pull/1540
|
non_defect
|
need a way to specify an appbar detail view whose size is based on the screen s width the figure on the right side of the detail view section of says that the size of the area below the toolbar statusbar should have an aspect ration of the current api really needs a fixed appbar height see
| 0
|
64,061
| 7,761,266,791
|
IssuesEvent
|
2018-06-01 09:18:08
|
researchstudio-sat/webofneeds
|
https://api.github.com/repos/researchstudio-sat/webofneeds
|
closed
|
Won.js as seperate npm-module/package
|
design decision suggestion
|
We could bundle everything that's contained in the won.js (and anything else that's not angular-dependent) into sub-projects/modules/packages and publish them to npm once they're finished, so others have an easier time writing their own clients.
|
1.0
|
Won.js as seperate npm-module/package - We could bundle everything that's contained in the won.js (and anything else that's not angular-dependent) into sub-projects/modules/packages and publish them to npm once they're finished, so others have an easier time writing their own clients.
|
non_defect
|
won js as seperate npm module package we could bundle everything that s contained in the won js and anything else that s not angular dependent into sub projects modules packages and publish them to npm once they re finished so others have an easier time writing their own clients
| 0
|
503,836
| 14,598,944,105
|
IssuesEvent
|
2020-12-21 02:38:03
|
magento/magento2
|
https://api.github.com/repos/magento/magento2
|
closed
|
Redundant header generated while add a product to grouped product
|
Component: Admin Component: GroupedProduct Fixed in 2.4.x Issue: Clear Description Issue: Confirmed Issue: ready for confirmation Priority: P2 Progress: PR in progress Reproduced on 2.4.x Severity: S3
|
### Preconditions (*)
1. M2.3.x
2. M2.4.0
### Steps to reproduce (*)
1. Go Admin > Catalog > Products
2. Select Grouped Product from drop down of Add Product button
3. In new product page, at Attribute Set select, choose Top
4. Scroll down to tab Grouped Products and try add new product.
5. Got an issue [https://imgur.com/porq478](https://imgur.com/porq478)
6. Loom: https://www.loom.com/share/b28e534af9254eec99bf5fc4e77a0376
### Expected result (*)
1. [https://imgur.com/iIjvWuq](https://imgur.com/iIjvWuq)
### Actual result (*)
1. [https://imgur.com/porq478](https://imgur.com/porq478)
---
Please provide [Severity](https://devdocs.magento.com/guides/v2.3/contributor-guide/contributing.html#backlog) assessment for the Issue as Reporter. This information will help during Confirmation and Issue triage processes.
- [ ] Severity: **S0** _- Affects critical data or functionality and leaves users without workaround._
- [ ] Severity: **S1** _- Affects critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S2** _- Affects non-critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S3** _- Affects non-critical data or functionality and does not force users to employ a workaround._
- [ ] Severity: **S4** _- Affects aesthetics, professional look and feel, “quality” or “usability”.
|
1.0
|
Redundant header generated while add a product to grouped product - ### Preconditions (*)
1. M2.3.x
2. M2.4.0
### Steps to reproduce (*)
1. Go Admin > Catalog > Products
2. Select Grouped Product from drop down of Add Product button
3. In new product page, at Attribute Set select, choose Top
4. Scroll down to tab Grouped Products and try add new product.
5. Got an issue [https://imgur.com/porq478](https://imgur.com/porq478)
6. Loom: https://www.loom.com/share/b28e534af9254eec99bf5fc4e77a0376
### Expected result (*)
1. [https://imgur.com/iIjvWuq](https://imgur.com/iIjvWuq)
### Actual result (*)
1. [https://imgur.com/porq478](https://imgur.com/porq478)
---
Please provide [Severity](https://devdocs.magento.com/guides/v2.3/contributor-guide/contributing.html#backlog) assessment for the Issue as Reporter. This information will help during Confirmation and Issue triage processes.
- [ ] Severity: **S0** _- Affects critical data or functionality and leaves users without workaround._
- [ ] Severity: **S1** _- Affects critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S2** _- Affects non-critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S3** _- Affects non-critical data or functionality and does not force users to employ a workaround._
- [ ] Severity: **S4** _- Affects aesthetics, professional look and feel, “quality” or “usability”.
|
non_defect
|
redundant header generated while add a product to grouped product preconditions x steps to reproduce go admin catalog products select grouped product from drop down of add product button in new product page at attribute set select choose top scroll down to tab grouped products and try add new product got an issue loom expected result actual result please provide assessment for the issue as reporter this information will help during confirmation and issue triage processes severity affects critical data or functionality and leaves users without workaround severity affects critical data or functionality and forces users to employ a workaround severity affects non critical data or functionality and forces users to employ a workaround severity affects non critical data or functionality and does not force users to employ a workaround severity affects aesthetics professional look and feel “quality” or “usability”
| 0
|
25,834
| 4,469,203,545
|
IssuesEvent
|
2016-08-25 12:16:58
|
contao/core
|
https://api.github.com/repos/contao/core
|
closed
|
Hyperlink empty anchor hash
|
defect feedback required
|
If you input `{{link_url::*}}#` as url in a hyperlink (may be the same for every url input field), the `#` is stripped and the resulting url is `{{link_url::*}}`.
But `{{link_url::*}}#` and `{{link_url::*}}` are two completely different URLs for the browser.
`{{link_url::*}}#` will do a jump to page top.
`{{link_url::*}}` will reload the current page.
`{{link_url::*}}#` is a valid url, I don't see any reason why the empty `#` should be stripped.
Workaround: use `{{link_url::*}}#top` instead of `{{link_url::*}}#`.
|
1.0
|
Hyperlink empty anchor hash - If you input `{{link_url::*}}#` as url in a hyperlink (may be the same for every url input field), the `#` is stripped and the resulting url is `{{link_url::*}}`.
But `{{link_url::*}}#` and `{{link_url::*}}` are two completely different URLs for the browser.
`{{link_url::*}}#` will do a jump to page top.
`{{link_url::*}}` will reload the current page.
`{{link_url::*}}#` is a valid url, I don't see any reason why the empty `#` should be stripped.
Workaround: use `{{link_url::*}}#top` instead of `{{link_url::*}}#`.
|
defect
|
hyperlink empty anchor hash if you input link url as url in a hyperlink may be the same for every url input field the is stripped and the resulting url is link url but link url and link url are two completely different urls for the browser link url will do a jump to page top link url will reload the current page link url is a valid url i don t see any reason why the empty should be stripped workaround use link url top instead of link url
| 1
|
697,682
| 23,948,867,871
|
IssuesEvent
|
2022-09-12 09:44:24
|
wp-media/wp-rocket
|
https://api.github.com/repos/wp-media/wp-rocket
|
closed
|
Guard against exceeding in-progress items in the cache table
|
type: enhancement module: preload priority: medium
|
**Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version
- Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
When the batch size is bigger than what server is processing, we're stashing the `in-progress` entries
**To Reproduce**
Steps to reproduce the behavior:
1. Open the database manager
2. Run the preload on the bigger website
3. Observe `in-progress` entries for some times
4. More and more is being added
**Expected behavior**
The maximum of `in-progress` items should be exactly as the batch size defined in `rocket_preload_cache_pending_jobs_cron_rows_count`
**Additional context**
It's not causing CPU problems, but we applied that kind of guard in RUCSS.
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
|
1.0
|
Guard against exceeding in-progress items in the cache table - **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version
- Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
When the batch size is bigger than what server is processing, we're stashing the `in-progress` entries
**To Reproduce**
Steps to reproduce the behavior:
1. Open the database manager
2. Run the preload on the bigger website
3. Observe `in-progress` entries for some times
4. More and more is being added
**Expected behavior**
The maximum of `in-progress` items should be exactly as the batch size defined in `rocket_preload_cache_pending_jobs_cron_rows_count`
**Additional context**
It's not causing CPU problems, but we applied that kind of guard in RUCSS.
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
|
non_defect
|
guard against exceeding in progress items in the cache table before submitting an issue please check that you’ve completed the following steps made sure you’re on the latest version used the search feature to ensure that the bug hasn’t been reported before describe the bug when the batch size is bigger than what server is processing we re stashing the in progress entries to reproduce steps to reproduce the behavior open the database manager run the preload on the bigger website observe in progress entries for some times more and more is being added expected behavior the maximum of in progress items should be exactly as the batch size defined in rocket preload cache pending jobs cron rows count additional context it s not causing cpu problems but we applied that kind of guard in rucss backlog grooming for wp media dev team use only reproduce the problem identify the root cause scope a solution estimate the effort
| 0
|
59,352
| 17,023,095,542
|
IssuesEvent
|
2021-07-03 00:21:39
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
javascript slippy map continues to download images after zooming
|
Component: website Priority: major Resolution: wontfix Type: defect
|
**[Submitted to the original trac issue database at 9.23am, Tuesday, 17th January 2006]**
one solution: set up an image queue
another: maintain same map object during zoom. and same set of visible images. and simply change source on those images, and swap them around, as needed.
|
1.0
|
javascript slippy map continues to download images after zooming - **[Submitted to the original trac issue database at 9.23am, Tuesday, 17th January 2006]**
one solution: set up an image queue
another: maintain same map object during zoom. and same set of visible images. and simply change source on those images, and swap them around, as needed.
|
defect
|
javascript slippy map continues to download images after zooming one solution set up an image queue another maintain same map object during zoom and same set of visible images and simply change source on those images and swap them around as needed
| 1
|
63,902
| 18,051,864,692
|
IssuesEvent
|
2021-09-19 21:58:01
|
kewisch/gdata-provider
|
https://api.github.com/repos/kewisch/gdata-provider
|
closed
|
Event dialog asks to save a new event with no modifications (Bugzilla Bug 1568096)
|
defect
|
# [Bugzilla Bug 1568096](https://bugzilla.mozilla.org/show_bug.cgi?id=1568096)
Date: 2019-07-23T04:05:52+02:00
From: @darktrojan
Assigned To: nobody
Last updated: 2019-07-23T04:05:52+02:00
|
1.0
|
Event dialog asks to save a new event with no modifications (Bugzilla Bug 1568096) - # [Bugzilla Bug 1568096](https://bugzilla.mozilla.org/show_bug.cgi?id=1568096)
Date: 2019-07-23T04:05:52+02:00
From: @darktrojan
Assigned To: nobody
Last updated: 2019-07-23T04:05:52+02:00
|
defect
|
event dialog asks to save a new event with no modifications bugzilla bug date from darktrojan assigned to nobody last updated
| 1
|
75,124
| 25,542,025,694
|
IssuesEvent
|
2022-11-29 15:56:55
|
martinrotter/rssguard
|
https://api.github.com/repos/martinrotter/rssguard
|
opened
|
[BUG]: Issue with duplicate feeds
|
Type-Defect
|
### Brief description of the issue
Experiencing duplicate feeds.
### How to reproduce the bug?
1. add filter.
`function filterMessage() {
if (msg.isAlreadyInDatabase(MessageObject.SameTitle)) {
return MessageObject.Ignore;
}
else {
return MessageObject.Accept;
}
}`
Add news feeds from similar sources. Ex. CTV.
### What was the expected result?
Feeds without duplicated titles.
### What actually happened?
Got duplicate feeds.
### Debug log
NA
### Operating system and version
* OS: KDE Neon
* RSS Guard version:
|
1.0
|
[BUG]: Issue with duplicate feeds - ### Brief description of the issue
Experiencing duplicate feeds.
### How to reproduce the bug?
1. add filter.
`function filterMessage() {
if (msg.isAlreadyInDatabase(MessageObject.SameTitle)) {
return MessageObject.Ignore;
}
else {
return MessageObject.Accept;
}
}`
Add news feeds from similar sources. Ex. CTV.
### What was the expected result?
Feeds without duplicated titles.
### What actually happened?
Got duplicate feeds.
### Debug log
NA
### Operating system and version
* OS: KDE Neon
* RSS Guard version:
|
defect
|
issue with duplicate feeds brief description of the issue experiencing duplicate feeds how to reproduce the bug add filter function filtermessage if msg isalreadyindatabase messageobject sametitle return messageobject ignore else return messageobject accept add news feeds from similar sources ex ctv what was the expected result feeds without duplicated titles what actually happened got duplicate feeds debug log na operating system and version os kde neon rss guard version
| 1
|
47,922
| 13,066,374,536
|
IssuesEvent
|
2020-07-30 21:33:45
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
[PROPOSAL] should not build tables in I3_SRC (Trac #1430)
|
Migrated from Trac combo simulation defect
|
One of the targets for PROPOSAL to build tables to is `$I3_BUILD/PROPOSAL/resources/tables`. This is a symlink into `$I3_SRC` almost all the time. Two problems with this:
1. Philosophically, if I specify a separate build directory that means I don't want you touching the source.
2. The source could potentially be read-only.
Solution: write somewhere else in the build directory.
Migrated from https://code.icecube.wisc.edu/ticket/1430
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:35",
"description": "One of the targets for PROPOSAL to build tables to is `$I3_BUILD/PROPOSAL/resources/tables`. This is a symlink into `$I3_SRC` almost all the time. Two problems with this:\n\n1. Philosophically, if I specify a separate build directory that means I don't want you touching the source.\n\n2. The source could potentially be read-only.\n\nSolution: write somewhere else in the build directory.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1550067215093672",
"component": "combo simulation",
"summary": "[PROPOSAL] should not build tables in I3_SRC",
"priority": "major",
"keywords": "",
"time": "2015-11-10T16:28:21",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
|
1.0
|
[PROPOSAL] should not build tables in I3_SRC (Trac #1430) - One of the targets for PROPOSAL to build tables to is `$I3_BUILD/PROPOSAL/resources/tables`. This is a symlink into `$I3_SRC` almost all the time. Two problems with this:
1. Philosophically, if I specify a separate build directory that means I don't want you touching the source.
2. The source could potentially be read-only.
Solution: write somewhere else in the build directory.
Migrated from https://code.icecube.wisc.edu/ticket/1430
```json
{
"status": "closed",
"changetime": "2019-02-13T14:13:35",
"description": "One of the targets for PROPOSAL to build tables to is `$I3_BUILD/PROPOSAL/resources/tables`. This is a symlink into `$I3_SRC` almost all the time. Two problems with this:\n\n1. Philosophically, if I specify a separate build directory that means I don't want you touching the source.\n\n2. The source could potentially be read-only.\n\nSolution: write somewhere else in the build directory.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1550067215093672",
"component": "combo simulation",
"summary": "[PROPOSAL] should not build tables in I3_SRC",
"priority": "major",
"keywords": "",
"time": "2015-11-10T16:28:21",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
|
defect
|
should not build tables in src trac one of the targets for proposal to build tables to is build proposal resources tables this is a symlink into src almost all the time two problems with this philosophically if i specify a separate build directory that means i don t want you touching the source the source could potentially be read only solution write somewhere else in the build directory migrated from json status closed changetime description one of the targets for proposal to build tables to is build proposal resources tables this is a symlink into src almost all the time two problems with this n philosophically if i specify a separate build directory that means i don t want you touching the source n the source could potentially be read only n nsolution write somewhere else in the build directory reporter david schultz cc resolution fixed ts component combo simulation summary should not build tables in src priority major keywords time milestone owner olivas type defect
| 1
|
266,588
| 23,247,443,263
|
IssuesEvent
|
2022-08-03 21:50:11
|
astropy/astropy
|
https://api.github.com/repos/astropy/astropy
|
closed
|
TST: test_day_frac_exact in astropy/time/tests/test_precision.py fails now (hypothesis 6.46.9)
|
testing time Bug
|
- [ ] Undo #13277 and backport that to 5.0.x and 5.1.x
I reran a previously passing build on `main` branch (passed 2 days ago) today and it fails with this error:
```
_____________________________ test_day_frac_exact ______________________________
@given(integers(-2**52+2, 2**52-2), floats(-0.5, 0.5))
> @example(i=65536, f=3.637978807091714e-12)
../../astropy/time/tests/test_precision.py:341:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
i = 1, f = 0.49999999999999994
@given(integers(-2**52+2, 2**52-2), floats(-0.5, 0.5))
@example(i=65536, f=3.637978807091714e-12)
def test_day_frac_exact(i, f):
assume(abs(f) < 0.5 or i % 2 == 0)
i_d, f_d = day_frac(i, f)
> assert i == i_d
E assert 1 == 2.0
../../astropy/time/tests/test_precision.py:345: AssertionError
---------------------------------- Hypothesis ----------------------------------
Falsifying example: test_day_frac_exact(
i=1, f=0.49999999999999994,
)
You can reproduce this example by temporarily adding @reproduce_failure('6.46.9', b'AAABAAABAAAAwB////////8=') as a decorator on your test case
```
This blocks CI in unrelated PRs.
@mhvk , do you know what is going on there? Thanks! 🙏
|
1.0
|
TST: test_day_frac_exact in astropy/time/tests/test_precision.py fails now (hypothesis 6.46.9) - - [ ] Undo #13277 and backport that to 5.0.x and 5.1.x
I reran a previously passing build on `main` branch (passed 2 days ago) today and it fails with this error:
```
_____________________________ test_day_frac_exact ______________________________
@given(integers(-2**52+2, 2**52-2), floats(-0.5, 0.5))
> @example(i=65536, f=3.637978807091714e-12)
../../astropy/time/tests/test_precision.py:341:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
i = 1, f = 0.49999999999999994
@given(integers(-2**52+2, 2**52-2), floats(-0.5, 0.5))
@example(i=65536, f=3.637978807091714e-12)
def test_day_frac_exact(i, f):
assume(abs(f) < 0.5 or i % 2 == 0)
i_d, f_d = day_frac(i, f)
> assert i == i_d
E assert 1 == 2.0
../../astropy/time/tests/test_precision.py:345: AssertionError
---------------------------------- Hypothesis ----------------------------------
Falsifying example: test_day_frac_exact(
i=1, f=0.49999999999999994,
)
You can reproduce this example by temporarily adding @reproduce_failure('6.46.9', b'AAABAAABAAAAwB////////8=') as a decorator on your test case
```
This blocks CI in unrelated PRs.
@mhvk , do you know what is going on there? Thanks! 🙏
|
non_defect
|
tst test day frac exact in astropy time tests test precision py fails now hypothesis undo and backport that to x and x i reran a previously passing build on main branch passed days ago today and it fails with this error test day frac exact given integers floats example i f astropy time tests test precision py i f given integers floats example i f def test day frac exact i f assume abs f or i i d f d day frac i f assert i i d e assert astropy time tests test precision py assertionerror hypothesis falsifying example test day frac exact i f you can reproduce this example by temporarily adding reproduce failure b aaabaaabaaaawb as a decorator on your test case this blocks ci in unrelated prs mhvk do you know what is going on there thanks 🙏
| 0
|
28,155
| 6,955,456,033
|
IssuesEvent
|
2017-12-07 07:44:07
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
[4.0] Show custom fields
|
No Code Attached Yet
|
### Steps to reproduce the issue
In com_contact options
<img width="531" alt="screenshotr12-21-49" src="https://user-images.githubusercontent.com/1296369/28498926-92be56fe-6fa1-11e7-8b19-db97f7b973ec.png">
As seen in the image above there is a warning message being displayed in this field
`
Warning</b>: htmlspecialchars() expects parameter 1 to be string, array given in <b>C:\Apache24\htdocs\cms4\layouts\joomla\form\field\text.php</b> on line <b>94</b><br />`
|
1.0
|
[4.0] Show custom fields - ### Steps to reproduce the issue
In com_contact options
<img width="531" alt="screenshotr12-21-49" src="https://user-images.githubusercontent.com/1296369/28498926-92be56fe-6fa1-11e7-8b19-db97f7b973ec.png">
As seen in the image above there is a warning message being displayed in this field
`
Warning</b>: htmlspecialchars() expects parameter 1 to be string, array given in <b>C:\Apache24\htdocs\cms4\layouts\joomla\form\field\text.php</b> on line <b>94</b><br />`
|
non_defect
|
show custom fields steps to reproduce the issue in com contact options img width alt src as seen in the image above there is a warning message being displayed in this field warning htmlspecialchars expects parameter to be string array given in c htdocs layouts joomla form field text php on line
| 0
|
20,203
| 3,315,072,885
|
IssuesEvent
|
2015-11-06 09:58:46
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
dart2js doesn't allow const null in string interpolation
|
Area-Dart2JS Type-Defect
|
I happened to notice that string interpolation with a const null variable results in a const string in dart but not dart2js. (According to the spec, dart's behavior is correct.) Here's the code:
```
const aConstNull = null; // OK with dart but not with dart2js.
const aConstNum = 0;
const aConstBool = true;
const aConstString = 'a constant string';
...
const validConstString = '$aConstNull $aConstNum $aConstBool $aConstString';
```
Here's the error I see when I try to run it in Chrome, using WebStorm (which uses `pub serve`):
```
[Info from Dart2JS]:
Compiling const_test|web/main.dart...
[Error from Dart2JS]:
web/main.dart:20:30:
Not a compile-time constant.
const validConstString = '$aConstNull $aConstNum $aConstBool $aConstString';
^^^^^^^^^^
[Error from Dart2JS]:
web/main.dart:20:30:
Not a compile-time constant.
const validConstString = '$aConstNull $aConstNum $aConstBool $aConstString';
^^^^^^^^^^
```
I noticed this because I happened to try [the code in DartPad](https://dartpad.dartlang.org/e9c0a860fc21e4adee7f).
|
1.0
|
dart2js doesn't allow const null in string interpolation - I happened to notice that string interpolation with a const null variable results in a const string in dart but not dart2js. (According to the spec, dart's behavior is correct.) Here's the code:
```
const aConstNull = null; // OK with dart but not with dart2js.
const aConstNum = 0;
const aConstBool = true;
const aConstString = 'a constant string';
...
const validConstString = '$aConstNull $aConstNum $aConstBool $aConstString';
```
Here's the error I see when I try to run it in Chrome, using WebStorm (which uses `pub serve`):
```
[Info from Dart2JS]:
Compiling const_test|web/main.dart...
[Error from Dart2JS]:
web/main.dart:20:30:
Not a compile-time constant.
const validConstString = '$aConstNull $aConstNum $aConstBool $aConstString';
^^^^^^^^^^
[Error from Dart2JS]:
web/main.dart:20:30:
Not a compile-time constant.
const validConstString = '$aConstNull $aConstNum $aConstBool $aConstString';
^^^^^^^^^^
```
I noticed this because I happened to try [the code in DartPad](https://dartpad.dartlang.org/e9c0a860fc21e4adee7f).
|
defect
|
doesn t allow const null in string interpolation i happened to notice that string interpolation with a const null variable results in a const string in dart but not according to the spec dart s behavior is correct here s the code const aconstnull null ok with dart but not with const aconstnum const aconstbool true const aconststring a constant string const validconststring aconstnull aconstnum aconstbool aconststring here s the error i see when i try to run it in chrome using webstorm which uses pub serve compiling const test web main dart web main dart not a compile time constant const validconststring aconstnull aconstnum aconstbool aconststring web main dart not a compile time constant const validconststring aconstnull aconstnum aconstbool aconststring i noticed this because i happened to try
| 1
|
48,242
| 5,950,265,989
|
IssuesEvent
|
2017-05-26 16:15:25
|
MyersResearchGroup/iBioSim
|
https://api.github.com/repos/MyersResearchGroup/iBioSim
|
closed
|
CHANGE: remove unused array namespace on export
|
CHANGE Needs Testing
|
If there are no arrays, remove arrays namespace.
|
1.0
|
CHANGE: remove unused array namespace on export - If there are no arrays, remove arrays namespace.
|
non_defect
|
change remove unused array namespace on export if there are no arrays remove arrays namespace
| 0
|
81,217
| 30,755,651,428
|
IssuesEvent
|
2023-07-29 02:45:38
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
4.x: Division by zero in ProgressHelper
|
defect console
|
### Description
Bake a command via `bin/cake bake Test`
Just like the docs say in https://book.cakephp.org/4/en/console-commands/input-output.html#command-helpers add
```
$progress = $io->helper('Progress');
$progress->increment(10);
$progress->draw();
```
to the execute method and execute the command bin `bin/cake test`
This will result in the following error:
```
2023-07-27 09:51:38 error: [DivisionByZeroError] Division by zero in /myapp/vendor/cakephp/cakephp/src/Shell/Helper/ProgressHelper.php on line 134
Stack Trace:
- /myapp/src/Command/TestCommand.php:25
- /myapp/vendor/cakephp/cakephp/src/Console/BaseCommand.php:190
- /myapp/vendor/cakephp/cakephp/src/Console/CommandRunner.php:334
- /myapp/vendor/cakephp/cakephp/src/Console/CommandRunner.php:172
- /myapp/alfred/bin/cake.php:12
- [main]:
[DivisionByZeroError] Division by zero in /myapp/vendor/cakephp/cakephp/src/Shell/Helper/ProgressHelper.php on line 134
```
Should this be adjusted in the docs or is this actually a core issue?
### CakePHP Version
4.4.15
### PHP Version
8.2
|
1.0
|
4.x: Division by zero in ProgressHelper - ### Description
Bake a command via `bin/cake bake Test`
Just like the docs say in https://book.cakephp.org/4/en/console-commands/input-output.html#command-helpers add
```
$progress = $io->helper('Progress');
$progress->increment(10);
$progress->draw();
```
to the execute method and execute the command bin `bin/cake test`
This will result in the following error:
```
2023-07-27 09:51:38 error: [DivisionByZeroError] Division by zero in /myapp/vendor/cakephp/cakephp/src/Shell/Helper/ProgressHelper.php on line 134
Stack Trace:
- /myapp/src/Command/TestCommand.php:25
- /myapp/vendor/cakephp/cakephp/src/Console/BaseCommand.php:190
- /myapp/vendor/cakephp/cakephp/src/Console/CommandRunner.php:334
- /myapp/vendor/cakephp/cakephp/src/Console/CommandRunner.php:172
- /myapp/alfred/bin/cake.php:12
- [main]:
[DivisionByZeroError] Division by zero in /myapp/vendor/cakephp/cakephp/src/Shell/Helper/ProgressHelper.php on line 134
```
Should this be adjusted in the docs or is this actually a core issue?
### CakePHP Version
4.4.15
### PHP Version
8.2
|
defect
|
x division by zero in progresshelper description bake a command via bin cake bake test just like the docs say in add progress io helper progress progress increment progress draw to the execute method and execute the command bin bin cake test this will result in the following error error division by zero in myapp vendor cakephp cakephp src shell helper progresshelper php on line stack trace myapp src command testcommand php myapp vendor cakephp cakephp src console basecommand php myapp vendor cakephp cakephp src console commandrunner php myapp vendor cakephp cakephp src console commandrunner php myapp alfred bin cake php division by zero in myapp vendor cakephp cakephp src shell helper progresshelper php on line should this be adjusted in the docs or is this actually a core issue cakephp version php version
| 1
|
22,467
| 3,652,766,866
|
IssuesEvent
|
2016-02-17 02:25:12
|
melt-umn/silver
|
https://api.github.com/repos/melt-umn/silver
|
closed
|
Silver Eclipse should ship with additional library grammars
|
auto-migrated Priority-Medium Type-Defect
|
```
We currently only ship core, ide, lib from the silver:grammars space. We
should also include silver:langutil as it has features that other languages
besides Silver use.
Ming, please indicate here where the code is that does this. I will try to
change it myself. I want to know where this code is in case this changes in
the future anyway.
Thanks,
Eric
```
Original issue reported on code.google.com by `e...@umn.edu` on 31 Jul 2013 at 7:03
|
1.0
|
Silver Eclipse should ship with additional library grammars - ```
We currently only ship core, ide, lib from the silver:grammars space. We
should also include silver:langutil as it has features that other languages
besides Silver use.
Ming, please indicate here where the code is that does this. I will try to
change it myself. I want to know where this code is in case this changes in
the future anyway.
Thanks,
Eric
```
Original issue reported on code.google.com by `e...@umn.edu` on 31 Jul 2013 at 7:03
|
defect
|
silver eclipse should ship with additional library grammars we currently only ship core ide lib from the silver grammars space we should also include silver langutil as it has features that other languages besides silver use ming please indicate here where the code is that does this i will try to change it myself i want to know where this code is in case this changes in the future anyway thanks eric original issue reported on code google com by e umn edu on jul at
| 1
|
536,577
| 15,710,936,650
|
IssuesEvent
|
2021-03-27 04:27:12
|
open-wa/wa-automate-nodejs
|
https://api.github.com/repos/open-wa/wa-automate-nodejs
|
closed
|
small enhancements
|
PRIORITY
|
launch logs:
- [x] `navigating to web page`
- [x] `setting up the browser`
- [x] `starting: version VERSION_NUMBER`
- [x] `found existing session data`
- [x] `injecting session data`
- [x] allow user to choose to see session info as object instead of console.table
- [x] change `debugInfo` to `sessionInfo` in `initializer`
perf:
- [x] set a timeout for patch requests
error:
- [ ] throw error `Cannot send messages to this chat`
linting:
- [x] fix some linting issues on `Client.ts`
|
1.0
|
small enhancements - launch logs:
- [x] `navigating to web page`
- [x] `setting up the browser`
- [x] `starting: version VERSION_NUMBER`
- [x] `found existing session data`
- [x] `injecting session data`
- [x] allow user to choose to see session info as object instead of console.table
- [x] change `debugInfo` to `sessionInfo` in `initializer`
perf:
- [x] set a timeout for patch requests
error:
- [ ] throw error `Cannot send messages to this chat`
linting:
- [x] fix some linting issues on `Client.ts`
|
non_defect
|
small enhancements launch logs navigating to web page setting up the browser starting version version number found existing session data injecting session data allow user to choose to see session info as object instead of console table change debuginfo to sessioninfo in initializer perf set a timeout for patch requests error throw error cannot send messages to this chat linting fix some linting issues on client ts
| 0
|
19,652
| 3,228,755,420
|
IssuesEvent
|
2015-10-12 06:43:19
|
contao/core
|
https://api.github.com/repos/contao/core
|
closed
|
ModuleWizard does not execute save_callback
|
defect
|
The ModuleWizard has `$blnSubmitInput` set to `false`, because it internally stores the values in the database.
Unfortunately, this does not execute `save_callback` and there's no way to check/manipulate the values.
|
1.0
|
ModuleWizard does not execute save_callback - The ModuleWizard has `$blnSubmitInput` set to `false`, because it internally stores the values in the database.
Unfortunately, this does not execute `save_callback` and there's no way to check/manipulate the values.
|
defect
|
modulewizard does not execute save callback the modulewizard has blnsubmitinput set to false because it internally stores the values in the database unfortunately this does not execute save callback and there s no way to check manipulate the values
| 1
|
40,512
| 12,799,553,140
|
IssuesEvent
|
2020-07-02 15:34:24
|
TreyM-WSS/WhiteSource-Demo
|
https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo
|
opened
|
CVE-2016-6816 (High) detected in tomcat-embed-core-8.5.4.jar
|
security vulnerability
|
## CVE-2016-6816 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.4.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p>
<p>Path to dependency file: /tmp/ws-scm/WhiteSource-Demo/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.4/tomcat-embed-core-8.5.4.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-1.4.0.RELEASE.jar
- :x: **tomcat-embed-core-8.5.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo/commits/75659f691fb82d67ecd666ba6076394defeb92d0">75659f691fb82d67ecd666ba6076394defeb92d0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The code in Apache Tomcat 9.0.0.M1 to 9.0.0.M11, 8.5.0 to 8.5.6, 8.0.0.RC1 to 8.0.38, 7.0.0 to 7.0.72, and 6.0.0 to 6.0.47 that parsed the HTTP request line permitted invalid characters. This could be exploited, in conjunction with a proxy that also permitted the invalid characters but with a different interpretation, to inject data into the HTTP response. By manipulating the HTTP response the attacker could poison a web-cache, perform an XSS attack and/or obtain sensitive information from requests other then their own.
<p>Publish Date: 2017-03-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-6816>CVE-2016-6816</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-6816">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-6816</a></p>
<p>Release Date: 2017-03-20</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M13,8.5.8,8.0.39,7.0.73,org.apache.tomcat:tomcat-catalina:9.0.0.M13,8.5.8,8.0.39,7.0.73</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.4","isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.4.0.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:1.4.0.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:8.5.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M13,8.5.8,8.0.39,7.0.73,org.apache.tomcat:tomcat-catalina:9.0.0.M13,8.5.8,8.0.39,7.0.73"}],"vulnerabilityIdentifier":"CVE-2016-6816","vulnerabilityDetails":"The code in Apache Tomcat 9.0.0.M1 to 9.0.0.M11, 8.5.0 to 8.5.6, 8.0.0.RC1 to 8.0.38, 7.0.0 to 7.0.72, and 6.0.0 to 6.0.47 that parsed the HTTP request line permitted invalid characters. This could be exploited, in conjunction with a proxy that also permitted the invalid characters but with a different interpretation, to inject data into the HTTP response. By manipulating the HTTP response the attacker could poison a web-cache, perform an XSS attack and/or obtain sensitive information from requests other then their own.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-6816","cvss3Severity":"high","cvss3Score":"7.1","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2016-6816 (High) detected in tomcat-embed-core-8.5.4.jar - ## CVE-2016-6816 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.4.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p>
<p>Path to dependency file: /tmp/ws-scm/WhiteSource-Demo/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.4/tomcat-embed-core-8.5.4.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-1.4.0.RELEASE.jar
- :x: **tomcat-embed-core-8.5.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo/commits/75659f691fb82d67ecd666ba6076394defeb92d0">75659f691fb82d67ecd666ba6076394defeb92d0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The code in Apache Tomcat 9.0.0.M1 to 9.0.0.M11, 8.5.0 to 8.5.6, 8.0.0.RC1 to 8.0.38, 7.0.0 to 7.0.72, and 6.0.0 to 6.0.47 that parsed the HTTP request line permitted invalid characters. This could be exploited, in conjunction with a proxy that also permitted the invalid characters but with a different interpretation, to inject data into the HTTP response. By manipulating the HTTP response the attacker could poison a web-cache, perform an XSS attack and/or obtain sensitive information from requests other then their own.
<p>Publish Date: 2017-03-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-6816>CVE-2016-6816</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-6816">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-6816</a></p>
<p>Release Date: 2017-03-20</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M13,8.5.8,8.0.39,7.0.73,org.apache.tomcat:tomcat-catalina:9.0.0.M13,8.5.8,8.0.39,7.0.73</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.4","isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.4.0.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:1.4.0.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:8.5.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M13,8.5.8,8.0.39,7.0.73,org.apache.tomcat:tomcat-catalina:9.0.0.M13,8.5.8,8.0.39,7.0.73"}],"vulnerabilityIdentifier":"CVE-2016-6816","vulnerabilityDetails":"The code in Apache Tomcat 9.0.0.M1 to 9.0.0.M11, 8.5.0 to 8.5.6, 8.0.0.RC1 to 8.0.38, 7.0.0 to 7.0.72, and 6.0.0 to 6.0.47 that parsed the HTTP request line permitted invalid characters. This could be exploited, in conjunction with a proxy that also permitted the invalid characters but with a different interpretation, to inject data into the HTTP response. By manipulating the HTTP response the attacker could poison a web-cache, perform an XSS attack and/or obtain sensitive information from requests other then their own.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-6816","cvss3Severity":"high","cvss3Score":"7.1","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file tmp ws scm whitesource demo pom xml path to vulnerable library home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library found in head commit a href vulnerability details the code in apache tomcat to to to to and to that parsed the http request line permitted invalid characters this could be exploited in conjunction with a proxy that also permitted the invalid characters but with a different interpretation to inject data into the http response by manipulating the http response the attacker could poison a web cache perform an xss attack and or obtain sensitive information from requests other then their own publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails the code in apache tomcat to to to to and to that parsed the http request line permitted invalid characters this could be exploited in conjunction with a proxy that also permitted the invalid characters but with a different interpretation to inject data into the http response by manipulating the http response the attacker could poison a web cache perform an xss attack and or obtain sensitive information from requests other then their own vulnerabilityurl
| 0
|
56,196
| 14,973,918,291
|
IssuesEvent
|
2021-01-28 02:15:46
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
closed
|
Unable to 'unallow' a user that has been deleted
|
Type: Defect
|
### System information
Type | Version/Name
--- | ---
Distribution Name | Debian
Distribution Version | 10
Linux Kernel | 4.19.0-12
Architecture | amd64
ZFS Version | 0.8.4-2~bpo10+1
SPL Version | 0.8.4-2~bpo10+1
### Describe the problem you're observing
There doesn't appear to be a way to 'unallow' a user that no longer exists on the system.
### Describe how to reproduce the problem
Add a test user to the system, create a test folder, and grant that user create permissions on the test folder:
```
root@uswuxsdbkp00:~# useradd test
root@uswuxsdbkp00:~# zfs create tank/myfolder
root@uswuxsdbkp00:~# zfs allow -u test create tank/myfolder
root@uswuxsdbkp00:~# zfs allow tank/myfolder
---- Permissions on tank/myfolder ------------------------------------
Local+Descendent permissions:
user test create
root@uswuxsdbkp00:~#
```
Unallow that user to show permissions can be removed:
```
root@uswuxsdbkp00:~# zfs unallow -u test tank/myfolder
root@uswuxsdbkp00:~# zfs allow tank/myfolder
root@uswuxsdbkp00:~#
```
Re-add the create permission and verify it exists:
```
root@uswuxsdbkp00:~# zfs allow -u test create tank/myfolder
root@uswuxsdbkp00:~# zfs allow tank/myfolder
---- Permissions on tank/myfolder ------------------------------------
Local+Descendent permissions:
user test create
root@uswuxsdbkp00:~#
```
Remove the test user from the system and look at the permissions again:
```
root@uswuxsdbkp00:~# userdel test
root@uswuxsdbkp00:~# zfs allow tank/myfolder
---- Permissions on tank/myfolder ------------------------------------
Local+Descendent permissions:
user create
root@uswuxsdbkp00:~#
```
I think there are a few problems here:
* ZFS doesn't display the UID/GID when it can't resolve the name
* ZFS doesn't appear to have a flag to force returning the UID/GID instead of the name
* Maybe the docs are confusing, but there doesn't appear to be a way to say "remove all permissions from this dataset" (```zfs unallow -e tank/myfolder``` doesn't do it)
|
1.0
|
Unable to 'unallow' a user that has been deleted - ### System information
Type | Version/Name
--- | ---
Distribution Name | Debian
Distribution Version | 10
Linux Kernel | 4.19.0-12
Architecture | amd64
ZFS Version | 0.8.4-2~bpo10+1
SPL Version | 0.8.4-2~bpo10+1
### Describe the problem you're observing
There doesn't appear to be a way to 'unallow' a user that no longer exists on the system.
### Describe how to reproduce the problem
Add a test user to the system, create a test folder, and grant that user create permissions on the test folder:
```
root@uswuxsdbkp00:~# useradd test
root@uswuxsdbkp00:~# zfs create tank/myfolder
root@uswuxsdbkp00:~# zfs allow -u test create tank/myfolder
root@uswuxsdbkp00:~# zfs allow tank/myfolder
---- Permissions on tank/myfolder ------------------------------------
Local+Descendent permissions:
user test create
root@uswuxsdbkp00:~#
```
Unallow that user to show permissions can be removed:
```
root@uswuxsdbkp00:~# zfs unallow -u test tank/myfolder
root@uswuxsdbkp00:~# zfs allow tank/myfolder
root@uswuxsdbkp00:~#
```
Re-add the create permission and verify it exists:
```
root@uswuxsdbkp00:~# zfs allow -u test create tank/myfolder
root@uswuxsdbkp00:~# zfs allow tank/myfolder
---- Permissions on tank/myfolder ------------------------------------
Local+Descendent permissions:
user test create
root@uswuxsdbkp00:~#
```
Remove the test user from the system and look at the permissions again:
```
root@uswuxsdbkp00:~# userdel test
root@uswuxsdbkp00:~# zfs allow tank/myfolder
---- Permissions on tank/myfolder ------------------------------------
Local+Descendent permissions:
user create
root@uswuxsdbkp00:~#
```
I think there are a few problems here:
* ZFS doesn't display the UID/GID when it can't resolve the name
* ZFS doesn't appear to have a flag to force returning the UID/GID instead of the name
* Maybe the docs are confusing, but there doesn't appear to be a way to say "remove all permissions from this dataset" (```zfs unallow -e tank/myfolder``` doesn't do it)
|
defect
|
unable to unallow a user that has been deleted system information type version name distribution name debian distribution version linux kernel architecture zfs version spl version describe the problem you re observing there doesn t appear to be a way to unallow a user that no longer exists on the system describe how to reproduce the problem add a test user to the system create a test folder and grant that user create permissions on the test folder root useradd test root zfs create tank myfolder root zfs allow u test create tank myfolder root zfs allow tank myfolder permissions on tank myfolder local descendent permissions user test create root unallow that user to show permissions can be removed root zfs unallow u test tank myfolder root zfs allow tank myfolder root re add the create permission and verify it exists root zfs allow u test create tank myfolder root zfs allow tank myfolder permissions on tank myfolder local descendent permissions user test create root remove the test user from the system and look at the permissions again root userdel test root zfs allow tank myfolder permissions on tank myfolder local descendent permissions user create root i think there are a few problems here zfs doesn t display the uid gid when it can t resolve the name zfs doesn t appear to have a flag to force returning the uid gid instead of the name maybe the docs are confusing but there doesn t appear to be a way to say remove all permissions from this dataset zfs unallow e tank myfolder doesn t do it
| 1
|
40,500
| 8,796,653,879
|
IssuesEvent
|
2018-12-23 10:08:14
|
happycube/ld-decode
|
https://api.github.com/repos/happycube/ld-decode
|
closed
|
Improvements to ld-analyse
|
enhancement ld-decode-tools
|
The ld-analyse application should make it clearer as to what metadata is present. For the VBI "not present" should be "no metadata". For highlight dropouts, the option should be unavailable if no dropout metadata is present. Same for the VBI and NTSC dialogue windows.
|
1.0
|
Improvements to ld-analyse - The ld-analyse application should make it clearer as to what metadata is present. For the VBI "not present" should be "no metadata". For highlight dropouts, the option should be unavailable if no dropout metadata is present. Same for the VBI and NTSC dialogue windows.
|
non_defect
|
improvements to ld analyse the ld analyse application should make it clearer as to what metadata is present for the vbi not present should be no metadata for highlight dropouts the option should be unavailable if no dropout metadata is present same for the vbi and ntsc dialogue windows
| 0
|
29,494
| 14,145,667,545
|
IssuesEvent
|
2020-11-10 18:03:42
|
firebase/firebase-ios-sdk
|
https://api.github.com/repos/firebase/firebase-ios-sdk
|
closed
|
Suspected Firebase SDK v6.34 crash from [APMMeasurement uploadData]
|
Carthage api: performance
|
<!-- DO NOT DELETE
validate_template=true
template_path=.github/ISSUE_TEMPLATE/bug_report.md
-->
### [REQUIRED] Step 1: Describe your environment
* Xcode version: 12.0
* Firebase SDK version: 6.34.0
* Installation method: Carthage
* Firebase Component: iOS SDK/Performance
### [REQUIRED] Step 2: Describe the problem
Since a recent app update, we are seeing a new number 1 crash in our Ecommerce app. This occurred in the first build since I'd refreshed the Firebase SDK (and other dependencies) to the latest versions. I am not certain it is caused by Firebase but from the stack trace it seems the most likely cause. I do not have steps to reproduce the issue.
My analysis is:
1. The crashed thread (number 19) was calling back on a KVO change from CFNetwork-ing
2. Looking in our crash tracker (not Crashlytics for commercial reasons) I can see one of the last network calls before the crash was to https://app-measurement.com/a
3. Reviewing other background threads, I noticed thread 14 was doing a SQL-Lite delete initiated from [APMMeasurement uploadData]
4. Googling the method found in point 3 reveals historic Firebase issues (nothing applicable to this crash)
.crash file can be made available by email if required.
The complete stack trace is here:
Incident Identifier: 06FD8CD6-304B-4DF3-8894-E27C258C8A5A
CrashReporter Key: 0000000000000000000000000000000000000000
Hardware Model: iPhone9,3
Code Type: arm64
Date/Time: 2020-11-09T08:35:49+00:00
Launch Time: 2020-11-09T08:35:49+00:00
OS Version: iOS 14.1 (18A8395)
Report Version: 105
Exception Type: SIGSEGV
Exception Subtype:
Thread 0 name:
Thread 0:
0 libsystem_kernel.dylib 0x00000001b9c948c4 mach_msg_trap (libsystem_kernel.dylib)
1 CoreFoundation 0x000000018f2c074c __CFRunLoopServiceMachPort (CoreFoundation)
2 CoreFoundation 0x000000018f2babd0 __CFRunLoopRun (CoreFoundation)
3 CoreFoundation 0x000000018f2ba200 CFRunLoopRunSpecific (CoreFoundation)
4 GraphicsServices 0x00000001a53b5598 GSEventRunModal (GraphicsServices)
5 UIKitCore 0x0000000191b80004 -[UIApplication _run] (UIKitCore)
6 UIKitCore 0x0000000191b855d8 UIApplicationMain (UIKitCore)
7 JDBuild 0x000000010022764c main (coreAppDelegate.swift:14)
8 libdyld.dylib 0x000000018ef99598 start (libdyld.dylib)
Thread 1 name:
Thread 1:
0 libsystem_kernel.dylib 0x00000001b9c948c4 mach_msg_trap (libsystem_kernel.dylib)
1 CoreFoundation 0x000000018f2c074c __CFRunLoopServiceMachPort (CoreFoundation)
2 CoreFoundation 0x000000018f2babd0 __CFRunLoopRun (CoreFoundation)
3 CoreFoundation 0x000000018f2ba200 CFRunLoopRunSpecific (CoreFoundation)
4 Foundation 0x00000001904c4278 -[NSRunLoop(NSRunLoop) runMode:beforeDate:] (Foundation)
5 Foundation 0x00000001904c4158 -[NSRunLoop(NSRunLoop) runUntilDate:] (Foundation)
6 UIKitCore 0x0000000191c2b9fc -[UIEventFetcher threadMain] (UIKitCore)
7 Foundation 0x0000000190620c48 __NSThread__start__ (Foundation)
8 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
9 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 2 name:
Thread 2:
0 libsystem_kernel.dylib 0x00000001b9cb772c __psynch_cvwait (libsystem_kernel.dylib)
1 libc++.1.dylib 0x00000001a2c03e0c std::__1::condition_variable::wait(std::__1::unique_lock<std::__1::mutex>&) (libc++.1.dylib)
2 NewRelic 0x0000000100d7d88c 0x100cf4000 + 563340
3 NewRelic 0x0000000100d7db04 0x100cf4000 + 563972
4 NewRelic 0x0000000100d7da1c 0x100cf4000 + 563740
5 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
6 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 3 name:
Thread 3:
0 libsystem_kernel.dylib 0x00000001b9cb772c __psynch_cvwait (libsystem_kernel.dylib)
1 libc++.1.dylib 0x00000001a2c03e0c std::__1::condition_variable::wait(std::__1::unique_lock<std::__1::mutex>&) (libc++.1.dylib)
2 NewRelic 0x0000000100d7d88c 0x100cf4000 + 563340
3 NewRelic 0x0000000100d7db04 0x100cf4000 + 563972
4 NewRelic 0x0000000100d7da1c 0x100cf4000 + 563740
5 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
6 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 4 name:
Thread 4:
0 libsystem_kernel.dylib 0x00000001b9c948c4 mach_msg_trap (libsystem_kernel.dylib)
1 CoreFoundation 0x000000018f2c074c __CFRunLoopServiceMachPort (CoreFoundation)
2 CoreFoundation 0x000000018f2babd0 __CFRunLoopRun (CoreFoundation)
3 CoreFoundation 0x000000018f2ba200 CFRunLoopRunSpecific (CoreFoundation)
4 CoreFoundation 0x000000018f2bb2c0 CFRunLoopRun (CoreFoundation)
5 CoreMotion 0x0000000199ab2374 (null) (CoreMotion)
6 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
7 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 5 name:
Thread 5:
0 libsystem_kernel.dylib 0x00000001b9c948c4 mach_msg_trap (libsystem_kernel.dylib)
1 CoreFoundation 0x000000018f2c074c __CFRunLoopServiceMachPort (CoreFoundation)
2 CoreFoundation 0x000000018f2babd0 __CFRunLoopRun (CoreFoundation)
3 CoreFoundation 0x000000018f2ba200 CFRunLoopRunSpecific (CoreFoundation)
4 AudioSession 0x00000001967b9158 GenericRunLoopThread::Entry(void*) (AudioSession)
5 AudioSession 0x00000001967bb0b8 CAPThread::Entry(CAPThread*) (AudioSession)
6 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
7 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 6 name:
Thread 6:
0 libsystem_kernel.dylib 0x00000001b9cb858c __workq_kernreturn (libsystem_kernel.dylib)
1 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 7 name:
Thread 7:
0 libsystem_kernel.dylib 0x00000001b9cb772c __psynch_cvwait (libsystem_kernel.dylib)
1 libc++.1.dylib 0x00000001a2c03e0c std::__1::condition_variable::wait(std::__1::unique_lock<std::__1::mutex>&) (libc++.1.dylib)
2 NewRelic 0x0000000100d7d88c 0x100cf4000 + 563340
3 NewRelic 0x0000000100d7db04 0x100cf4000 + 563972
4 NewRelic 0x0000000100d7da1c 0x100cf4000 + 563740
5 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
6 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 8 name:
Thread 8:
0 libsystem_kernel.dylib 0x00000001b9c948c4 mach_msg_trap (libsystem_kernel.dylib)
1 CoreFoundation 0x000000018f2c074c __CFRunLoopServiceMachPort (CoreFoundation)
2 CoreFoundation 0x000000018f2babd0 __CFRunLoopRun (CoreFoundation)
3 CoreFoundation 0x000000018f2ba200 CFRunLoopRunSpecific (CoreFoundation)
4 CFNetwork 0x000000018fb449a4 (null) (CFNetwork)
5 Foundation 0x0000000190620c48 __NSThread__start__ (Foundation)
6 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
7 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 9 name:
Thread 9:
0 libsystem_kernel.dylib 0x00000001b9c948c4 mach_msg_trap (libsystem_kernel.dylib)
1 CoreFoundation 0x000000018f2c074c __CFRunLoopServiceMachPort (CoreFoundation)
2 CoreFoundation 0x000000018f2babd0 __CFRunLoopRun (CoreFoundation)
3 CoreFoundation 0x000000018f2ba200 CFRunLoopRunSpecific (CoreFoundation)
4 Foundation 0x00000001904c4278 -[NSRunLoop(NSRunLoop) runMode:beforeDate:] (Foundation)
5 Foundation 0x00000001904f53e4 -[NSRunLoop(NSRunLoop) run] (Foundation)
6 JDBuild 0x0000000100424210 +[GAI threadMain:]
7 Foundation 0x0000000190620c48 __NSThread__start__ (Foundation)
8 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
9 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 10 name:
Thread 10:
0 libsystem_kernel.dylib 0x00000001b9cb858c __workq_kernreturn (libsystem_kernel.dylib)
1 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 11 name:
Thread 11:
0 libsystem_kernel.dylib 0x00000001b9cb858c __workq_kernreturn (libsystem_kernel.dylib)
1 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 12 name:
Thread 12:
0 libsystem_kernel.dylib 0x00000001b9cb858c __workq_kernreturn (libsystem_kernel.dylib)
1 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 13 name:
Thread 13:
0 libsystem_kernel.dylib 0x00000001b9cb858c __workq_kernreturn (libsystem_kernel.dylib)
1 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 14 name:
Thread 14:
**0 libsystem_kernel.dylib 0x00000001b9cb6b18 __guarded_open_dprotected_np (libsystem_kernel.dylib)
1 libsqlite3.dylib 0x00000001a760d444 (null) (libsqlite3.dylib)
2 libsqlite3.dylib 0x00000001a7561d0c (null) (libsqlite3.dylib)
3 libsqlite3.dylib 0x00000001a75c33d4 (null) (libsqlite3.dylib)
4 libsqlite3.dylib 0x00000001a75c2c04 (null) (libsqlite3.dylib)
5 libsqlite3.dylib 0x00000001a75a2130 (null) (libsqlite3.dylib)
6 libsqlite3.dylib 0x00000001a759c46c sqlite3_step (libsqlite3.dylib)
7 JDBuild 0x00000001005146f4 -[APMSqliteStore deleteRecordsFromTableWithName:condition:parameterValues:error:]
8 JDBuild 0x00000001004e0454 -[APMDatabase deleteBundlesWithRowIDs:error:]
9 JDBuild 0x00000001004f55e8 -[APMMeasurement removeBundlesFromDatabaseWithIDs:]
10 JDBuild 0x00000001004f1cbc -[APMMeasurement networkUploadCompletionHandlerWithResponse:error:]
11 JDBuild 0x00000001004f1670 __28-[APMMeasurement uploadData]_block_invoke**
12 libdispatch.dylib 0x000000018ef79298 _dispatch_call_block_and_release (libdispatch.dylib)
13 libdispatch.dylib 0x000000018ef7a280 _dispatch_client_callout (libdispatch.dylib)
14 libdispatch.dylib 0x000000018ef22fa8 _dispatch_lane_serial_drain$VARIANT$mp (libdispatch.dylib)
15 libdispatch.dylib 0x000000018ef23a84 _dispatch_lane_invoke$VARIANT$mp (libdispatch.dylib)
16 libdispatch.dylib 0x000000018ef2d518 _dispatch_workloop_worker_thread (libdispatch.dylib)
17 libsystem_pthread.dylib 0x00000001d4c245a4 _pthread_wqthread (libsystem_pthread.dylib)
18 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 15 name:
Thread 15:
0 libsystem_kernel.dylib 0x00000001b9c94918 semaphore_timedwait_trap (libsystem_kernel.dylib)
1 libdispatch.dylib 0x000000018ef1d97c _dispatch_semaphore_wait_slow (libdispatch.dylib)
2 libdispatch.dylib 0x000000018ef2bef4 _dispatch_worker_thread (libdispatch.dylib)
3 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
4 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 16 name:
Thread 16:
0 libsystem_kernel.dylib 0x00000001b9c94918 semaphore_timedwait_trap (libsystem_kernel.dylib)
1 libdispatch.dylib 0x000000018ef1d97c _dispatch_semaphore_wait_slow (libdispatch.dylib)
2 libdispatch.dylib 0x000000018ef2bef4 _dispatch_worker_thread (libdispatch.dylib)
3 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
4 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 17 name:
Thread 17:
0 libsystem_kernel.dylib 0x00000001b9c94918 semaphore_timedwait_trap (libsystem_kernel.dylib)
1 libdispatch.dylib 0x000000018ef1d97c _dispatch_semaphore_wait_slow (libdispatch.dylib)
2 AirshipCore 0x0000000100c2857c -[UADelay start] (UADelay.m:31)
3 AirshipCore 0x0000000100c39a10 __34-[UADelayOperation initWithDelay:]_block_invoke (UADelayOperation.m:17)
4 Foundation 0x00000001905ddcc8 __NSBLOCKOPERATION_IS_CALLING_OUT_TO_A_BLOCK__ (Foundation)
5 Foundation 0x00000001904dab18 -[NSBlockOperation main] (Foundation)
6 Foundation 0x00000001905dff80 __NSOPERATION_IS_INVOKING_MAIN__ (Foundation)
7 Foundation 0x00000001904da7d4 -[NSOperation start] (Foundation)
8 Foundation 0x00000001905e0978 __NSOPERATIONQUEUE_IS_STARTING_AN_OPERATION__ (Foundation)
9 Foundation 0x00000001905e0444 __NSOQSchedule_f (Foundation)
10 libdispatch.dylib 0x000000018ef2993c _dispatch_block_async_invoke2 (libdispatch.dylib)
11 libdispatch.dylib 0x000000018ef7a280 _dispatch_client_callout (libdispatch.dylib)
12 libdispatch.dylib 0x000000018ef1f56c _dispatch_continuation_pop$VARIANT$mp (libdispatch.dylib)
13 libdispatch.dylib 0x000000018ef1ecb0 _dispatch_async_redirect_invoke (libdispatch.dylib)
14 libdispatch.dylib 0x000000018ef2c11c _dispatch_root_queue_drain (libdispatch.dylib)
15 libdispatch.dylib 0x000000018ef2c8e4 _dispatch_worker_thread2 (libdispatch.dylib)
16 libsystem_pthread.dylib 0x00000001d4c24568 _pthread_wqthread (libsystem_pthread.dylib)
17 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 18 name:
Thread 18:
0 libsystem_kernel.dylib 0x00000001b9cb858c __workq_kernreturn (libsystem_kernel.dylib)
1 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 19 name:
**Thread 19 Crashed:**
0 libobjc.A.dylib 0x00000001a2b84df0 objc_release (libobjc.A.dylib)
1 Foundation 0x00000001905bb57c NSKeyValuePopPendingNotificationLocal (Foundation)
2 Foundation 0x00000001905bb4f8 NSKeyValueDidChange (Foundation)
3 Foundation 0x00000001905baf64 -[NSObject(NSKeyValueObservingPrivate) _changeValueForKeys:count:maybeOldValuesDict:maybeNewValuesDict:usingBlock:] (Foundation)
4 Foundation 0x00000001904ff800 -[NSObject(NSKeyValueObservingPrivate) _changeValueForKey:key:key:usingBlock:] (Foundation)
5 Foundation 0x00000001905b63d8 _NSSetLongLongValueAndNotify (Foundation)
6 CFNetwork 0x000000018f918470 (null) (CFNetwork)
7 CFNetwork 0x000000018f9b06c8 (null) (CFNetwork)
8 CFNetwork 0x000000018f91d630 (null) (CFNetwork)
9 CFNetwork 0x000000018f91d4c0 (null) (CFNetwork)
10 libdispatch.dylib 0x000000018ef2993c _dispatch_block_async_invoke2 (libdispatch.dylib)
11 libdispatch.dylib 0x000000018ef7a280 _dispatch_client_callout (libdispatch.dylib)
12 libdispatch.dylib 0x000000018ef22fa8 _dispatch_lane_serial_drain$VARIANT$mp (libdispatch.dylib)
13 libdispatch.dylib 0x000000018ef23ab4 _dispatch_lane_invoke$VARIANT$mp (libdispatch.dylib)
14 libdispatch.dylib 0x000000018ef2d518 _dispatch_workloop_worker_thread (libdispatch.dylib)
15 libsystem_pthread.dylib 0x00000001d4c245a4 _pthread_wqthread (libsystem_pthread.dylib)
16 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 21 name:
Thread 21:
0 libsystem_kernel.dylib 0x00000001b9c94918 semaphore_timedwait_trap (libsystem_kernel.dylib)
1 libdispatch.dylib 0x000000018ef1d97c _dispatch_semaphore_wait_slow (libdispatch.dylib)
2 libdispatch.dylib 0x000000018ef2bef4 _dispatch_worker_thread (libdispatch.dylib)
3 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
4 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
#### Steps to reproduce:
Unknown
|
True
|
Suspected Firebase SDK v6.34 crash from [APMMeasurement uploadData] - <!-- DO NOT DELETE
validate_template=true
template_path=.github/ISSUE_TEMPLATE/bug_report.md
-->
### [REQUIRED] Step 1: Describe your environment
* Xcode version: 12.0
* Firebase SDK version: 6.34.0
* Installation method: Carthage
* Firebase Component: iOS SDK/Performance
### [REQUIRED] Step 2: Describe the problem
Since a recent app update, we are seeing a new number 1 crash in our Ecommerce app. This occurred in the first build since I'd refreshed the Firebase SDK (and other dependencies) to the latest versions. I am not certain it is caused by Firebase but from the stack trace it seems the most likely cause. I do not have steps to reproduce the issue.
My analysis is:
1. The crashed thread (number 19) was calling back on a KVO change from CFNetwork-ing
2. Looking in our crash tracker (not Crashlytics for commercial reasons) I can see one of the last network calls before the crash was to https://app-measurement.com/a
3. Reviewing other background threads, I noticed thread 14 was doing a SQL-Lite delete initiated from [APMMeasurement uploadData]
4. Googling the method found in point 3 reveals historic Firebase issues (nothing applicable to this crash)
.crash file can be made available by email if required.
The complete stack trace is here:
Incident Identifier: 06FD8CD6-304B-4DF3-8894-E27C258C8A5A
CrashReporter Key: 0000000000000000000000000000000000000000
Hardware Model: iPhone9,3
Code Type: arm64
Date/Time: 2020-11-09T08:35:49+00:00
Launch Time: 2020-11-09T08:35:49+00:00
OS Version: iOS 14.1 (18A8395)
Report Version: 105
Exception Type: SIGSEGV
Exception Subtype:
Thread 0 name:
Thread 0:
0 libsystem_kernel.dylib 0x00000001b9c948c4 mach_msg_trap (libsystem_kernel.dylib)
1 CoreFoundation 0x000000018f2c074c __CFRunLoopServiceMachPort (CoreFoundation)
2 CoreFoundation 0x000000018f2babd0 __CFRunLoopRun (CoreFoundation)
3 CoreFoundation 0x000000018f2ba200 CFRunLoopRunSpecific (CoreFoundation)
4 GraphicsServices 0x00000001a53b5598 GSEventRunModal (GraphicsServices)
5 UIKitCore 0x0000000191b80004 -[UIApplication _run] (UIKitCore)
6 UIKitCore 0x0000000191b855d8 UIApplicationMain (UIKitCore)
7 JDBuild 0x000000010022764c main (coreAppDelegate.swift:14)
8 libdyld.dylib 0x000000018ef99598 start (libdyld.dylib)
Thread 1 name:
Thread 1:
0 libsystem_kernel.dylib 0x00000001b9c948c4 mach_msg_trap (libsystem_kernel.dylib)
1 CoreFoundation 0x000000018f2c074c __CFRunLoopServiceMachPort (CoreFoundation)
2 CoreFoundation 0x000000018f2babd0 __CFRunLoopRun (CoreFoundation)
3 CoreFoundation 0x000000018f2ba200 CFRunLoopRunSpecific (CoreFoundation)
4 Foundation 0x00000001904c4278 -[NSRunLoop(NSRunLoop) runMode:beforeDate:] (Foundation)
5 Foundation 0x00000001904c4158 -[NSRunLoop(NSRunLoop) runUntilDate:] (Foundation)
6 UIKitCore 0x0000000191c2b9fc -[UIEventFetcher threadMain] (UIKitCore)
7 Foundation 0x0000000190620c48 __NSThread__start__ (Foundation)
8 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
9 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 2 name:
Thread 2:
0 libsystem_kernel.dylib 0x00000001b9cb772c __psynch_cvwait (libsystem_kernel.dylib)
1 libc++.1.dylib 0x00000001a2c03e0c std::__1::condition_variable::wait(std::__1::unique_lock<std::__1::mutex>&) (libc++.1.dylib)
2 NewRelic 0x0000000100d7d88c 0x100cf4000 + 563340
3 NewRelic 0x0000000100d7db04 0x100cf4000 + 563972
4 NewRelic 0x0000000100d7da1c 0x100cf4000 + 563740
5 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
6 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 3 name:
Thread 3:
0 libsystem_kernel.dylib 0x00000001b9cb772c __psynch_cvwait (libsystem_kernel.dylib)
1 libc++.1.dylib 0x00000001a2c03e0c std::__1::condition_variable::wait(std::__1::unique_lock<std::__1::mutex>&) (libc++.1.dylib)
2 NewRelic 0x0000000100d7d88c 0x100cf4000 + 563340
3 NewRelic 0x0000000100d7db04 0x100cf4000 + 563972
4 NewRelic 0x0000000100d7da1c 0x100cf4000 + 563740
5 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
6 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 4 name:
Thread 4:
0 libsystem_kernel.dylib 0x00000001b9c948c4 mach_msg_trap (libsystem_kernel.dylib)
1 CoreFoundation 0x000000018f2c074c __CFRunLoopServiceMachPort (CoreFoundation)
2 CoreFoundation 0x000000018f2babd0 __CFRunLoopRun (CoreFoundation)
3 CoreFoundation 0x000000018f2ba200 CFRunLoopRunSpecific (CoreFoundation)
4 CoreFoundation 0x000000018f2bb2c0 CFRunLoopRun (CoreFoundation)
5 CoreMotion 0x0000000199ab2374 (null) (CoreMotion)
6 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
7 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 5 name:
Thread 5:
0 libsystem_kernel.dylib 0x00000001b9c948c4 mach_msg_trap (libsystem_kernel.dylib)
1 CoreFoundation 0x000000018f2c074c __CFRunLoopServiceMachPort (CoreFoundation)
2 CoreFoundation 0x000000018f2babd0 __CFRunLoopRun (CoreFoundation)
3 CoreFoundation 0x000000018f2ba200 CFRunLoopRunSpecific (CoreFoundation)
4 AudioSession 0x00000001967b9158 GenericRunLoopThread::Entry(void*) (AudioSession)
5 AudioSession 0x00000001967bb0b8 CAPThread::Entry(CAPThread*) (AudioSession)
6 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
7 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 6 name:
Thread 6:
0 libsystem_kernel.dylib 0x00000001b9cb858c __workq_kernreturn (libsystem_kernel.dylib)
1 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 7 name:
Thread 7:
0 libsystem_kernel.dylib 0x00000001b9cb772c __psynch_cvwait (libsystem_kernel.dylib)
1 libc++.1.dylib 0x00000001a2c03e0c std::__1::condition_variable::wait(std::__1::unique_lock<std::__1::mutex>&) (libc++.1.dylib)
2 NewRelic 0x0000000100d7d88c 0x100cf4000 + 563340
3 NewRelic 0x0000000100d7db04 0x100cf4000 + 563972
4 NewRelic 0x0000000100d7da1c 0x100cf4000 + 563740
5 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
6 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 8 name:
Thread 8:
0 libsystem_kernel.dylib 0x00000001b9c948c4 mach_msg_trap (libsystem_kernel.dylib)
1 CoreFoundation 0x000000018f2c074c __CFRunLoopServiceMachPort (CoreFoundation)
2 CoreFoundation 0x000000018f2babd0 __CFRunLoopRun (CoreFoundation)
3 CoreFoundation 0x000000018f2ba200 CFRunLoopRunSpecific (CoreFoundation)
4 CFNetwork 0x000000018fb449a4 (null) (CFNetwork)
5 Foundation 0x0000000190620c48 __NSThread__start__ (Foundation)
6 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
7 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 9 name:
Thread 9:
0 libsystem_kernel.dylib 0x00000001b9c948c4 mach_msg_trap (libsystem_kernel.dylib)
1 CoreFoundation 0x000000018f2c074c __CFRunLoopServiceMachPort (CoreFoundation)
2 CoreFoundation 0x000000018f2babd0 __CFRunLoopRun (CoreFoundation)
3 CoreFoundation 0x000000018f2ba200 CFRunLoopRunSpecific (CoreFoundation)
4 Foundation 0x00000001904c4278 -[NSRunLoop(NSRunLoop) runMode:beforeDate:] (Foundation)
5 Foundation 0x00000001904f53e4 -[NSRunLoop(NSRunLoop) run] (Foundation)
6 JDBuild 0x0000000100424210 +[GAI threadMain:]
7 Foundation 0x0000000190620c48 __NSThread__start__ (Foundation)
8 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
9 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 10 name:
Thread 10:
0 libsystem_kernel.dylib 0x00000001b9cb858c __workq_kernreturn (libsystem_kernel.dylib)
1 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 11 name:
Thread 11:
0 libsystem_kernel.dylib 0x00000001b9cb858c __workq_kernreturn (libsystem_kernel.dylib)
1 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 12 name:
Thread 12:
0 libsystem_kernel.dylib 0x00000001b9cb858c __workq_kernreturn (libsystem_kernel.dylib)
1 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 13 name:
Thread 13:
0 libsystem_kernel.dylib 0x00000001b9cb858c __workq_kernreturn (libsystem_kernel.dylib)
1 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 14 name:
Thread 14:
**0 libsystem_kernel.dylib 0x00000001b9cb6b18 __guarded_open_dprotected_np (libsystem_kernel.dylib)
1 libsqlite3.dylib 0x00000001a760d444 (null) (libsqlite3.dylib)
2 libsqlite3.dylib 0x00000001a7561d0c (null) (libsqlite3.dylib)
3 libsqlite3.dylib 0x00000001a75c33d4 (null) (libsqlite3.dylib)
4 libsqlite3.dylib 0x00000001a75c2c04 (null) (libsqlite3.dylib)
5 libsqlite3.dylib 0x00000001a75a2130 (null) (libsqlite3.dylib)
6 libsqlite3.dylib 0x00000001a759c46c sqlite3_step (libsqlite3.dylib)
7 JDBuild 0x00000001005146f4 -[APMSqliteStore deleteRecordsFromTableWithName:condition:parameterValues:error:]
8 JDBuild 0x00000001004e0454 -[APMDatabase deleteBundlesWithRowIDs:error:]
9 JDBuild 0x00000001004f55e8 -[APMMeasurement removeBundlesFromDatabaseWithIDs:]
10 JDBuild 0x00000001004f1cbc -[APMMeasurement networkUploadCompletionHandlerWithResponse:error:]
11 JDBuild 0x00000001004f1670 __28-[APMMeasurement uploadData]_block_invoke**
12 libdispatch.dylib 0x000000018ef79298 _dispatch_call_block_and_release (libdispatch.dylib)
13 libdispatch.dylib 0x000000018ef7a280 _dispatch_client_callout (libdispatch.dylib)
14 libdispatch.dylib 0x000000018ef22fa8 _dispatch_lane_serial_drain$VARIANT$mp (libdispatch.dylib)
15 libdispatch.dylib 0x000000018ef23a84 _dispatch_lane_invoke$VARIANT$mp (libdispatch.dylib)
16 libdispatch.dylib 0x000000018ef2d518 _dispatch_workloop_worker_thread (libdispatch.dylib)
17 libsystem_pthread.dylib 0x00000001d4c245a4 _pthread_wqthread (libsystem_pthread.dylib)
18 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 15 name:
Thread 15:
0 libsystem_kernel.dylib 0x00000001b9c94918 semaphore_timedwait_trap (libsystem_kernel.dylib)
1 libdispatch.dylib 0x000000018ef1d97c _dispatch_semaphore_wait_slow (libdispatch.dylib)
2 libdispatch.dylib 0x000000018ef2bef4 _dispatch_worker_thread (libdispatch.dylib)
3 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
4 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 16 name:
Thread 16:
0 libsystem_kernel.dylib 0x00000001b9c94918 semaphore_timedwait_trap (libsystem_kernel.dylib)
1 libdispatch.dylib 0x000000018ef1d97c _dispatch_semaphore_wait_slow (libdispatch.dylib)
2 libdispatch.dylib 0x000000018ef2bef4 _dispatch_worker_thread (libdispatch.dylib)
3 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
4 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
Thread 17 name:
Thread 17:
0 libsystem_kernel.dylib 0x00000001b9c94918 semaphore_timedwait_trap (libsystem_kernel.dylib)
1 libdispatch.dylib 0x000000018ef1d97c _dispatch_semaphore_wait_slow (libdispatch.dylib)
2 AirshipCore 0x0000000100c2857c -[UADelay start] (UADelay.m:31)
3 AirshipCore 0x0000000100c39a10 __34-[UADelayOperation initWithDelay:]_block_invoke (UADelayOperation.m:17)
4 Foundation 0x00000001905ddcc8 __NSBLOCKOPERATION_IS_CALLING_OUT_TO_A_BLOCK__ (Foundation)
5 Foundation 0x00000001904dab18 -[NSBlockOperation main] (Foundation)
6 Foundation 0x00000001905dff80 __NSOPERATION_IS_INVOKING_MAIN__ (Foundation)
7 Foundation 0x00000001904da7d4 -[NSOperation start] (Foundation)
8 Foundation 0x00000001905e0978 __NSOPERATIONQUEUE_IS_STARTING_AN_OPERATION__ (Foundation)
9 Foundation 0x00000001905e0444 __NSOQSchedule_f (Foundation)
10 libdispatch.dylib 0x000000018ef2993c _dispatch_block_async_invoke2 (libdispatch.dylib)
11 libdispatch.dylib 0x000000018ef7a280 _dispatch_client_callout (libdispatch.dylib)
12 libdispatch.dylib 0x000000018ef1f56c _dispatch_continuation_pop$VARIANT$mp (libdispatch.dylib)
13 libdispatch.dylib 0x000000018ef1ecb0 _dispatch_async_redirect_invoke (libdispatch.dylib)
14 libdispatch.dylib 0x000000018ef2c11c _dispatch_root_queue_drain (libdispatch.dylib)
15 libdispatch.dylib 0x000000018ef2c8e4 _dispatch_worker_thread2 (libdispatch.dylib)
16 libsystem_pthread.dylib 0x00000001d4c24568 _pthread_wqthread (libsystem_pthread.dylib)
17 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 18 name:
Thread 18:
0 libsystem_kernel.dylib 0x00000001b9cb858c __workq_kernreturn (libsystem_kernel.dylib)
1 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 19 name:
**Thread 19 Crashed:**
0 libobjc.A.dylib 0x00000001a2b84df0 objc_release (libobjc.A.dylib)
1 Foundation 0x00000001905bb57c NSKeyValuePopPendingNotificationLocal (Foundation)
2 Foundation 0x00000001905bb4f8 NSKeyValueDidChange (Foundation)
3 Foundation 0x00000001905baf64 -[NSObject(NSKeyValueObservingPrivate) _changeValueForKeys:count:maybeOldValuesDict:maybeNewValuesDict:usingBlock:] (Foundation)
4 Foundation 0x00000001904ff800 -[NSObject(NSKeyValueObservingPrivate) _changeValueForKey:key:key:usingBlock:] (Foundation)
5 Foundation 0x00000001905b63d8 _NSSetLongLongValueAndNotify (Foundation)
6 CFNetwork 0x000000018f918470 (null) (CFNetwork)
7 CFNetwork 0x000000018f9b06c8 (null) (CFNetwork)
8 CFNetwork 0x000000018f91d630 (null) (CFNetwork)
9 CFNetwork 0x000000018f91d4c0 (null) (CFNetwork)
10 libdispatch.dylib 0x000000018ef2993c _dispatch_block_async_invoke2 (libdispatch.dylib)
11 libdispatch.dylib 0x000000018ef7a280 _dispatch_client_callout (libdispatch.dylib)
12 libdispatch.dylib 0x000000018ef22fa8 _dispatch_lane_serial_drain$VARIANT$mp (libdispatch.dylib)
13 libdispatch.dylib 0x000000018ef23ab4 _dispatch_lane_invoke$VARIANT$mp (libdispatch.dylib)
14 libdispatch.dylib 0x000000018ef2d518 _dispatch_workloop_worker_thread (libdispatch.dylib)
15 libsystem_pthread.dylib 0x00000001d4c245a4 _pthread_wqthread (libsystem_pthread.dylib)
16 libsystem_pthread.dylib 0x00000001d4c27874 start_wqthread (libsystem_pthread.dylib)
Thread 21 name:
Thread 21:
0 libsystem_kernel.dylib 0x00000001b9c94918 semaphore_timedwait_trap (libsystem_kernel.dylib)
1 libdispatch.dylib 0x000000018ef1d97c _dispatch_semaphore_wait_slow (libdispatch.dylib)
2 libdispatch.dylib 0x000000018ef2bef4 _dispatch_worker_thread (libdispatch.dylib)
3 libsystem_pthread.dylib 0x00000001d4c22b70 _pthread_start (libsystem_pthread.dylib)
4 libsystem_pthread.dylib 0x00000001d4c27880 thread_start (libsystem_pthread.dylib)
#### Steps to reproduce:
Unknown
|
non_defect
|
suspected firebase sdk crash from do not delete validate template true template path github issue template bug report md step describe your environment xcode version firebase sdk version installation method carthage firebase component ios sdk performance step describe the problem since a recent app update we are seeing a new number crash in our ecommerce app this occurred in the first build since i d refreshed the firebase sdk and other dependencies to the latest versions i am not certain it is caused by firebase but from the stack trace it seems the most likely cause i do not have steps to reproduce the issue my analysis is the crashed thread number was calling back on a kvo change from cfnetwork ing looking in our crash tracker not crashlytics for commercial reasons i can see one of the last network calls before the crash was to reviewing other background threads i noticed thread was doing a sql lite delete initiated from googling the method found in point reveals historic firebase issues nothing applicable to this crash crash file can be made available by email if required the complete stack trace is here incident identifier crashreporter key hardware model code type date time launch time os version ios report version exception type sigsegv exception subtype thread name thread libsystem kernel dylib mach msg trap libsystem kernel dylib corefoundation cfrunloopservicemachport corefoundation corefoundation cfrunlooprun corefoundation corefoundation cfrunlooprunspecific corefoundation graphicsservices gseventrunmodal graphicsservices uikitcore uikitcore uikitcore uiapplicationmain uikitcore jdbuild main coreappdelegate swift libdyld dylib start libdyld dylib thread name thread libsystem kernel dylib mach msg trap libsystem kernel dylib corefoundation cfrunloopservicemachport corefoundation corefoundation cfrunlooprun corefoundation corefoundation cfrunlooprunspecific corefoundation foundation foundation foundation foundation uikitcore uikitcore foundation nsthread start foundation libsystem pthread dylib pthread start libsystem pthread dylib libsystem pthread dylib thread start libsystem pthread dylib thread name thread libsystem kernel dylib psynch cvwait libsystem kernel dylib libc dylib std condition variable wait std unique lock libc dylib newrelic newrelic newrelic libsystem pthread dylib pthread start libsystem pthread dylib libsystem pthread dylib thread start libsystem pthread dylib thread name thread libsystem kernel dylib psynch cvwait libsystem kernel dylib libc dylib std condition variable wait std unique lock libc dylib newrelic newrelic newrelic libsystem pthread dylib pthread start libsystem pthread dylib libsystem pthread dylib thread start libsystem pthread dylib thread name thread libsystem kernel dylib mach msg trap libsystem kernel dylib corefoundation cfrunloopservicemachport corefoundation corefoundation cfrunlooprun corefoundation corefoundation cfrunlooprunspecific corefoundation corefoundation cfrunlooprun corefoundation coremotion null coremotion libsystem pthread dylib pthread start libsystem pthread dylib libsystem pthread dylib thread start libsystem pthread dylib thread name thread libsystem kernel dylib mach msg trap libsystem kernel dylib corefoundation cfrunloopservicemachport corefoundation corefoundation cfrunlooprun corefoundation corefoundation cfrunlooprunspecific corefoundation audiosession genericrunloopthread entry void audiosession audiosession capthread entry capthread audiosession libsystem pthread dylib pthread start libsystem pthread dylib libsystem pthread dylib thread start libsystem pthread dylib thread name thread libsystem kernel dylib workq kernreturn libsystem kernel dylib libsystem pthread dylib start wqthread libsystem pthread dylib thread name thread libsystem kernel dylib psynch cvwait libsystem kernel dylib libc dylib std condition variable wait std unique lock libc dylib newrelic newrelic newrelic libsystem pthread dylib pthread start libsystem pthread dylib libsystem pthread dylib thread start libsystem pthread dylib thread name thread libsystem kernel dylib mach msg trap libsystem kernel dylib corefoundation cfrunloopservicemachport corefoundation corefoundation cfrunlooprun corefoundation corefoundation cfrunlooprunspecific corefoundation cfnetwork null cfnetwork foundation nsthread start foundation libsystem pthread dylib pthread start libsystem pthread dylib libsystem pthread dylib thread start libsystem pthread dylib thread name thread libsystem kernel dylib mach msg trap libsystem kernel dylib corefoundation cfrunloopservicemachport corefoundation corefoundation cfrunlooprun corefoundation corefoundation cfrunlooprunspecific corefoundation foundation foundation foundation foundation jdbuild foundation nsthread start foundation libsystem pthread dylib pthread start libsystem pthread dylib libsystem pthread dylib thread start libsystem pthread dylib thread name thread libsystem kernel dylib workq kernreturn libsystem kernel dylib libsystem pthread dylib start wqthread libsystem pthread dylib thread name thread libsystem kernel dylib workq kernreturn libsystem kernel dylib libsystem pthread dylib start wqthread libsystem pthread dylib thread name thread libsystem kernel dylib workq kernreturn libsystem kernel dylib libsystem pthread dylib start wqthread libsystem pthread dylib thread name thread libsystem kernel dylib workq kernreturn libsystem kernel dylib libsystem pthread dylib start wqthread libsystem pthread dylib thread name thread libsystem kernel dylib guarded open dprotected np libsystem kernel dylib dylib null dylib dylib null dylib dylib null dylib dylib null dylib dylib null dylib dylib step dylib jdbuild jdbuild jdbuild jdbuild jdbuild block invoke libdispatch dylib dispatch call block and release libdispatch dylib libdispatch dylib dispatch client callout libdispatch dylib libdispatch dylib dispatch lane serial drain variant mp libdispatch dylib libdispatch dylib dispatch lane invoke variant mp libdispatch dylib libdispatch dylib dispatch workloop worker thread libdispatch dylib libsystem pthread dylib pthread wqthread libsystem pthread dylib libsystem pthread dylib start wqthread libsystem pthread dylib thread name thread libsystem kernel dylib semaphore timedwait trap libsystem kernel dylib libdispatch dylib dispatch semaphore wait slow libdispatch dylib libdispatch dylib dispatch worker thread libdispatch dylib libsystem pthread dylib pthread start libsystem pthread dylib libsystem pthread dylib thread start libsystem pthread dylib thread name thread libsystem kernel dylib semaphore timedwait trap libsystem kernel dylib libdispatch dylib dispatch semaphore wait slow libdispatch dylib libdispatch dylib dispatch worker thread libdispatch dylib libsystem pthread dylib pthread start libsystem pthread dylib libsystem pthread dylib thread start libsystem pthread dylib thread name thread libsystem kernel dylib semaphore timedwait trap libsystem kernel dylib libdispatch dylib dispatch semaphore wait slow libdispatch dylib airshipcore uadelay m airshipcore block invoke uadelayoperation m foundation nsblockoperation is calling out to a block foundation foundation foundation foundation nsoperation is invoking main foundation foundation foundation foundation nsoperationqueue is starting an operation foundation foundation nsoqschedule f foundation libdispatch dylib dispatch block async libdispatch dylib libdispatch dylib dispatch client callout libdispatch dylib libdispatch dylib dispatch continuation pop variant mp libdispatch dylib libdispatch dylib dispatch async redirect invoke libdispatch dylib libdispatch dylib dispatch root queue drain libdispatch dylib libdispatch dylib dispatch worker libdispatch dylib libsystem pthread dylib pthread wqthread libsystem pthread dylib libsystem pthread dylib start wqthread libsystem pthread dylib thread name thread libsystem kernel dylib workq kernreturn libsystem kernel dylib libsystem pthread dylib start wqthread libsystem pthread dylib thread name thread crashed libobjc a dylib objc release libobjc a dylib foundation nskeyvaluepoppendingnotificationlocal foundation foundation nskeyvaluedidchange foundation foundation foundation foundation foundation foundation nssetlonglongvalueandnotify foundation cfnetwork null cfnetwork cfnetwork null cfnetwork cfnetwork null cfnetwork cfnetwork null cfnetwork libdispatch dylib dispatch block async libdispatch dylib libdispatch dylib dispatch client callout libdispatch dylib libdispatch dylib dispatch lane serial drain variant mp libdispatch dylib libdispatch dylib dispatch lane invoke variant mp libdispatch dylib libdispatch dylib dispatch workloop worker thread libdispatch dylib libsystem pthread dylib pthread wqthread libsystem pthread dylib libsystem pthread dylib start wqthread libsystem pthread dylib thread name thread libsystem kernel dylib semaphore timedwait trap libsystem kernel dylib libdispatch dylib dispatch semaphore wait slow libdispatch dylib libdispatch dylib dispatch worker thread libdispatch dylib libsystem pthread dylib pthread start libsystem pthread dylib libsystem pthread dylib thread start libsystem pthread dylib steps to reproduce unknown
| 0
|
31,896
| 6,657,858,644
|
IssuesEvent
|
2017-09-30 11:42:39
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
Mobile page navigation doesn't work if the pages are not the immediate children of the `body`
|
defect
|
```
<h:body>
<h:panelGroup id="my_page_container">
<pm:page>
...
```
Attempting to navigate between pages (p:link, p:button or PrimeFaces.Mobile.navigate) results only in a console error
> Uncaught Error: cannot call methods on pagecontainer prior to initialization; attempted to call method 'change'
jQuery Mobile automatically correctly determines the page container to be `#my_page_container`, and that's where it creates the widget instance. But the problem is PrimeFaces.Mobile.navigate expects the page container to be on the `body` tag. Probably should use `$(":mobile-pagecontainer")` instead.
Encountered with PF5.2.12, seems to be relevant in 5.3 judging by the core.mobile.js.
|
1.0
|
Mobile page navigation doesn't work if the pages are not the immediate children of the `body` - ```
<h:body>
<h:panelGroup id="my_page_container">
<pm:page>
...
```
Attempting to navigate between pages (p:link, p:button or PrimeFaces.Mobile.navigate) results only in a console error
> Uncaught Error: cannot call methods on pagecontainer prior to initialization; attempted to call method 'change'
jQuery Mobile automatically correctly determines the page container to be `#my_page_container`, and that's where it creates the widget instance. But the problem is PrimeFaces.Mobile.navigate expects the page container to be on the `body` tag. Probably should use `$(":mobile-pagecontainer")` instead.
Encountered with PF5.2.12, seems to be relevant in 5.3 judging by the core.mobile.js.
|
defect
|
mobile page navigation doesn t work if the pages are not the immediate children of the body attempting to navigate between pages p link p button or primefaces mobile navigate results only in a console error uncaught error cannot call methods on pagecontainer prior to initialization attempted to call method change jquery mobile automatically correctly determines the page container to be my page container and that s where it creates the widget instance but the problem is primefaces mobile navigate expects the page container to be on the body tag probably should use mobile pagecontainer instead encountered with seems to be relevant in judging by the core mobile js
| 1
|
34,429
| 7,451,357,426
|
IssuesEvent
|
2018-03-29 02:28:04
|
kerdokullamae/test_koik_issued
|
https://api.github.com/repos/kerdokullamae/test_koik_issued
|
closed
|
nimistute arv all-arhiivide korral vigane
|
C: AIS P: high R: fixed T: defect
|
**Reported by katrin vesterblom on 27 Mar 2013 08:59 UTC**
aadress rahvusarhiiv.tietotest.ee .
ERA.185 nimistute arvuks näitab 89, tegelikult peab olema 2. Enamus ridu (87) on absurdsed.
Kas viga on tulnud sellest, et arhiiv jaguneb esmalt all-arhiivideks ja lisaks on iga all-arhiiviga seotud oma arhiivimoodustaja (isik/organisatsioon)?
Sama lugu on arhiiviga ERA.99 ja ERA.660. Ka need jagunevad esmalt all-arhiivideks, mis omavad moodustajaid.
Arhiiv ERA.1 on korras, kuigi ka tema jaguneb all-arhiivideks, aga temal ei ole all-arhiividega seotud moodustajaid (isikuid).
|
1.0
|
nimistute arv all-arhiivide korral vigane - **Reported by katrin vesterblom on 27 Mar 2013 08:59 UTC**
aadress rahvusarhiiv.tietotest.ee .
ERA.185 nimistute arvuks näitab 89, tegelikult peab olema 2. Enamus ridu (87) on absurdsed.
Kas viga on tulnud sellest, et arhiiv jaguneb esmalt all-arhiivideks ja lisaks on iga all-arhiiviga seotud oma arhiivimoodustaja (isik/organisatsioon)?
Sama lugu on arhiiviga ERA.99 ja ERA.660. Ka need jagunevad esmalt all-arhiivideks, mis omavad moodustajaid.
Arhiiv ERA.1 on korras, kuigi ka tema jaguneb all-arhiivideks, aga temal ei ole all-arhiividega seotud moodustajaid (isikuid).
|
defect
|
nimistute arv all arhiivide korral vigane reported by katrin vesterblom on mar utc aadress rahvusarhiiv tietotest ee era nimistute arvuks näitab tegelikult peab olema enamus ridu on absurdsed kas viga on tulnud sellest et arhiiv jaguneb esmalt all arhiivideks ja lisaks on iga all arhiiviga seotud oma arhiivimoodustaja isik organisatsioon sama lugu on arhiiviga era ja era ka need jagunevad esmalt all arhiivideks mis omavad moodustajaid arhiiv era on korras kuigi ka tema jaguneb all arhiivideks aga temal ei ole all arhiividega seotud moodustajaid isikuid
| 1
|
6,784
| 7,754,206,505
|
IssuesEvent
|
2018-05-31 05:28:20
|
terraform-providers/terraform-provider-aws
|
https://api.github.com/repos/terraform-providers/terraform-provider-aws
|
closed
|
Bug in 1.9.0 when creating route53 entries for ACM approvals
|
bug service/acm waiting-response
|
This occurs every time on the first run of a terraform apply...on a subsequent run it works fine (So our current work around is run it again)
NOTE: I Have censored our actual DNS names for this as it might contain information we don't want posted to a public location
* module.service_definition.aws_route53_record.cn_cert_validation: aws_route53_record.cn_cert_validation: diffs didn't match during apply. This is a bug with Terraform and should be reported as a GitHub Issue.
Please include the following information in your report:
Terraform Version: 0.11.1
Resource ID: aws_route53_record.cn_cert_validation
Mismatch reason: attribute mismatch: records.942692827
Diff One (usually from plan): *terraform.InstanceDiff{mu:sync.Mutex{state:0, sema:0x0}, Attributes:map[string]*terraform.ResourceAttrDiff{"ttl":*terraform.ResourceAttrDiff{Old:"", New:"60", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "type":*terraform.ResourceAttrDiff{Old:"", New:"CNAME", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "records.#":*terraform.ResourceAttrDiff{Old:"", New:"1", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "records.942692827":*terraform.ResourceAttrDiff{Old:"", New:"_1480da86a40a658a93bd31e006759e1a.acm-validations.aws.", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "name":*terraform.ResourceAttrDiff{Old:"", New:"_8154b685e678779c8f64f2bd964294b0.foo.bar.com", NewComputed:false, NewRemoved:false, NewExtra:"_8154b685e678779c8f64f2bd964294b0.foo.bar.com.", RequiresNew:true, Sensitive:false, Type:0x0}, "fqdn":*terraform.ResourceAttrDiff{Old:"", New:"", NewComputed:true, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "zone_id":*terraform.ResourceAttrDiff{Old:"", New:"Z3PNQYIZM8851A", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:true, Sensitive:false, Type:0x0}}, Destroy:false, DestroyDeposed:false, DestroyTainted:false, Meta:map[string]interface {}(nil)}
Diff Two (usually from apply): *terraform.InstanceDiff{mu:sync.Mutex{state:0, sema:0x0}, Attributes:map[string]*terraform.ResourceAttrDiff{"name":*terraform.ResourceAttrDiff{Old:"", New:"_1b72220dfd38e5357f1e4da5bcf452f4.zsryq6c8.foo.bar.com", NewComputed:false, NewRemoved:false, NewExtra:"_1b72220dfd38e5357f1e4da5bcf452f4.zsryq6c8.foo.bar.com.", RequiresNew:true, Sensitive:false, Type:0x0}, "records.#":*terraform.ResourceAttrDiff{Old:"", New:"1", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "records.3982558149":*terraform.ResourceAttrDiff{Old:"", New:"_7d301af98c6434cfd04ed0550f2a7164.acm-validations.aws.", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "fqdn":*terraform.ResourceAttrDiff{Old:"", New:"", NewComputed:true, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "ttl":*terraform.ResourceAttrDiff{Old:"", New:"60", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "type":*terraform.ResourceAttrDiff{Old:"", New:"CNAME", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "zone_id":*terraform.ResourceAttrDiff{Old:"", New:"Z3PNQYIZM8851A", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:true, Sensitive:false, Type:0x0}}, Destroy:false, DestroyDeposed:false, DestroyTainted:false, Meta:map[string]interface {}(nil)}
Also include as much context as you can about your config, state, and the steps you performed to trigger this error.
This is what the code looks like
```hcl
resource "aws_acm_certificate" "certificate" {
# Using app name for common name to avoid CN character limitations
domain_name = "${aws_route53_record.random_subdomain.fqdn}"
validation_method = "DNS"
# Your DNS will be set as the first SAN in the list
subject_alternative_names = ["${aws_route53_record.alb.fqdn}"]
tags {
Owner = "${var.owner}"
Name = "The Certificate for the ${var.app_name} service in ${var.vpc} deployed to ${var.region}"
}
lifecycle {
create_before_destroy = true
}
}
resource "aws_route53_record" "cn_cert_validation" {
name = "${aws_acm_certificate.certificate.domain_validation_options.0.resource_record_name}"
type = "${aws_acm_certificate.certificate.domain_validation_options.0.resource_record_type}"
zone_id = "${data.aws_route53_zone.approval_zone.id}"
records = ["${aws_acm_certificate.certificate.domain_validation_options.0.resource_record_value}"]
ttl = 60
}
resource "aws_route53_record" "san_cert_validation" {
name = "${aws_acm_certificate.certificate.domain_validation_options.1.resource_record_name}"
type = "${aws_acm_certificate.certificate.domain_validation_options.1.resource_record_type}"
zone_id = "${data.aws_route53_zone.approval_zone.id}"
records = ["${aws_acm_certificate.certificate.domain_validation_options.1.resource_record_value}"]
ttl = 60
}
```
Thanks for any assistance on this issue.
-Jeff
|
1.0
|
Bug in 1.9.0 when creating route53 entries for ACM approvals - This occurs every time on the first run of a terraform apply...on a subsequent run it works fine (So our current work around is run it again)
NOTE: I Have censored our actual DNS names for this as it might contain information we don't want posted to a public location
* module.service_definition.aws_route53_record.cn_cert_validation: aws_route53_record.cn_cert_validation: diffs didn't match during apply. This is a bug with Terraform and should be reported as a GitHub Issue.
Please include the following information in your report:
Terraform Version: 0.11.1
Resource ID: aws_route53_record.cn_cert_validation
Mismatch reason: attribute mismatch: records.942692827
Diff One (usually from plan): *terraform.InstanceDiff{mu:sync.Mutex{state:0, sema:0x0}, Attributes:map[string]*terraform.ResourceAttrDiff{"ttl":*terraform.ResourceAttrDiff{Old:"", New:"60", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "type":*terraform.ResourceAttrDiff{Old:"", New:"CNAME", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "records.#":*terraform.ResourceAttrDiff{Old:"", New:"1", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "records.942692827":*terraform.ResourceAttrDiff{Old:"", New:"_1480da86a40a658a93bd31e006759e1a.acm-validations.aws.", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "name":*terraform.ResourceAttrDiff{Old:"", New:"_8154b685e678779c8f64f2bd964294b0.foo.bar.com", NewComputed:false, NewRemoved:false, NewExtra:"_8154b685e678779c8f64f2bd964294b0.foo.bar.com.", RequiresNew:true, Sensitive:false, Type:0x0}, "fqdn":*terraform.ResourceAttrDiff{Old:"", New:"", NewComputed:true, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "zone_id":*terraform.ResourceAttrDiff{Old:"", New:"Z3PNQYIZM8851A", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:true, Sensitive:false, Type:0x0}}, Destroy:false, DestroyDeposed:false, DestroyTainted:false, Meta:map[string]interface {}(nil)}
Diff Two (usually from apply): *terraform.InstanceDiff{mu:sync.Mutex{state:0, sema:0x0}, Attributes:map[string]*terraform.ResourceAttrDiff{"name":*terraform.ResourceAttrDiff{Old:"", New:"_1b72220dfd38e5357f1e4da5bcf452f4.zsryq6c8.foo.bar.com", NewComputed:false, NewRemoved:false, NewExtra:"_1b72220dfd38e5357f1e4da5bcf452f4.zsryq6c8.foo.bar.com.", RequiresNew:true, Sensitive:false, Type:0x0}, "records.#":*terraform.ResourceAttrDiff{Old:"", New:"1", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "records.3982558149":*terraform.ResourceAttrDiff{Old:"", New:"_7d301af98c6434cfd04ed0550f2a7164.acm-validations.aws.", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "fqdn":*terraform.ResourceAttrDiff{Old:"", New:"", NewComputed:true, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "ttl":*terraform.ResourceAttrDiff{Old:"", New:"60", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "type":*terraform.ResourceAttrDiff{Old:"", New:"CNAME", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:false, Sensitive:false, Type:0x0}, "zone_id":*terraform.ResourceAttrDiff{Old:"", New:"Z3PNQYIZM8851A", NewComputed:false, NewRemoved:false, NewExtra:interface {}(nil), RequiresNew:true, Sensitive:false, Type:0x0}}, Destroy:false, DestroyDeposed:false, DestroyTainted:false, Meta:map[string]interface {}(nil)}
Also include as much context as you can about your config, state, and the steps you performed to trigger this error.
This is what the code looks like
```hcl
resource "aws_acm_certificate" "certificate" {
# Using app name for common name to avoid CN character limitations
domain_name = "${aws_route53_record.random_subdomain.fqdn}"
validation_method = "DNS"
# Your DNS will be set as the first SAN in the list
subject_alternative_names = ["${aws_route53_record.alb.fqdn}"]
tags {
Owner = "${var.owner}"
Name = "The Certificate for the ${var.app_name} service in ${var.vpc} deployed to ${var.region}"
}
lifecycle {
create_before_destroy = true
}
}
resource "aws_route53_record" "cn_cert_validation" {
name = "${aws_acm_certificate.certificate.domain_validation_options.0.resource_record_name}"
type = "${aws_acm_certificate.certificate.domain_validation_options.0.resource_record_type}"
zone_id = "${data.aws_route53_zone.approval_zone.id}"
records = ["${aws_acm_certificate.certificate.domain_validation_options.0.resource_record_value}"]
ttl = 60
}
resource "aws_route53_record" "san_cert_validation" {
name = "${aws_acm_certificate.certificate.domain_validation_options.1.resource_record_name}"
type = "${aws_acm_certificate.certificate.domain_validation_options.1.resource_record_type}"
zone_id = "${data.aws_route53_zone.approval_zone.id}"
records = ["${aws_acm_certificate.certificate.domain_validation_options.1.resource_record_value}"]
ttl = 60
}
```
Thanks for any assistance on this issue.
-Jeff
|
non_defect
|
bug in when creating entries for acm approvals this occurs every time on the first run of a terraform apply on a subsequent run it works fine so our current work around is run it again note i have censored our actual dns names for this as it might contain information we don t want posted to a public location module service definition aws record cn cert validation aws record cn cert validation diffs didn t match during apply this is a bug with terraform and should be reported as a github issue please include the following information in your report terraform version resource id aws record cn cert validation mismatch reason attribute mismatch records diff one usually from plan terraform instancediff mu sync mutex state sema attributes map terraform resourceattrdiff ttl terraform resourceattrdiff old new newcomputed false newremoved false newextra interface nil requiresnew false sensitive false type type terraform resourceattrdiff old new cname newcomputed false newremoved false newextra interface nil requiresnew false sensitive false type records terraform resourceattrdiff old new newcomputed false newremoved false newextra interface nil requiresnew false sensitive false type records terraform resourceattrdiff old new acm validations aws newcomputed false newremoved false newextra interface nil requiresnew false sensitive false type name terraform resourceattrdiff old new foo bar com newcomputed false newremoved false newextra foo bar com requiresnew true sensitive false type fqdn terraform resourceattrdiff old new newcomputed true newremoved false newextra interface nil requiresnew false sensitive false type zone id terraform resourceattrdiff old new newcomputed false newremoved false newextra interface nil requiresnew true sensitive false type destroy false destroydeposed false destroytainted false meta map interface nil diff two usually from apply terraform instancediff mu sync mutex state sema attributes map terraform resourceattrdiff name terraform resourceattrdiff old new foo bar com newcomputed false newremoved false newextra foo bar com requiresnew true sensitive false type records terraform resourceattrdiff old new newcomputed false newremoved false newextra interface nil requiresnew false sensitive false type records terraform resourceattrdiff old new acm validations aws newcomputed false newremoved false newextra interface nil requiresnew false sensitive false type fqdn terraform resourceattrdiff old new newcomputed true newremoved false newextra interface nil requiresnew false sensitive false type ttl terraform resourceattrdiff old new newcomputed false newremoved false newextra interface nil requiresnew false sensitive false type type terraform resourceattrdiff old new cname newcomputed false newremoved false newextra interface nil requiresnew false sensitive false type zone id terraform resourceattrdiff old new newcomputed false newremoved false newextra interface nil requiresnew true sensitive false type destroy false destroydeposed false destroytainted false meta map interface nil also include as much context as you can about your config state and the steps you performed to trigger this error this is what the code looks like hcl resource aws acm certificate certificate using app name for common name to avoid cn character limitations domain name aws record random subdomain fqdn validation method dns your dns will be set as the first san in the list subject alternative names tags owner var owner name the certificate for the var app name service in var vpc deployed to var region lifecycle create before destroy true resource aws record cn cert validation name aws acm certificate certificate domain validation options resource record name type aws acm certificate certificate domain validation options resource record type zone id data aws zone approval zone id records ttl resource aws record san cert validation name aws acm certificate certificate domain validation options resource record name type aws acm certificate certificate domain validation options resource record type zone id data aws zone approval zone id records ttl thanks for any assistance on this issue jeff
| 0
|
80,243
| 30,183,486,765
|
IssuesEvent
|
2023-07-04 10:25:28
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
opened
|
TST: test_axis_nan_policy_full failure
|
defect
|
### Describe your issue.
test_axis_nan_policy_full tests fails.
I built SciPy 1.11.1 manually with Intel OneAPI MKL as BLAS on Ubuntu 22.04. CPU with AVX512 support (NumPy/Scipy built with --cpu-baseline="native" flag).
Similar to resolved earlier https://github.com/scipy/scipy/issues/16337
Failure log for **test_axis_nan_policy_full[all_finite-2-propagate-jarque_bera-args17-kwds17-1-2-False-None]**, **test_axis_nan_policy_full[all_finite-2-omit-jarque_bera-args17-kwds17-1-2-False-None]**, **test_axis_nan_policy_full[all_finite-2-raise-jarque_bera-args17-kwds17-1-2-False-None]** tests is omitted as there was error "There was an error creating your issue: body is too long (maximum is 65536 characters)."
### Reproducing Code Example
```python
import scipy as sp
sp.test('full')
```
### Error message
```shell
_______________________________________________________________________ test_axis_nan_policy_full[all_finite--3-propagate-jarque_bera-args17-kwds17-1-2-False-None] ________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -3
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'propagate'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -3
data = [array([[[0.63696169, 0.02831967],
[0.57152983, 0.40455184],
[0.75772885, 0.47998792]],
[[0.26...0.84131728]],
[[0.42268722, 0.3577952 ],
[0.0520213 , 0.46004514],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -3, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'propagate'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F49E9D3D540
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f4954df24d0>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a2f80>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49550a2710>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
__________________________________________________________________________ test_axis_nan_policy_full[all_finite--3-omit-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -3
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'omit'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -3
data = [array([[[0.63696169, 0.02831967],
[0.57152983, 0.40455184],
[0.75772885, 0.47998792]],
[[0.26...0.84131728]],
[[0.42268722, 0.3577952 ],
[0.0520213 , 0.46004514],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -3, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'omit'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967C0C120
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f4955420a30>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a1240>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49550a1a20>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
_________________________________________________________________________ test_axis_nan_policy_full[all_finite--3-raise-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -3
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'raise'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -3
data = [array([[[0.63696169, 0.02831967],
[0.57152983, 0.40455184],
[0.75772885, 0.47998792]],
[[0.26...0.84131728]],
[[0.42268722, 0.3577952 ],
[0.0520213 , 0.46004514],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -3, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'raise'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEFE60
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f4954399300>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a0b80>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49550a0dc0>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
_______________________________________________________________________ test_axis_nan_policy_full[all_finite--2-propagate-jarque_bera-args17-kwds17-1-2-False-None] ________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -2
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'propagate'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -2
data = [array([[[0.63696169, 0.02831967],
[0.26978671, 0.12428328],
[0.04097352, 0.67062441],
[0.0165..., 0.92910422],
[0.97262881, 0.0660825 ],
[0.88993556, 0.84131728],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -2, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'propagate'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEFA00
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f49545b4b80>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a0430>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49550a0310>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
__________________________________________________________________________ test_axis_nan_policy_full[all_finite--2-omit-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -2
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'omit'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -2
data = [array([[[0.63696169, 0.02831967],
[0.26978671, 0.12428328],
[0.04097352, 0.67062441],
[0.0165..., 0.92910422],
[0.97262881, 0.0660825 ],
[0.88993556, 0.84131728],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -2, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'omit'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEF680
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f49544ddc90>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a1480>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b5bd0>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
_________________________________________________________________________ test_axis_nan_policy_full[all_finite--2-raise-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -2
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'raise'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -2
data = [array([[[0.63696169, 0.02831967],
[0.26978671, 0.12428328],
[0.04097352, 0.67062441],
[0.0165..., 0.92910422],
[0.97262881, 0.0660825 ],
[0.88993556, 0.84131728],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -2, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'raise'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEF3E0
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f49554e11e0>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a2d40>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b7880>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
_______________________________________________________________________ test_axis_nan_policy_full[all_finite--1-propagate-jarque_bera-args17-kwds17-1-2-False-None] ________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -1
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'propagate'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -1
data = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -1, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'propagate'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEF140
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f4954787130>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a2680>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b6f80>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
__________________________________________________________________________ test_axis_nan_policy_full[all_finite--1-omit-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -1
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'omit'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -1
data = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -1, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'omit'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEEEA0
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f49542d4820>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a2cb0>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b6830>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
_________________________________________________________________________ test_axis_nan_policy_full[all_finite--1-raise-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -1
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'raise'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -1
data = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -1, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'raise'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEEC00
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f495440b760>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a2440>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b5cf0>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
________________________________________________________________________ test_axis_nan_policy_full[all_finite-0-propagate-jarque_bera-args17-kwds17-1-2-False-None] ________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = 0
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'propagate'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = 0
data = [array([[[0.63696169, 0.02831967],
[0.57152983, 0.40455184],
[0.75772885, 0.47998792]],
[[0.26...0.84131728]],
[[0.42268722, 0.3577952 ],
[0.0520213 , 0.46004514],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': 0, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'propagate'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEE960
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f495414d150>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a25f0>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b5750>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
__________________________________________________________________________ test_axis_nan_policy_full[all_finite-0-omit-jarque_bera-args17-kwds17-1-2-False-None] ___________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = 0
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'omit'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = 0
data = [array([[[0.63696169, 0.02831967],
[0.57152983, 0.40455184],
[0.75772885, 0.47998792]],
[[0.26...0.84131728]],
[[0.42268722, 0.3577952 ],
[0.0520213 , 0.46004514],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': 0, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'omit'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEE6C0
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f495471efe0>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a2830>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b5090>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
__________________________________________________________________________ test_axis_nan_policy_full[all_finite-0-raise-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = 0
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'raise'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = 0
data = [array([[[0.63696169, 0.02831967],
[0.57152983, 0.40455184],
[0.75772885, 0.47998792]],
[[0.26...0.84131728]],
[[0.42268722, 0.3577952 ],
[0.0520213 , 0.46004514],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': 0, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'raise'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEE7A0
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f4947f11e70>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49551b4d30>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b4a60>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
________________________________________________________________________ test_axis_nan_policy_full[all_finite-1-propagate-jarque_bera-args17-kwds17-1-2-False-None] ________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = 1
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'propagate'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = 1
data = [array([[[0.63696169, 0.02831967],
[0.26978671, 0.12428328],
[0.04097352, 0.67062441],
[0.0165..., 0.92910422],
[0.97262881, 0.0660825 ],
[0.88993556, 0.84131728],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': 1, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'propagate'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEE5E0
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f49547b43a0>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49551b49d0>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49552c3e20>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
__________________________________________________________________________ test_axis_nan_policy_full[all_finite-1-omit-jarque_bera-args17-kwds17-1-2-False-None] ___________________________________________________________________________
...
__________________________________________________________________________ test_axis_nan_policy_full[all_finite-1-raise-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
...
________________________________________________________________________ test_axis_nan_policy_full[all_finite-2-propagate-jarque_bera-args17-kwds17-1-2-False-None] ________________________________________________________________________
...
__________________________________________________________________________ test_axis_nan_policy_full[all_finite-2-omit-jarque_bera-args17-kwds17-1-2-False-None] ___________________________________________________________________________
...
__________________________________________________________________________ test_axis_nan_policy_full[all_finite-2-raise-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
...
========================================================================================================= short test summary info ==========================================================================================================
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--3-propagate-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--3-omit-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--3-raise-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--2-propagate-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--2-omit-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--2-raise-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--1-propagate-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--1-omit-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--1-raise-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-0-propagate-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-0-omit-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-0-raise-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-1-propagate-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-1-omit-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-1-raise-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-2-propagate-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-2-omit-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-2-raise-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
```
### SciPy/NumPy/Python version and system information
```shell
1.11.1 1.25.0 sys.version_info(major=3, minor=10, micro=6, releaselevel='final', serial=0)
lapack_armpl_info:
NOT AVAILABLE
lapack_mkl_info:
libraries = ['mkl_rt', 'pthread', 'mkl_rt']
library_dirs = ['/opt/intel/oneapi/mkl/latest/lib/intel64']
define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)]
include_dirs = ['/opt/intel/oneapi/mkl/latest/include']
lapack_opt_info:
libraries = ['mkl_rt', 'pthread', 'mkl_rt']
library_dirs = ['/opt/intel/oneapi/mkl/latest/lib/intel64']
define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)]
include_dirs = ['/opt/intel/oneapi/mkl/latest/include']
blas_armpl_info:
NOT AVAILABLE
blas_mkl_info:
libraries = ['mkl_rt', 'pthread', 'mkl_rt']
library_dirs = ['/opt/intel/oneapi/mkl/latest/lib/intel64']
define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)]
include_dirs = ['/opt/intel/oneapi/mkl/latest/include']
blas_opt_info:
libraries = ['mkl_rt', 'pthread', 'mkl_rt']
library_dirs = ['/opt/intel/oneapi/mkl/latest/lib/intel64']
define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)]
include_dirs = ['/opt/intel/oneapi/mkl/latest/include']
Supported SIMD extensions in this NumPy install:
baseline = SSE,SSE2,SSE3,SSSE3,SSE41,POPCNT,SSE42,AVX,F16C,FMA3,AVX2,AVX512F,AVX512CD,AVX512_SKX,AVX512_CLX,AVX512_CNL,AVX512_ICL
found =
not found = AVX512_KNL,AVX512_KNM
```
|
1.0
|
TST: test_axis_nan_policy_full failure - ### Describe your issue.
test_axis_nan_policy_full tests fails.
I built SciPy 1.11.1 manually with Intel OneAPI MKL as BLAS on Ubuntu 22.04. CPU with AVX512 support (NumPy/Scipy built with --cpu-baseline="native" flag).
Similar to resolved earlier https://github.com/scipy/scipy/issues/16337
Failure log for **test_axis_nan_policy_full[all_finite-2-propagate-jarque_bera-args17-kwds17-1-2-False-None]**, **test_axis_nan_policy_full[all_finite-2-omit-jarque_bera-args17-kwds17-1-2-False-None]**, **test_axis_nan_policy_full[all_finite-2-raise-jarque_bera-args17-kwds17-1-2-False-None]** tests is omitted as there was error "There was an error creating your issue: body is too long (maximum is 65536 characters)."
### Reproducing Code Example
```python
import scipy as sp
sp.test('full')
```
### Error message
```shell
_______________________________________________________________________ test_axis_nan_policy_full[all_finite--3-propagate-jarque_bera-args17-kwds17-1-2-False-None] ________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -3
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'propagate'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -3
data = [array([[[0.63696169, 0.02831967],
[0.57152983, 0.40455184],
[0.75772885, 0.47998792]],
[[0.26...0.84131728]],
[[0.42268722, 0.3577952 ],
[0.0520213 , 0.46004514],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -3, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'propagate'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F49E9D3D540
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f4954df24d0>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a2f80>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49550a2710>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
__________________________________________________________________________ test_axis_nan_policy_full[all_finite--3-omit-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -3
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'omit'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -3
data = [array([[[0.63696169, 0.02831967],
[0.57152983, 0.40455184],
[0.75772885, 0.47998792]],
[[0.26...0.84131728]],
[[0.42268722, 0.3577952 ],
[0.0520213 , 0.46004514],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -3, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'omit'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967C0C120
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f4955420a30>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a1240>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49550a1a20>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
_________________________________________________________________________ test_axis_nan_policy_full[all_finite--3-raise-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -3
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'raise'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -3
data = [array([[[0.63696169, 0.02831967],
[0.57152983, 0.40455184],
[0.75772885, 0.47998792]],
[[0.26...0.84131728]],
[[0.42268722, 0.3577952 ],
[0.0520213 , 0.46004514],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -3, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'raise'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEFE60
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f4954399300>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a0b80>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49550a0dc0>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
_______________________________________________________________________ test_axis_nan_policy_full[all_finite--2-propagate-jarque_bera-args17-kwds17-1-2-False-None] ________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -2
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'propagate'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -2
data = [array([[[0.63696169, 0.02831967],
[0.26978671, 0.12428328],
[0.04097352, 0.67062441],
[0.0165..., 0.92910422],
[0.97262881, 0.0660825 ],
[0.88993556, 0.84131728],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -2, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'propagate'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEFA00
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f49545b4b80>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a0430>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49550a0310>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
__________________________________________________________________________ test_axis_nan_policy_full[all_finite--2-omit-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -2
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'omit'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -2
data = [array([[[0.63696169, 0.02831967],
[0.26978671, 0.12428328],
[0.04097352, 0.67062441],
[0.0165..., 0.92910422],
[0.97262881, 0.0660825 ],
[0.88993556, 0.84131728],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -2, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'omit'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEF680
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f49544ddc90>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a1480>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b5bd0>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
_________________________________________________________________________ test_axis_nan_policy_full[all_finite--2-raise-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -2
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'raise'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -2
data = [array([[[0.63696169, 0.02831967],
[0.26978671, 0.12428328],
[0.04097352, 0.67062441],
[0.0165..., 0.92910422],
[0.97262881, 0.0660825 ],
[0.88993556, 0.84131728],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -2, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'raise'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEF3E0
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f49554e11e0>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a2d40>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b7880>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
_______________________________________________________________________ test_axis_nan_policy_full[all_finite--1-propagate-jarque_bera-args17-kwds17-1-2-False-None] ________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -1
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'propagate'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -1
data = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -1, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'propagate'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEF140
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f4954787130>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a2680>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b6f80>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
__________________________________________________________________________ test_axis_nan_policy_full[all_finite--1-omit-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -1
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'omit'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -1
data = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -1, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'omit'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEEEA0
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f49542d4820>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a2cb0>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b6830>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
_________________________________________________________________________ test_axis_nan_policy_full[all_finite--1-raise-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = -1
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'raise'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = -1
data = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': -1, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'raise'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEEC00
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f495440b760>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a2440>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b5cf0>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
________________________________________________________________________ test_axis_nan_policy_full[all_finite-0-propagate-jarque_bera-args17-kwds17-1-2-False-None] ________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = 0
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'propagate'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = 0
data = [array([[[0.63696169, 0.02831967],
[0.57152983, 0.40455184],
[0.75772885, 0.47998792]],
[[0.26...0.84131728]],
[[0.42268722, 0.3577952 ],
[0.0520213 , 0.46004514],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': 0, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'propagate'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEE960
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f495414d150>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a25f0>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b5750>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
__________________________________________________________________________ test_axis_nan_policy_full[all_finite-0-omit-jarque_bera-args17-kwds17-1-2-False-None] ___________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = 0
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'omit'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = 0
data = [array([[[0.63696169, 0.02831967],
[0.57152983, 0.40455184],
[0.75772885, 0.47998792]],
[[0.26...0.84131728]],
[[0.42268722, 0.3577952 ],
[0.0520213 , 0.46004514],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': 0, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'omit'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEE6C0
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f495471efe0>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49550a2830>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b5090>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
__________________________________________________________________________ test_axis_nan_policy_full[all_finite-0-raise-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = 0
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'raise'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = 0
data = [array([[[0.63696169, 0.02831967],
[0.57152983, 0.40455184],
[0.75772885, 0.47998792]],
[[0.26...0.84131728]],
[[0.42268722, 0.3577952 ],
[0.0520213 , 0.46004514],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': 0, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'raise'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEE7A0
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f4947f11e70>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49551b4d30>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49551b4a60>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
________________________________________________________________________ test_axis_nan_policy_full[all_finite-1-propagate-jarque_bera-args17-kwds17-1-2-False-None] ________________________________________________________________________
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:208: in test_axis_nan_policy_full
_axis_nan_policy_test(hypotest, args, kwds, n_samples, n_outputs, paired,
args = ()
axis = 1
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
kwds = {}
n_outputs = 2
n_samples = 1
nan_policy = 'propagate'
paired = False
unpacker = None
/usr/local/lib/python3.10/dist-packages/scipy/stats/tests/test_axis_nan_policy.py:328: in _axis_nan_policy_test
assert_allclose(res[1], pvalues, rtol=1e-15)
_ = 0.0
args = ()
axis = 1
data = [array([[[0.63696169, 0.02831967],
[0.26978671, 0.12428328],
[0.04097352, 0.67062441],
[0.0165..., 0.92910422],
[0.97262881, 0.0660825 ],
[0.88993556, 0.84131728],
[0.82237383, 0.06669001]]])]
data1d = [array([0.47998792, 0.23237292, 0.80188058, 0.92353016, 0.26613027,
0.53893441, 0.44275283, 0.93101732, 0.04051...5, 0.02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001])]
data_b = [array([[[0.63696169, 0.26978671, 0.04097352, 0.01652764, 0.81327024,
0.91275558, 0.60663578, 0.72949656, 0.5....02836537, 0.71921977, 0.01599173, 0.757951 ,
0.51275872, 0.92910422, 0.0660825 , 0.84131728, 0.06669001]]])]
data_gen_kwds = {'all_nans': False, 'axis': 1, 'n_repetitions': 3, 'n_samples': 1, ...}
data_generator = 'all_finite'
hypotest = <function jarque_bera at 0x7f49e50fa710>
i = (2, 1)
inherent_size = 2
kwds = {}
n_outputs = 2
n_repetitions = 3
n_samples = 1
nan_policy = 'propagate'
output_shape = [3, 2]
paired = False
pvalues = array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]])
res = SignificanceResult(statistic=array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]]), pvalue=array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
res1d = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
res1db = SignificanceResult(statistic=1.7796245292853161, pvalue=0.41073285459397746)
rng = Generator(PCG64) at 0x7F4967DEE5E0
statistics = array([[1.87063092, 0.69307536],
[1.56279006, 1.15076854],
[3.68466423, 1.77962453]])
sup = <numpy.testing._private.utils.suppress_warnings object at 0x7f49547b43a0>
unpacker = <function _axis_nan_policy_test.<locals>.unpacker at 0x7f49551b49d0>
/usr/lib/python3.10/contextlib.py:79: in inner
return func(*args, **kwds)
E AssertionError:
E Not equal to tolerance rtol=1e-15, atol=0
E
E Mismatched elements: 1 / 6 (16.7%)
E Max absolute difference: 1.66533454e-16
E Max relative difference: 1.05103253e-15
E x: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
E y: array([[0.392462, 0.707132],
E [0.457767, 0.562489],
E [0.158447, 0.410733]])
args = (<function assert_allclose.<locals>.compare at 0x7f49552c3e20>, array([[0.39246204, 0.70713217],
[0.45776697, 0...48, 0.41073285]]), array([[0.39246204, 0.70713217],
[0.45776697, 0.56248868],
[0.15844748, 0.41073285]]))
func = <function assert_array_compare at 0x7f49eb42e200>
kwds = {'equal_nan': True, 'err_msg': '', 'header': 'Not equal to tolerance rtol=1e-15, atol=0', 'verbose': True}
self = <contextlib._GeneratorContextManager object at 0x7f49ebae1e10>
__________________________________________________________________________ test_axis_nan_policy_full[all_finite-1-omit-jarque_bera-args17-kwds17-1-2-False-None] ___________________________________________________________________________
...
__________________________________________________________________________ test_axis_nan_policy_full[all_finite-1-raise-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
...
________________________________________________________________________ test_axis_nan_policy_full[all_finite-2-propagate-jarque_bera-args17-kwds17-1-2-False-None] ________________________________________________________________________
...
__________________________________________________________________________ test_axis_nan_policy_full[all_finite-2-omit-jarque_bera-args17-kwds17-1-2-False-None] ___________________________________________________________________________
...
__________________________________________________________________________ test_axis_nan_policy_full[all_finite-2-raise-jarque_bera-args17-kwds17-1-2-False-None] __________________________________________________________________________
...
========================================================================================================= short test summary info ==========================================================================================================
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--3-propagate-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--3-omit-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--3-raise-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--2-propagate-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--2-omit-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--2-raise-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--1-propagate-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--1-omit-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite--1-raise-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-0-propagate-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-0-omit-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-0-raise-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-1-propagate-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-1-omit-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-1-raise-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-2-propagate-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-2-omit-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
FAILED stats/tests/test_axis_nan_policy.py::test_axis_nan_policy_full[all_finite-2-raise-jarque_bera-args17-kwds17-1-2-False-None] - AssertionError:
```
### SciPy/NumPy/Python version and system information
```shell
1.11.1 1.25.0 sys.version_info(major=3, minor=10, micro=6, releaselevel='final', serial=0)
lapack_armpl_info:
NOT AVAILABLE
lapack_mkl_info:
libraries = ['mkl_rt', 'pthread', 'mkl_rt']
library_dirs = ['/opt/intel/oneapi/mkl/latest/lib/intel64']
define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)]
include_dirs = ['/opt/intel/oneapi/mkl/latest/include']
lapack_opt_info:
libraries = ['mkl_rt', 'pthread', 'mkl_rt']
library_dirs = ['/opt/intel/oneapi/mkl/latest/lib/intel64']
define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)]
include_dirs = ['/opt/intel/oneapi/mkl/latest/include']
blas_armpl_info:
NOT AVAILABLE
blas_mkl_info:
libraries = ['mkl_rt', 'pthread', 'mkl_rt']
library_dirs = ['/opt/intel/oneapi/mkl/latest/lib/intel64']
define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)]
include_dirs = ['/opt/intel/oneapi/mkl/latest/include']
blas_opt_info:
libraries = ['mkl_rt', 'pthread', 'mkl_rt']
library_dirs = ['/opt/intel/oneapi/mkl/latest/lib/intel64']
define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)]
include_dirs = ['/opt/intel/oneapi/mkl/latest/include']
Supported SIMD extensions in this NumPy install:
baseline = SSE,SSE2,SSE3,SSSE3,SSE41,POPCNT,SSE42,AVX,F16C,FMA3,AVX2,AVX512F,AVX512CD,AVX512_SKX,AVX512_CLX,AVX512_CNL,AVX512_ICL
found =
not found = AVX512_KNL,AVX512_KNM
```
|
defect
|
tst test axis nan policy full failure describe your issue test axis nan policy full tests fails i built scipy manually with intel oneapi mkl as blas on ubuntu cpu with support numpy scipy built with cpu baseline native flag similar to resolved earlier failure log for test axis nan policy full test axis nan policy full test axis nan policy full tests is omitted as there was error there was an error creating your issue body is too long maximum is characters reproducing code example python import scipy as sp sp test full error message shell test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy propagate paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy propagate output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy omit paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy omit output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy raise paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy raise output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy propagate paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy propagate output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy omit paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy omit output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy raise paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy raise output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy propagate paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy propagate output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy omit paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy omit output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy raise paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy raise output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy propagate paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy propagate output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy omit paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy omit output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy raise paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy raise output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full usr local lib dist packages scipy stats tests test axis nan policy py in test axis nan policy full axis nan policy test hypotest args kwds n samples n outputs paired args axis data generator all finite hypotest kwds n outputs n samples nan policy propagate paired false unpacker none usr local lib dist packages scipy stats tests test axis nan policy py in axis nan policy test assert allclose res pvalues rtol args axis data array data b array data gen kwds all nans false axis n repetitions n samples data generator all finite hypotest i inherent size kwds n outputs n repetitions n samples nan policy propagate output shape paired false pvalues array res significanceresult statistic array pvalue array significanceresult statistic pvalue significanceresult statistic pvalue rng generator at statistics array sup unpacker unpacker at usr lib contextlib py in inner return func args kwds e assertionerror e not equal to tolerance rtol atol e e mismatched elements e max absolute difference e max relative difference e x array e e e y array e e args compare at array array func kwds equal nan true err msg header not equal to tolerance rtol atol verbose true self test axis nan policy full test axis nan policy full test axis nan policy full test axis nan policy full test axis nan policy full short test summary info failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror failed stats tests test axis nan policy py test axis nan policy full assertionerror scipy numpy python version and system information shell sys version info major minor micro releaselevel final serial lapack armpl info not available lapack mkl info libraries library dirs define macros include dirs lapack opt info libraries library dirs define macros include dirs blas armpl info not available blas mkl info libraries library dirs define macros include dirs blas opt info libraries library dirs define macros include dirs supported simd extensions in this numpy install baseline sse popcnt avx skx clx cnl icl found not found knl knm
| 1
|
78,978
| 9,813,204,551
|
IssuesEvent
|
2019-06-13 07:22:45
|
semi-technologies/weaviate
|
https://api.github.com/repos/semi-technologies/weaviate
|
opened
|
Suggestion for UX timeout behavior
|
API design & UX Bad UX graphql
|
Take the following example:
```graphql
{
Local {
Get {
Things {
Place(where: {
path: ["name"],
operator: Equal,
valueString: "Shell"
}) {
name
InCity {
... on City {
name
}
}
}
}
}
}
}
```
On a large graph with many places, this is going to timeout. Which is understandable. But the UX is not very nice.
```json
{
"data": {
"Local": {
"Get": {
"Things": {
"Place": null
}
}
}
},
"errors": [
{
"locations": [
{
"column": 9,
"line": 5
}
],
"message": "Janusgraph.LocalGetClass: &errors.errorString{s:\"could not process fetch query: executing the query failed: server error: {\\\"message\\\":\\\"Script evaluation exceeded the configured 'scriptEvaluationTimeout' threshold of 30000 ms or evaluation was otherwise cancelled directly for request [g.V().has(\\\\\\\"kind\\\\\\\", \\\\\\\"thing\\\\\\\").has(\\\\\\\"classId\\\\\\\", \\\\\\\"class_3\\\\\\\").union(has(\\\\\\\"prop_f\\\\\\\", eq(\\\\\\\"Shell qqqq\\\\\\\"))).union(optional(outE(\\\\\\\"prop_1f\\\\\\\").inV().has(\\\\\\\"classId\\\\\\\", \\\\\\\"class_2\\\\\\\"))).limit(100).path().by(valueMap())]\\\",\\\"Exception-Class\\\":\\\"java.util.concurrent.TimeoutException\\\",\\\"exceptions\\\":[\\\"java.util.concurrent.TimeoutException\\\"],\\\"stackTrace\\\":\\\"java.util.concurrent.TimeoutException: Script evaluation exceeded the configured 'scriptEvaluationTimeout' threshold of 30000 ms or evaluation was otherwise cancelled directly for request [g.V().has(\\\\\\\"kind\\\\\\\", \\\\\\\"thing\\\\\\\").has(\\\\\\\"classId\\\\\\\", \\\\\\\"class_3\\\\\\\").union(has(\\\\\\\"prop_f\\\\\\\", eq(\\\\\\\"Shell qqqq\\\\\\\"))).union(optional(outE(\\\\\\\"prop_1f\\\\\\\").inV().has(\\\\\\\"classId\\\\\\\", \\\\\\\"class_2\\\\\\\"))).limit(100).path().by(valueMap())]\\\\n\\\\tat org.apache.tinkerpop.gremlin.groovy.engine.GremlinExecutor.lambda$eval$1(GremlinExecutor.java:310)\\\\n\\\\tat io.netty.util.concurrent.PromiseTask$RunnableAdapter.call(PromiseTask.java:38)\\\\n\\\\tat io.netty.util.concurrent.ScheduledFutureTask.run(ScheduledFutureTask.java:120)\\\\n\\\\tat io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:399)\\\\n\\\\tat io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:464)\\\\n\\\\tat io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)\\\\n\\\\tat java.lang.Thread.run(Thread.java:748)\\\\n\\\"}\"}",
"path": [
"Local",
"Get",
"Things",
"Place"
]
}
]
}
```
## Proposed Solution
Can we do the following things to avoid the bad user experience;
1. Have people set a manual `timeOut` in seconds (see below)?
2. If a timeout occurs, send out _the results it has found_ combined with an informative error message with suggestions. For example;
```json
{
"data": {
"Local": {
"Get": {
"Things": {
"Place": [
{
"InCity": [
{
"name": "Berlin"
}
],
"name": "Shell"
},
{
"InCity": [
{
"name": "Langenselbold"
}
],
"name": "Shell"
}
]
}
}
}
},
"errors": [
{
"locations": [
{
"column": 9,
"line": 5
}
],
"message": "We have been unable to fulfill the complete request, we are showing what we have found. Maybe your query was to large, or there are no results in the Graph. Try to limit, increase the timeout or more accureatly pinpoint your query. You can also use our Spark integration",
"path": [
"Local",
"Get",
"Things",
"Place"
]
}
]
}
```
#### Timeout suggestion:
```graphql
{
Local {
Get {
Things {
Place(where: {
path: ["name"],
operator: Equal,
valueString: "Shell"
},
timeOut: 120) { # <=== Note the example
name
InCity {
... on City {
name
}
}
}
}
}
}
}
```
|
1.0
|
Suggestion for UX timeout behavior - Take the following example:
```graphql
{
Local {
Get {
Things {
Place(where: {
path: ["name"],
operator: Equal,
valueString: "Shell"
}) {
name
InCity {
... on City {
name
}
}
}
}
}
}
}
```
On a large graph with many places, this is going to timeout. Which is understandable. But the UX is not very nice.
```json
{
"data": {
"Local": {
"Get": {
"Things": {
"Place": null
}
}
}
},
"errors": [
{
"locations": [
{
"column": 9,
"line": 5
}
],
"message": "Janusgraph.LocalGetClass: &errors.errorString{s:\"could not process fetch query: executing the query failed: server error: {\\\"message\\\":\\\"Script evaluation exceeded the configured 'scriptEvaluationTimeout' threshold of 30000 ms or evaluation was otherwise cancelled directly for request [g.V().has(\\\\\\\"kind\\\\\\\", \\\\\\\"thing\\\\\\\").has(\\\\\\\"classId\\\\\\\", \\\\\\\"class_3\\\\\\\").union(has(\\\\\\\"prop_f\\\\\\\", eq(\\\\\\\"Shell qqqq\\\\\\\"))).union(optional(outE(\\\\\\\"prop_1f\\\\\\\").inV().has(\\\\\\\"classId\\\\\\\", \\\\\\\"class_2\\\\\\\"))).limit(100).path().by(valueMap())]\\\",\\\"Exception-Class\\\":\\\"java.util.concurrent.TimeoutException\\\",\\\"exceptions\\\":[\\\"java.util.concurrent.TimeoutException\\\"],\\\"stackTrace\\\":\\\"java.util.concurrent.TimeoutException: Script evaluation exceeded the configured 'scriptEvaluationTimeout' threshold of 30000 ms or evaluation was otherwise cancelled directly for request [g.V().has(\\\\\\\"kind\\\\\\\", \\\\\\\"thing\\\\\\\").has(\\\\\\\"classId\\\\\\\", \\\\\\\"class_3\\\\\\\").union(has(\\\\\\\"prop_f\\\\\\\", eq(\\\\\\\"Shell qqqq\\\\\\\"))).union(optional(outE(\\\\\\\"prop_1f\\\\\\\").inV().has(\\\\\\\"classId\\\\\\\", \\\\\\\"class_2\\\\\\\"))).limit(100).path().by(valueMap())]\\\\n\\\\tat org.apache.tinkerpop.gremlin.groovy.engine.GremlinExecutor.lambda$eval$1(GremlinExecutor.java:310)\\\\n\\\\tat io.netty.util.concurrent.PromiseTask$RunnableAdapter.call(PromiseTask.java:38)\\\\n\\\\tat io.netty.util.concurrent.ScheduledFutureTask.run(ScheduledFutureTask.java:120)\\\\n\\\\tat io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:399)\\\\n\\\\tat io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:464)\\\\n\\\\tat io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)\\\\n\\\\tat java.lang.Thread.run(Thread.java:748)\\\\n\\\"}\"}",
"path": [
"Local",
"Get",
"Things",
"Place"
]
}
]
}
```
## Proposed Solution
Can we do the following things to avoid the bad user experience;
1. Have people set a manual `timeOut` in seconds (see below)?
2. If a timeout occurs, send out _the results it has found_ combined with an informative error message with suggestions. For example;
```json
{
"data": {
"Local": {
"Get": {
"Things": {
"Place": [
{
"InCity": [
{
"name": "Berlin"
}
],
"name": "Shell"
},
{
"InCity": [
{
"name": "Langenselbold"
}
],
"name": "Shell"
}
]
}
}
}
},
"errors": [
{
"locations": [
{
"column": 9,
"line": 5
}
],
"message": "We have been unable to fulfill the complete request, we are showing what we have found. Maybe your query was to large, or there are no results in the Graph. Try to limit, increase the timeout or more accureatly pinpoint your query. You can also use our Spark integration",
"path": [
"Local",
"Get",
"Things",
"Place"
]
}
]
}
```
#### Timeout suggestion:
```graphql
{
Local {
Get {
Things {
Place(where: {
path: ["name"],
operator: Equal,
valueString: "Shell"
},
timeOut: 120) { # <=== Note the example
name
InCity {
... on City {
name
}
}
}
}
}
}
}
```
|
non_defect
|
suggestion for ux timeout behavior take the following example graphql local get things place where path operator equal valuestring shell name incity on city name on a large graph with many places this is going to timeout which is understandable but the ux is not very nice json data local get things place null errors locations column line message janusgraph localgetclass errors errorstring s could not process fetch query executing the query failed server error message script evaluation exceeded the configured scriptevaluationtimeout threshold of ms or evaluation was otherwise cancelled directly for request exception class java util concurrent timeoutexception exceptions stacktrace java util concurrent timeoutexception script evaluation exceeded the configured scriptevaluationtimeout threshold of ms or evaluation was otherwise cancelled directly for request n tat org apache tinkerpop gremlin groovy engine gremlinexecutor lambda eval gremlinexecutor java n tat io netty util concurrent promisetask runnableadapter call promisetask java n tat io netty util concurrent scheduledfuturetask run scheduledfuturetask java n tat io netty util concurrent singlethreadeventexecutor runalltasks singlethreadeventexecutor java n tat io netty channel nio nioeventloop run nioeventloop java n tat io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java n tat java lang thread run thread java n path local get things place proposed solution can we do the following things to avoid the bad user experience have people set a manual timeout in seconds see below if a timeout occurs send out the results it has found combined with an informative error message with suggestions for example json data local get things place incity name berlin name shell incity name langenselbold name shell errors locations column line message we have been unable to fulfill the complete request we are showing what we have found maybe your query was to large or there are no results in the graph try to limit increase the timeout or more accureatly pinpoint your query you can also use our spark integration path local get things place timeout suggestion graphql local get things place where path operator equal valuestring shell timeout note the example name incity on city name
| 0
|
161,103
| 20,120,401,081
|
IssuesEvent
|
2022-02-08 01:15:22
|
arohablue/BlockDockServer
|
https://api.github.com/repos/arohablue/BlockDockServer
|
closed
|
CVE-2020-36187 (High) detected in jackson-databind-2.8.8.jar - autoclosed
|
security vulnerability
|
## CVE-2020-36187 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /BlockDockServer/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.8/bf88c7b27e95cbadce4e7c316a56c3efffda8026/jackson-databind-2.8.8.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-actuator-1.5.4.RELEASE.jar (Root Library)
- spring-boot-actuator-1.5.4.RELEASE.jar
- :x: **jackson-databind-2.8.8.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.SharedPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36187>CVE-2020-36187</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2997">https://github.com/FasterXML/jackson-databind/issues/2997</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-36187 (High) detected in jackson-databind-2.8.8.jar - autoclosed - ## CVE-2020-36187 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /BlockDockServer/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.8/bf88c7b27e95cbadce4e7c316a56c3efffda8026/jackson-databind-2.8.8.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-actuator-1.5.4.RELEASE.jar (Root Library)
- spring-boot-actuator-1.5.4.RELEASE.jar
- :x: **jackson-databind-2.8.8.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.SharedPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36187>CVE-2020-36187</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2997">https://github.com/FasterXML/jackson-databind/issues/2997</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file blockdockserver build gradle path to vulnerable library root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter actuator release jar root library spring boot actuator release jar x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp dbcp datasources sharedpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
| 0
|
127,971
| 18,024,784,734
|
IssuesEvent
|
2021-09-17 02:03:20
|
AlexRogalskiy/qiitos
|
https://api.github.com/repos/AlexRogalskiy/qiitos
|
opened
|
CVE-2020-7598 (Medium) detected in minimist-0.1.0.tgz
|
security vulnerability
|
## CVE-2020-7598 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimist-0.1.0.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.1.0.tgz">https://registry.npmjs.org/minimist/-/minimist-0.1.0.tgz</a></p>
<p>Path to dependency file: qiitos/package.json</p>
<p>Path to vulnerable library: qiitos/node_modules/strong-log-transformer/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- lerna-2.11.0.tgz (Root Library)
- strong-log-transformer-1.0.6.tgz
- :x: **minimist-0.1.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload.
<p>Publish Date: 2020-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94">https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94</a></p>
<p>Release Date: 2020-03-11</p>
<p>Fix Resolution: minimist - 0.2.1,1.2.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7598 (Medium) detected in minimist-0.1.0.tgz - ## CVE-2020-7598 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimist-0.1.0.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.1.0.tgz">https://registry.npmjs.org/minimist/-/minimist-0.1.0.tgz</a></p>
<p>Path to dependency file: qiitos/package.json</p>
<p>Path to vulnerable library: qiitos/node_modules/strong-log-transformer/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- lerna-2.11.0.tgz (Root Library)
- strong-log-transformer-1.0.6.tgz
- :x: **minimist-0.1.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload.
<p>Publish Date: 2020-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94">https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94</a></p>
<p>Release Date: 2020-03-11</p>
<p>Fix Resolution: minimist - 0.2.1,1.2.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in minimist tgz cve medium severity vulnerability vulnerable library minimist tgz parse argument options library home page a href path to dependency file qiitos package json path to vulnerable library qiitos node modules strong log transformer node modules minimist package json dependency hierarchy lerna tgz root library strong log transformer tgz x minimist tgz vulnerable library found in base branch master vulnerability details minimist before could be tricked into adding or modifying properties of object prototype using a constructor or proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution minimist step up your open source security game with whitesource
| 0
|
366,479
| 25,587,088,719
|
IssuesEvent
|
2022-12-01 10:08:17
|
r5py/r5py
|
https://api.github.com/repos/r5py/r5py
|
closed
|
[Upstream issue]: openjdk package in conda-forge overwrites libraries installed by other packages, breaks gdal
|
documentation installation reminder upstream
|
_This is a follow-up issue on #156 - which will be closed when our side of the bug has been fixed with PR #189_
Packaging for OpenJDK is currently broken in conda-forge, an upstream issue (https://github.com/conda-forge/openjdk-feedstock/issues/86) and pull request (https://github.com/conda-forge/openjdk-feedstock/pull/100) exist and, hopefully, will be closed and merged soon.
This issue is to remind us to then re-enable openjdk as a dependency for the mamba/conda environment files, and adjusting the installation instructions accordingly.
|
1.0
|
[Upstream issue]: openjdk package in conda-forge overwrites libraries installed by other packages, breaks gdal - _This is a follow-up issue on #156 - which will be closed when our side of the bug has been fixed with PR #189_
Packaging for OpenJDK is currently broken in conda-forge, an upstream issue (https://github.com/conda-forge/openjdk-feedstock/issues/86) and pull request (https://github.com/conda-forge/openjdk-feedstock/pull/100) exist and, hopefully, will be closed and merged soon.
This issue is to remind us to then re-enable openjdk as a dependency for the mamba/conda environment files, and adjusting the installation instructions accordingly.
|
non_defect
|
openjdk package in conda forge overwrites libraries installed by other packages breaks gdal this is a follow up issue on which will be closed when our side of the bug has been fixed with pr packaging for openjdk is currently broken in conda forge an upstream issue and pull request exist and hopefully will be closed and merged soon this issue is to remind us to then re enable openjdk as a dependency for the mamba conda environment files and adjusting the installation instructions accordingly
| 0
|
90,948
| 18,278,028,470
|
IssuesEvent
|
2021-10-04 21:27:00
|
microsoft/pxt-arcade
|
https://api.github.com/repos/microsoft/pxt-arcade
|
closed
|
Hour of Code 2021 Landing Page Design
|
ui design graphic design hour of code landing page css
|
For hour of code 2021, we may want to make styling and layout changes to our hour of code landing page on Arcade. This feature request will track those changes.
Here's the current task list:
- [ ] Update the navbar to be our standardized nav bar (just purple) (@shakao can you help with this?)
- [ ] Launch skillmap activity in a new tab
- [ ] Add code.org Trademark information
- [ ] Simplify footer
- [ ] Add CTA buttons to educator page sections
- [ ] Add how long, how difficult, etc data on educator page about activity
- [ ] Refine all page copy and educator page copy
- - [ ] First section copy
- - [ ] Social section (by footer) copy
- - [ ] Educator Guide Teacher Prep section
- [ ] Come up with metadata one-line description and icons for social sharing (@Jaqster for description, I can do the icons)
- [ ] Get display font approved (@riknoll can you help with this?) OR if not make display font Segoe
- [ ] Swap FontAwesome CDN with the fontawesome already include in Arcade (@riknoll can you help with this?)
- [ ] Update all tracking, tick events and usabila to be for HoC 2021 instead of 2020 (@jwunderl can you help with this?)
- [x] Make a script to animate the carousel
- [x] Educator's page theming
- [x] Header background image
- [x] Main activity image
- [ ] Global Game Jam image
- [ ] Carousel images + links
- - [x] Hour of Code 2020
- - [x] Beginner Skillmap
- - [ ] Rockstar Skillmap
- - [ ] Jungle Skillmap
- - [ ] Minecraft HoC
- - [x] Micro:bit
- [ ] Hooking up all the links
- - [x] Educator Guide
- - [x] Forest activity
- - [ ] Game Jam Landing Page
https://arcade.makecode.com/hour-of-code
|
1.0
|
Hour of Code 2021 Landing Page Design - For hour of code 2021, we may want to make styling and layout changes to our hour of code landing page on Arcade. This feature request will track those changes.
Here's the current task list:
- [ ] Update the navbar to be our standardized nav bar (just purple) (@shakao can you help with this?)
- [ ] Launch skillmap activity in a new tab
- [ ] Add code.org Trademark information
- [ ] Simplify footer
- [ ] Add CTA buttons to educator page sections
- [ ] Add how long, how difficult, etc data on educator page about activity
- [ ] Refine all page copy and educator page copy
- - [ ] First section copy
- - [ ] Social section (by footer) copy
- - [ ] Educator Guide Teacher Prep section
- [ ] Come up with metadata one-line description and icons for social sharing (@Jaqster for description, I can do the icons)
- [ ] Get display font approved (@riknoll can you help with this?) OR if not make display font Segoe
- [ ] Swap FontAwesome CDN with the fontawesome already include in Arcade (@riknoll can you help with this?)
- [ ] Update all tracking, tick events and usabila to be for HoC 2021 instead of 2020 (@jwunderl can you help with this?)
- [x] Make a script to animate the carousel
- [x] Educator's page theming
- [x] Header background image
- [x] Main activity image
- [ ] Global Game Jam image
- [ ] Carousel images + links
- - [x] Hour of Code 2020
- - [x] Beginner Skillmap
- - [ ] Rockstar Skillmap
- - [ ] Jungle Skillmap
- - [ ] Minecraft HoC
- - [x] Micro:bit
- [ ] Hooking up all the links
- - [x] Educator Guide
- - [x] Forest activity
- - [ ] Game Jam Landing Page
https://arcade.makecode.com/hour-of-code
|
non_defect
|
hour of code landing page design for hour of code we may want to make styling and layout changes to our hour of code landing page on arcade this feature request will track those changes here s the current task list update the navbar to be our standardized nav bar just purple shakao can you help with this launch skillmap activity in a new tab add code org trademark information simplify footer add cta buttons to educator page sections add how long how difficult etc data on educator page about activity refine all page copy and educator page copy first section copy social section by footer copy educator guide teacher prep section come up with metadata one line description and icons for social sharing jaqster for description i can do the icons get display font approved riknoll can you help with this or if not make display font segoe swap fontawesome cdn with the fontawesome already include in arcade riknoll can you help with this update all tracking tick events and usabila to be for hoc instead of jwunderl can you help with this make a script to animate the carousel educator s page theming header background image main activity image global game jam image carousel images links hour of code beginner skillmap rockstar skillmap jungle skillmap minecraft hoc micro bit hooking up all the links educator guide forest activity game jam landing page
| 0
|
31,307
| 6,496,338,934
|
IssuesEvent
|
2017-08-22 09:39:51
|
buildo/react-components
|
https://api.github.com/repos/buildo/react-components
|
closed
|
Don't use `t` from tcomb-react
|
defect waiting for merge
|
## description
We should be using `t` from `tcomb` directly as it has a type definition file for TS.
|
1.0
|
Don't use `t` from tcomb-react - ## description
We should be using `t` from `tcomb` directly as it has a type definition file for TS.
|
defect
|
don t use t from tcomb react description we should be using t from tcomb directly as it has a type definition file for ts
| 1
|
70,638
| 23,269,586,734
|
IssuesEvent
|
2022-08-04 21:09:17
|
primefaces/primeng
|
https://api.github.com/repos/primefaces/primeng
|
opened
|
Tab navigation doesn’t work on the second DynamicDialogModule
|
defect
|
### Describe the bug
I've have 2 dynamic dialogs. On the first one the tab navigation works, but when I open the second dynamic dialog from the first one, tab navigation stops working on the second dynamic dialog form. The demo:
https://angular-ivy-mpvxm5.stackblitz.io
https://stackblitz.com/edit/angular-ivy-mpvxm5?file=src/app/components/users/user-edit/user-edit.component.html
### Environment
os: macOS montereu, ver 12.5
node: v16.14.2
npm: 8.5.0
### Reproducer
https://angular-ivy-mpvxm5.stackblitz.io
### Angular version
13.0.0 and 14.0.0
### PrimeNG version
13.0.2
### Build / Runtime
Angular CLI App
### Language
TypeScript
### Node version (for AoT issues node --version)
16.14.2
### Browser(s)
Chrome 103
### Steps to reproduce the behavior
_No response_
### Expected behavior
_No response_
|
1.0
|
Tab navigation doesn’t work on the second DynamicDialogModule - ### Describe the bug
I've have 2 dynamic dialogs. On the first one the tab navigation works, but when I open the second dynamic dialog from the first one, tab navigation stops working on the second dynamic dialog form. The demo:
https://angular-ivy-mpvxm5.stackblitz.io
https://stackblitz.com/edit/angular-ivy-mpvxm5?file=src/app/components/users/user-edit/user-edit.component.html
### Environment
os: macOS montereu, ver 12.5
node: v16.14.2
npm: 8.5.0
### Reproducer
https://angular-ivy-mpvxm5.stackblitz.io
### Angular version
13.0.0 and 14.0.0
### PrimeNG version
13.0.2
### Build / Runtime
Angular CLI App
### Language
TypeScript
### Node version (for AoT issues node --version)
16.14.2
### Browser(s)
Chrome 103
### Steps to reproduce the behavior
_No response_
### Expected behavior
_No response_
|
defect
|
tab navigation doesn’t work on the second dynamicdialogmodule describe the bug i ve have dynamic dialogs on the first one the tab navigation works but when i open the second dynamic dialog from the first one tab navigation stops working on the second dynamic dialog form the demo environment os macos montereu ver node npm reproducer angular version and primeng version build runtime angular cli app language typescript node version for aot issues node version browser s chrome steps to reproduce the behavior no response expected behavior no response
| 1
|
387,253
| 26,717,518,638
|
IssuesEvent
|
2023-01-28 18:11:57
|
soyvieira/site-portfolio
|
https://api.github.com/repos/soyvieira/site-portfolio
|
closed
|
Melhorar nosso readme
|
documentation
|
# Descrição:
Melhorar nosso arquivo readme do projeto.
<h5>Teste título</h5>
|
1.0
|
Melhorar nosso readme - # Descrição:
Melhorar nosso arquivo readme do projeto.
<h5>Teste título</h5>
|
non_defect
|
melhorar nosso readme descrição melhorar nosso arquivo readme do projeto teste título
| 0
|
69,206
| 22,274,771,295
|
IssuesEvent
|
2022-06-10 15:31:35
|
vector-im/element-ios
|
https://api.github.com/repos/vector-im/element-ios
|
closed
|
Reply rendering bug for displaynames containing "<"
|
T-Defect A-Timeline Z-Papercuts S-Minor O-Occasional Z-WTF
|
### Steps to reproduce
1. Set your roomnick to a username containing "<"
2. Send a message in the room
3. Get a reply to that message
### Outcome
#### What did you expect?
Reply is rendered correctly
#### What happened instead?
- the username is cut off beginning with the "<"
- The whole original message is displayed in green
- You can click on the cited original message and get to the users "profile" as if you clicked on their username

### Your phone model
iPhone8
### Operating system version
iOS 15.3
### Application version
Element 1.7.0
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Reply rendering bug for displaynames containing "<" - ### Steps to reproduce
1. Set your roomnick to a username containing "<"
2. Send a message in the room
3. Get a reply to that message
### Outcome
#### What did you expect?
Reply is rendered correctly
#### What happened instead?
- the username is cut off beginning with the "<"
- The whole original message is displayed in green
- You can click on the cited original message and get to the users "profile" as if you clicked on their username

### Your phone model
iPhone8
### Operating system version
iOS 15.3
### Application version
Element 1.7.0
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
reply rendering bug for displaynames containing steps to reproduce set your roomnick to a username containing send a message in the room get a reply to that message outcome what did you expect reply is rendered correctly what happened instead the username is cut off beginning with the the whole original message is displayed in green you can click on the cited original message and get to the users profile as if you clicked on their username your phone model operating system version ios application version element homeserver no response will you send logs no
| 1
|
74,448
| 25,131,197,468
|
IssuesEvent
|
2022-11-09 15:15:40
|
SasView/sasmodels
|
https://api.github.com/repos/SasView/sasmodels
|
opened
|
Error in documentation for gel_fit model
|
SasModels Documentation defect
|
User ZhouQ noticed the following about the gel_fit model:
>From your manual, the scattered intensity is provided by

>However, from Shibayama’s paper, the Lorentzian part is given by

>Considering that when D = 2, the Lorentzian part will be reduced to the Ornstein-Zernicke equation, as you and Shibayama mentioned, I believe the bottom form is correct. But I am unsure whether it is a typo in the manual or an incorrect expression that may affect the fitting results (a bug?).
Examination of the source code shows:
```
// Lorentzian Term
////////////////////////double a(x[i]*x[i]*zeta*zeta);
double lorentzian_term = square(q*cor_length);
lorentzian_term = 1.0 + ((fractal_dim + 1.0)/3.0)*lorentzian_term;
lorentzian_term = pow(lorentzian_term, fractal_dim/2.0 );
```
which corresponds with the lower (Shibayama) expression so the formatting of the expression in the model doc string needs revising.
|
1.0
|
Error in documentation for gel_fit model - User ZhouQ noticed the following about the gel_fit model:
>From your manual, the scattered intensity is provided by

>However, from Shibayama’s paper, the Lorentzian part is given by

>Considering that when D = 2, the Lorentzian part will be reduced to the Ornstein-Zernicke equation, as you and Shibayama mentioned, I believe the bottom form is correct. But I am unsure whether it is a typo in the manual or an incorrect expression that may affect the fitting results (a bug?).
Examination of the source code shows:
```
// Lorentzian Term
////////////////////////double a(x[i]*x[i]*zeta*zeta);
double lorentzian_term = square(q*cor_length);
lorentzian_term = 1.0 + ((fractal_dim + 1.0)/3.0)*lorentzian_term;
lorentzian_term = pow(lorentzian_term, fractal_dim/2.0 );
```
which corresponds with the lower (Shibayama) expression so the formatting of the expression in the model doc string needs revising.
|
defect
|
error in documentation for gel fit model user zhouq noticed the following about the gel fit model from your manual the scattered intensity is provided by however from shibayama’s paper the lorentzian part is given by considering that when d the lorentzian part will be reduced to the ornstein zernicke equation as you and shibayama mentioned i believe the bottom form is correct but i am unsure whether it is a typo in the manual or an incorrect expression that may affect the fitting results a bug examination of the source code shows lorentzian term double a x x zeta zeta double lorentzian term square q cor length lorentzian term fractal dim lorentzian term lorentzian term pow lorentzian term fractal dim which corresponds with the lower shibayama expression so the formatting of the expression in the model doc string needs revising
| 1
|
149,461
| 13,281,764,895
|
IssuesEvent
|
2020-08-23 19:00:33
|
erleben/OpenTissue
|
https://api.github.com/repos/erleben/OpenTissue
|
closed
|
Update README.md with new build system
|
Documentation
|
Add a detailed explanation of how to use Conan, modern CMake, dashboard, azure pipeline etc.
|
1.0
|
Update README.md with new build system - Add a detailed explanation of how to use Conan, modern CMake, dashboard, azure pipeline etc.
|
non_defect
|
update readme md with new build system add a detailed explanation of how to use conan modern cmake dashboard azure pipeline etc
| 0
|
50,877
| 13,187,932,326
|
IssuesEvent
|
2020-08-13 05:04:03
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
[steamshovel] log viewer setting not saved (Trac #1587)
|
Migrated from Trac combo core defect
|
In window->configuration, there is a setting to "automatically open log viewer"
This doesn't get saved in the session.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1587">https://code.icecube.wisc.edu/ticket/1587</a>, reported by david.schultz and owned by hdembinski</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:26",
"description": "In window->configuration, there is a setting to \"automatically open log viewer\"\n\nThis doesn't get saved in the session.",
"reporter": "david.schultz",
"cc": "cweaver",
"resolution": "fixed",
"_ts": "1550067086520250",
"component": "combo core",
"summary": "[steamshovel] log viewer setting not saved",
"priority": "critical",
"keywords": "",
"time": "2016-03-16T02:29:29",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[steamshovel] log viewer setting not saved (Trac #1587) - In window->configuration, there is a setting to "automatically open log viewer"
This doesn't get saved in the session.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1587">https://code.icecube.wisc.edu/ticket/1587</a>, reported by david.schultz and owned by hdembinski</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:26",
"description": "In window->configuration, there is a setting to \"automatically open log viewer\"\n\nThis doesn't get saved in the session.",
"reporter": "david.schultz",
"cc": "cweaver",
"resolution": "fixed",
"_ts": "1550067086520250",
"component": "combo core",
"summary": "[steamshovel] log viewer setting not saved",
"priority": "critical",
"keywords": "",
"time": "2016-03-16T02:29:29",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
</p>
</details>
|
defect
|
log viewer setting not saved trac in window configuration there is a setting to automatically open log viewer this doesn t get saved in the session migrated from json status closed changetime description in window configuration there is a setting to automatically open log viewer n nthis doesn t get saved in the session reporter david schultz cc cweaver resolution fixed ts component combo core summary log viewer setting not saved priority critical keywords time milestone owner hdembinski type defect
| 1
|
662
| 3,130,906,920
|
IssuesEvent
|
2015-09-09 12:15:09
|
e-government-ua/i
|
https://api.github.com/repos/e-government-ua/i
|
closed
|
Доработать систему вспомогательной технической авторизации.
|
active hi priority In process of testing test
|
Доработать класс:
\i\wf-base\src\main\java\org\activity\rest\security\AccessKeyAuthFilter.java
и связанные с ним классы:
\i\wf-base\src\main\java\org\activity\rest\security\*.java
Чтоб по мимо пары параметров(для авторизации кроме фиксированного логина/пароля энкоденного в хеадере):
sAccessKey (ключ, по которому подтягивается сверяемый контент)
nID_Subject (сам контент, с которым происходит сверка)
Поддерживалась комбинация:
sAccessKey (ключ, по которому подтягивается сверяемый контент)
sAccessContract=Request (параметр, говорящий, что сверяемым контентом является сама строка запроса)
При которой вместо сверки совпадения контента nID_Subject с контентом, полученным по ключу(sAccessKey), в качестве контента будет выступать строка самого текущего запроса.
НО при этом, из строки реквеста важно будет вырезать сам параметр "&sAccessKey=(тот код, что пришел)"
Например:
"/wf-region/service/setPaymentStatus_TaskActiviti?sID_Order=TaskActiviti_105123&sID_PaymentSystem=Liqpay&sData=&sAccessContract=Request"
(хотя по логике там еще должен был быть параметр &sAccessKey=5d1184c7-2d84-4931-a8f5-1032b50ed699)
ВАЖНО:
- Если по ключу sAccessKey вытягивается контент, но он не совпадет со "сверяемым" выдавать эксепшин с сообщением:
"Contract data by sAccessKey is not equal !!! (sAccessContract=Request, sAccessKey=5d1184c7-2d84-4931-a8f5-1032b50ed699)"
(а в логах выводить один и другой контент, для сравнения, чтоб можно было дебажить)
- Если по ключу sAccessKey вообще не найден контента, выдавать эксепшин с сообщением:
"Content not found by sAccessKey !!! (sAccessContract=Request, sAccessKey=5d1184c7-2d84-4931-a8f5-1032b50ed699)"
П.С.: Это логическое продолжение развития предыдущей таски: https://github.com/e-government-ua/i/issues/502
|
1.0
|
Доработать систему вспомогательной технической авторизации. - Доработать класс:
\i\wf-base\src\main\java\org\activity\rest\security\AccessKeyAuthFilter.java
и связанные с ним классы:
\i\wf-base\src\main\java\org\activity\rest\security\*.java
Чтоб по мимо пары параметров(для авторизации кроме фиксированного логина/пароля энкоденного в хеадере):
sAccessKey (ключ, по которому подтягивается сверяемый контент)
nID_Subject (сам контент, с которым происходит сверка)
Поддерживалась комбинация:
sAccessKey (ключ, по которому подтягивается сверяемый контент)
sAccessContract=Request (параметр, говорящий, что сверяемым контентом является сама строка запроса)
При которой вместо сверки совпадения контента nID_Subject с контентом, полученным по ключу(sAccessKey), в качестве контента будет выступать строка самого текущего запроса.
НО при этом, из строки реквеста важно будет вырезать сам параметр "&sAccessKey=(тот код, что пришел)"
Например:
"/wf-region/service/setPaymentStatus_TaskActiviti?sID_Order=TaskActiviti_105123&sID_PaymentSystem=Liqpay&sData=&sAccessContract=Request"
(хотя по логике там еще должен был быть параметр &sAccessKey=5d1184c7-2d84-4931-a8f5-1032b50ed699)
ВАЖНО:
- Если по ключу sAccessKey вытягивается контент, но он не совпадет со "сверяемым" выдавать эксепшин с сообщением:
"Contract data by sAccessKey is not equal !!! (sAccessContract=Request, sAccessKey=5d1184c7-2d84-4931-a8f5-1032b50ed699)"
(а в логах выводить один и другой контент, для сравнения, чтоб можно было дебажить)
- Если по ключу sAccessKey вообще не найден контента, выдавать эксепшин с сообщением:
"Content not found by sAccessKey !!! (sAccessContract=Request, sAccessKey=5d1184c7-2d84-4931-a8f5-1032b50ed699)"
П.С.: Это логическое продолжение развития предыдущей таски: https://github.com/e-government-ua/i/issues/502
|
non_defect
|
доработать систему вспомогательной технической авторизации доработать класс i wf base src main java org activity rest security accesskeyauthfilter java и связанные с ним классы i wf base src main java org activity rest security java чтоб по мимо пары параметров для авторизации кроме фиксированного логина пароля энкоденного в хеадере saccesskey ключ по которому подтягивается сверяемый контент nid subject сам контент с которым происходит сверка поддерживалась комбинация saccesskey ключ по которому подтягивается сверяемый контент saccesscontract request параметр говорящий что сверяемым контентом является сама строка запроса при которой вместо сверки совпадения контента nid subject с контентом полученным по ключу saccesskey в качестве контента будет выступать строка самого текущего запроса но при этом из строки реквеста важно будет вырезать сам параметр saccesskey тот код что пришел например wf region service setpaymentstatus taskactiviti sid order taskactiviti sid paymentsystem liqpay sdata saccesscontract request хотя по логике там еще должен был быть параметр saccesskey важно если по ключу saccesskey вытягивается контент но он не совпадет со сверяемым выдавать эксепшин с сообщением contract data by saccesskey is not equal saccesscontract request saccesskey а в логах выводить один и другой контент для сравнения чтоб можно было дебажить если по ключу saccesskey вообще не найден контента выдавать эксепшин с сообщением content not found by saccesskey saccesscontract request saccesskey п с это логическое продолжение развития предыдущей таски
| 0
|
4,054
| 2,610,086,645
|
IssuesEvent
|
2015-02-26 18:26:19
|
chrsmith/dsdsdaadf
|
https://api.github.com/repos/chrsmith/dsdsdaadf
|
opened
|
深圳除祛痤疮收费
|
auto-migrated Priority-Medium Type-Defect
|
```
深圳除祛痤疮收费【深圳韩方科颜全国热线400-869-1818,24小时
QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘��
�——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方�
��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健
康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业��
�疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘�
��。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:13
|
1.0
|
深圳除祛痤疮收费 - ```
深圳除祛痤疮收费【深圳韩方科颜全国热线400-869-1818,24小时
QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘��
�——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方�
��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健
康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业��
�疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘�
��。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:13
|
defect
|
深圳除祛痤疮收费 深圳除祛痤疮收费【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘�� �——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方� ��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健 康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业�� �疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘� ��。 original issue reported on code google com by szft com on may at
| 1
|
152,854
| 19,697,876,182
|
IssuesEvent
|
2022-01-12 14:00:03
|
rarkins/spdx-maven-plugin-demo
|
https://api.github.com/repos/rarkins/spdx-maven-plugin-demo
|
opened
|
CVE-2020-35728 (High) detected in jackson-databind-2.9.10.4.jar
|
security vulnerability
|
## CVE-2020-35728 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.4/jackson-databind-2.9.10.4.jar</p>
<p>
Dependency Hierarchy:
- spdx-tools-2.2.1.jar (Root Library)
- :x: **jackson-databind-2.9.10.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rarkins/spdx-maven-plugin-demo/commit/7b3620983ed40ac208d55490d9322df547b7716a">7b3620983ed40ac208d55490d9322df547b7716a</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.oracle.wls.shaded.org.apache.xalan.lib.sql.JNDIConnectionPool (aka embedded Xalan in org.glassfish.web/javax.servlet.jsp.jstl).
<p>Publish Date: 2020-12-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35728>CVE-2020-35728</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35728">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35728</a></p>
<p>Release Date: 2020-12-27</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-35728 (High) detected in jackson-databind-2.9.10.4.jar - ## CVE-2020-35728 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.4/jackson-databind-2.9.10.4.jar</p>
<p>
Dependency Hierarchy:
- spdx-tools-2.2.1.jar (Root Library)
- :x: **jackson-databind-2.9.10.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rarkins/spdx-maven-plugin-demo/commit/7b3620983ed40ac208d55490d9322df547b7716a">7b3620983ed40ac208d55490d9322df547b7716a</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.oracle.wls.shaded.org.apache.xalan.lib.sql.JNDIConnectionPool (aka embedded Xalan in org.glassfish.web/javax.servlet.jsp.jstl).
<p>Publish Date: 2020-12-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35728>CVE-2020-35728</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35728">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-35728</a></p>
<p>Release Date: 2020-12-27</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spdx tools jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com oracle wls shaded org apache xalan lib sql jndiconnectionpool aka embedded xalan in org glassfish web javax servlet jsp jstl publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
| 0
|
33,968
| 12,232,632,046
|
IssuesEvent
|
2020-05-04 10:02:56
|
Neufund/platform-frontend
|
https://api.github.com/repos/Neufund/platform-frontend
|
closed
|
Always logout the user first when accessing the app via the login/activation link
|
Security Sprint
|
**Background**
Everytime the login/acitivation link is used we need to logout the user. No need for extra logic or any special mechanics
**Designs**
N/A
**Acceptance criteria**
- [ ] Given I'm a user entering with an Activation/Login link I will always need to sign or enter my password
- [ ] Given I'm a user entering without a link I should be able to login easily
**Technical notes**
**Testing notes**
|
True
|
Always logout the user first when accessing the app via the login/activation link - **Background**
Everytime the login/acitivation link is used we need to logout the user. No need for extra logic or any special mechanics
**Designs**
N/A
**Acceptance criteria**
- [ ] Given I'm a user entering with an Activation/Login link I will always need to sign or enter my password
- [ ] Given I'm a user entering without a link I should be able to login easily
**Technical notes**
**Testing notes**
|
non_defect
|
always logout the user first when accessing the app via the login activation link background everytime the login acitivation link is used we need to logout the user no need for extra logic or any special mechanics designs n a acceptance criteria given i m a user entering with an activation login link i will always need to sign or enter my password given i m a user entering without a link i should be able to login easily technical notes testing notes
| 0
|
74,455
| 14,261,032,587
|
IssuesEvent
|
2020-11-20 10:42:10
|
LoliKingdom/Toxic-Garbage-Island
|
https://api.github.com/repos/LoliKingdom/Toxic-Garbage-Island
|
opened
|
Add appropriate types to lacking biomes
|
bug codebase medium_priority
|
- The Aurorian
- The Midnight
- Glacidus
- NC Wasteland
|
1.0
|
Add appropriate types to lacking biomes - - The Aurorian
- The Midnight
- Glacidus
- NC Wasteland
|
non_defect
|
add appropriate types to lacking biomes the aurorian the midnight glacidus nc wasteland
| 0
|
52,620
| 13,224,866,501
|
IssuesEvent
|
2020-08-17 20:00:34
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
View -> Take Snapshot bad location (Trac #69)
|
Migrated from Trac defect glshovel
|
Let's try File -> Export Image
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/69">https://code.icecube.wisc.edu/projects/icecube/ticket/69</a>, reported by troyand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-11T03:35:42",
"_ts": "1194752142000000",
"description": "Let's try File -> Export Image",
"reporter": "troy",
"cc": "",
"resolution": "duplicate",
"time": "2007-06-21T11:31:10",
"component": "glshovel",
"summary": "View -> Take Snapshot bad location",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
View -> Take Snapshot bad location (Trac #69) - Let's try File -> Export Image
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/69">https://code.icecube.wisc.edu/projects/icecube/ticket/69</a>, reported by troyand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-11T03:35:42",
"_ts": "1194752142000000",
"description": "Let's try File -> Export Image",
"reporter": "troy",
"cc": "",
"resolution": "duplicate",
"time": "2007-06-21T11:31:10",
"component": "glshovel",
"summary": "View -> Take Snapshot bad location",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
|
defect
|
view take snapshot bad location trac let s try file export image migrated from json status closed changetime ts description let s try file export image reporter troy cc resolution duplicate time component glshovel summary view take snapshot bad location priority normal keywords milestone owner troy type defect
| 1
|
127,592
| 5,037,027,598
|
IssuesEvent
|
2016-12-17 12:15:21
|
Moridrin/ssv-frontend-members
|
https://api.github.com/repos/Moridrin/ssv-frontend-members
|
closed
|
Registration Page not showing
|
Minor Bug PRIORITY
|
Even with the "Custom Register Page" option unchecked it tries to use the custom registration page.
|
1.0
|
Registration Page not showing - Even with the "Custom Register Page" option unchecked it tries to use the custom registration page.
|
non_defect
|
registration page not showing even with the custom register page option unchecked it tries to use the custom registration page
| 0
|
4,266
| 2,610,090,479
|
IssuesEvent
|
2015-02-26 18:27:23
|
chrsmith/dsdsdaadf
|
https://api.github.com/repos/chrsmith/dsdsdaadf
|
opened
|
深圳痘痘怎么治疗
|
auto-migrated Priority-Medium Type-Defect
|
```
深圳痘痘怎么治疗【深圳韩方科颜全国热线400-869-1818,24小时
QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘��
�——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方�
��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健
康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业��
�疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘�
��。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:42
|
1.0
|
深圳痘痘怎么治疗 - ```
深圳痘痘怎么治疗【深圳韩方科颜全国热线400-869-1818,24小时
QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘��
�——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方�
��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健
康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业��
�疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘�
��。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:42
|
defect
|
深圳痘痘怎么治疗 深圳痘痘怎么治疗【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘�� �——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方� ��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健 康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业�� �疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘� ��。 original issue reported on code google com by szft com on may at
| 1
|
55,805
| 14,692,662,328
|
IssuesEvent
|
2021-01-03 03:45:12
|
Cockatrice/Cockatrice
|
https://api.github.com/repos/Cockatrice/Cockatrice
|
closed
|
Cards dragged to graveyard don't go to top with tokens
|
App - Protocol / API Defect - Game Rules Compliance High Priority
|
If I make a selection group consisting of one normal object, and one object that is destroyed as soon as it leaves the play area,, and drag that selection to the discard pile (graveyard), the normal object will not go on top of the pile as expected.
Instead, it will be second from the top.
(Note the order: `{card object, token}`)
----
Presumably the drop action is assuming that the graveyard increases by 2, not 1 and inserts the card using indexes based on this assumption.
|
1.0
|
Cards dragged to graveyard don't go to top with tokens - If I make a selection group consisting of one normal object, and one object that is destroyed as soon as it leaves the play area,, and drag that selection to the discard pile (graveyard), the normal object will not go on top of the pile as expected.
Instead, it will be second from the top.
(Note the order: `{card object, token}`)
----
Presumably the drop action is assuming that the graveyard increases by 2, not 1 and inserts the card using indexes based on this assumption.
|
defect
|
cards dragged to graveyard don t go to top with tokens if i make a selection group consisting of one normal object and one object that is destroyed as soon as it leaves the play area and drag that selection to the discard pile graveyard the normal object will not go on top of the pile as expected instead it will be second from the top note the order card object token presumably the drop action is assuming that the graveyard increases by not and inserts the card using indexes based on this assumption
| 1
|
123,175
| 4,858,331,401
|
IssuesEvent
|
2016-11-13 02:40:56
|
chaorace/cqui
|
https://api.github.com/repos/chaorace/cqui
|
closed
|
Improved resources dim even if the improvement is unrelated to the resource
|
bug low priority
|
Example: Building a farm on marble dims the marble resource icon even though it's not the proper improvement for utilizing the resource. The more correct behavior is for the resource icon to only dim if an appropriate improvement is present, as opposed to any. See discussion in #54 for details
|
1.0
|
Improved resources dim even if the improvement is unrelated to the resource - Example: Building a farm on marble dims the marble resource icon even though it's not the proper improvement for utilizing the resource. The more correct behavior is for the resource icon to only dim if an appropriate improvement is present, as opposed to any. See discussion in #54 for details
|
non_defect
|
improved resources dim even if the improvement is unrelated to the resource example building a farm on marble dims the marble resource icon even though it s not the proper improvement for utilizing the resource the more correct behavior is for the resource icon to only dim if an appropriate improvement is present as opposed to any see discussion in for details
| 0
|
20,890
| 3,427,707,354
|
IssuesEvent
|
2015-12-10 03:47:04
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
Use of 'super' in dart2js output breaks in Android browser
|
area-dart2js Priority-Medium triaged Type-Defect
|
Approximately 80% of all android web users use the Android browser (not the Chrome for Android browser). See http://developer.android.com/about/dashboards/index.html
dart2js emits code that uses the word 'super'. This has caused at least two problems.
First:
The minifier that's part of the new MVC4 bundling stuff fails with "run-time warning JS1028: Expected identifier or string: super" while trying to minify the dart output below:
$$._PendingSendPortFinder = {"":
["ports?", "_visited"],
super: "_MessageTraverser", //<--- this line specifically
visitSendPort$1: function(port) {
if (typeof port === 'object' && port !== null && !!port.is$_BufferingSendPort && port.get$_port() == null)
$.add$1(this.ports, port.get$_futurePort());
See this thread for more: https://groups.google.com/a/dartlang.org/forum/?fromgroups#!topic/misc/wVQ6Mi1dWtU
(luckily in this case, MVC4 was updated and fixed the issue)
Second:
I get a syntax error on line 8 and this is what the generated code (from hi.dart in sample code) looks like:
function Isolate() {}
init();
var $$ = {};
var $ = Isolate.$isolateProperties;
$$.ExceptionImplementation = {"":
["_msg"],
super: "Object",
toString$0: function() {
var t1 = this._msg;
return t1 == null ? 'Exception' : 'Exception: ' + $.S(t1);
},
is$Exception: true
A patch was submitted (note: the ++ and -- are reversed here) which claims to fix the issue:
--- lib/dart2js/lib/compiler/implementation/js_backend/emitter.dart 2012-08-27 18:47:33.152927918 -0700
+++ lib/dart2js/lib/compiler/implementation/js_backend/emitter.dart~ 2012-08-27 16:30:30.881870773 -0700
@@ -166,7 +166,7 @@
if (Object.prototype.hasOwnProperty.call(collectedClasses, cls)) {
var desc = collectedClasses[cls];
$isolatePropertiesName[cls] = $defineClassName(cls, desc[''], desc);
- if (desc['superduper'] !== "") $pendingClassesName[cls] = desc['superduper'];
+ if (desc['super'] !== "") $pendingClassesName[cls] = desc['super'];
}
}
var pendingClasses = $pendingClassesName;
@@ -197,7 +197,7 @@
hosOwnProperty instead. \*/'''
var hasOwnProperty = Object.prototype.hasOwnProperty;
for (var member in prototype) {
- if (member == '' || member == 'superduper') continue;
+ if (member == '' || member == 'super') continue;
if (hasOwnProperty.call(prototype, member)) {
newPrototype[member] = prototype[member];
}
@@ -567,7 +567,7 @@
// TODO(floitsch): the emitInstanceMember should simply always emit a ',\n'.
// That does currently not work because the native classes have a different
// syntax.
- buffer.add(' superduper: "$superName"');
+ buffer.add(' super: "$superName"');
emitInstanceMembers(classElement, buffer, true);
buffer.add('\n};\n\n');
}
@@ -746,7 +746,7 @@
boundClosureBuffer.add("""
$classesCollector.$mangledName = {'':
['self', 'target'],
- 'superduper': '$superName',
+ 'super': '$superName',
""");
// Now add the methods on the closure class. The instance method does not
// have the correct name. Since [addParameterStubs] use the name to create
See this thread for more: https://groups.google.com/a/dartlang.org/forum/#!topic/misc/hj8iL89-Wdg
|
1.0
|
Use of 'super' in dart2js output breaks in Android browser - Approximately 80% of all android web users use the Android browser (not the Chrome for Android browser). See http://developer.android.com/about/dashboards/index.html
dart2js emits code that uses the word 'super'. This has caused at least two problems.
First:
The minifier that's part of the new MVC4 bundling stuff fails with "run-time warning JS1028: Expected identifier or string: super" while trying to minify the dart output below:
$$._PendingSendPortFinder = {"":
["ports?", "_visited"],
super: "_MessageTraverser", //<--- this line specifically
visitSendPort$1: function(port) {
if (typeof port === 'object' && port !== null && !!port.is$_BufferingSendPort && port.get$_port() == null)
$.add$1(this.ports, port.get$_futurePort());
See this thread for more: https://groups.google.com/a/dartlang.org/forum/?fromgroups#!topic/misc/wVQ6Mi1dWtU
(luckily in this case, MVC4 was updated and fixed the issue)
Second:
I get a syntax error on line 8 and this is what the generated code (from hi.dart in sample code) looks like:
function Isolate() {}
init();
var $$ = {};
var $ = Isolate.$isolateProperties;
$$.ExceptionImplementation = {"":
["_msg"],
super: "Object",
toString$0: function() {
var t1 = this._msg;
return t1 == null ? 'Exception' : 'Exception: ' + $.S(t1);
},
is$Exception: true
A patch was submitted (note: the ++ and -- are reversed here) which claims to fix the issue:
--- lib/dart2js/lib/compiler/implementation/js_backend/emitter.dart 2012-08-27 18:47:33.152927918 -0700
+++ lib/dart2js/lib/compiler/implementation/js_backend/emitter.dart~ 2012-08-27 16:30:30.881870773 -0700
@@ -166,7 +166,7 @@
if (Object.prototype.hasOwnProperty.call(collectedClasses, cls)) {
var desc = collectedClasses[cls];
$isolatePropertiesName[cls] = $defineClassName(cls, desc[''], desc);
- if (desc['superduper'] !== "") $pendingClassesName[cls] = desc['superduper'];
+ if (desc['super'] !== "") $pendingClassesName[cls] = desc['super'];
}
}
var pendingClasses = $pendingClassesName;
@@ -197,7 +197,7 @@
hosOwnProperty instead. \*/'''
var hasOwnProperty = Object.prototype.hasOwnProperty;
for (var member in prototype) {
- if (member == '' || member == 'superduper') continue;
+ if (member == '' || member == 'super') continue;
if (hasOwnProperty.call(prototype, member)) {
newPrototype[member] = prototype[member];
}
@@ -567,7 +567,7 @@
// TODO(floitsch): the emitInstanceMember should simply always emit a ',\n'.
// That does currently not work because the native classes have a different
// syntax.
- buffer.add(' superduper: "$superName"');
+ buffer.add(' super: "$superName"');
emitInstanceMembers(classElement, buffer, true);
buffer.add('\n};\n\n');
}
@@ -746,7 +746,7 @@
boundClosureBuffer.add("""
$classesCollector.$mangledName = {'':
['self', 'target'],
- 'superduper': '$superName',
+ 'super': '$superName',
""");
// Now add the methods on the closure class. The instance method does not
// have the correct name. Since [addParameterStubs] use the name to create
See this thread for more: https://groups.google.com/a/dartlang.org/forum/#!topic/misc/hj8iL89-Wdg
|
defect
|
use of super in output breaks in android browser approximately of all android web users use the android browser not the chrome for android browser see emits code that uses the word super this has caused at least two problems first the minifier that s part of the new bundling stuff fails with quot run time warning expected identifier or string super quot while trying to minify the dart output below pendingsendportfinder quot quot nbsp nbsp super quot messagetraverser quot lt this line specifically nbsp visitsendport function port nbsp nbsp if typeof port equals equals equals object amp amp port null amp amp port is bufferingsendport amp amp port get port null nbsp nbsp nbsp nbsp add this ports port get futureport see this thread for more luckily in this case was updated and fixed the issue second i get a syntax error on line and this is what the generated code from hi dart in sample code looks like function isolate init var var isolate isolateproperties exceptionimplementation quot quot nbsp nbsp super quot object quot nbsp tostring function nbsp nbsp var this msg nbsp nbsp return null exception exception s nbsp is exception true a patch was submitted note the and are reversed here which claims to fix the issue lib lib compiler implementation js backend emitter dart lib lib compiler implementation js backend emitter dart nbsp nbsp nbsp nbsp nbsp if object prototype hasownproperty call collectedclasses cls nbsp nbsp nbsp nbsp nbsp nbsp nbsp var desc collectedclasses nbsp nbsp nbsp nbsp nbsp nbsp nbsp isolatepropertiesname defineclassname cls desc desc if desc quot quot pendingclassesname desc if desc quot quot pendingclassesname desc nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp var pendingclasses pendingclassesname nbsp nbsp nbsp nbsp nbsp nbsp nbsp hosownproperty instead nbsp nbsp nbsp nbsp nbsp nbsp nbsp var hasownproperty object prototype hasownproperty nbsp nbsp nbsp nbsp nbsp nbsp nbsp for var member in prototype if member member superduper continue if member member super continue nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp if hasownproperty call prototype member nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp newprototype prototype nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp todo floitsch the emitinstancemember should simply always emit a n nbsp nbsp nbsp nbsp nbsp that does currently not work because the native classes have a different nbsp nbsp nbsp nbsp nbsp syntax buffer add superduper quot supername quot buffer add super quot supername quot nbsp nbsp nbsp nbsp nbsp emitinstancemembers classelement buffer true nbsp nbsp nbsp nbsp nbsp buffer add n n n nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp nbsp boundclosurebuffer add quot quot quot nbsp classescollector mangledname nbsp nbsp superduper supername super supername nbsp quot quot quot nbsp nbsp nbsp nbsp nbsp nbsp nbsp now add the methods on the closure class the instance method does not nbsp nbsp nbsp nbsp nbsp nbsp nbsp have the correct name since use the name to create see this thread for more
| 1
|
227,568
| 18,070,016,660
|
IssuesEvent
|
2021-09-21 01:01:33
|
WordPress/gutenberg
|
https://api.github.com/repos/WordPress/gutenberg
|
closed
|
RaneControl block triggers onChange event during mouseMove event when focused in safari
|
[Feature] UI Components Needs Testing [Status] Duplicate [Package] Components
|
### Description
When working in safari browser (desktop), clicking the arrows within the RangeControl component will correctly trigger an onChange event, however subsequent mouseMove events above said RangeControl arrow elements will continue to trigger onChange events.
It looks like the RangeControl component ultimate renders an input of type number. I've looked and it doesn't seem that this is default behavior on Safari.
### Step-by-step reproduction instructions
1. Add any block that either uses a RangeControl component within its edit function or implements the component within the block's settings inspector. In my testing I've used the `core/spacer` block .
2. In Safari, manipulate the `Height in Pixels` within the Block tab in the settings panel by clicking the arrows to the right of the number within the input.
3. Observe the attributes are updated successfully
4. Move the mouse up and down while remaining in the dimensions of the two arrows.
5. Observe the attributes are continued to change without clicking anything.
### Screenshots, screen recording, code snippet
_No response_
### Environment info
- Safari , Wordpress 5.8
### Please confirm that you have searched existing issues in the repo.
Yes
### Please confirm that you have tested with all plugins deactivated except Gutenberg.
Yes
|
1.0
|
RaneControl block triggers onChange event during mouseMove event when focused in safari - ### Description
When working in safari browser (desktop), clicking the arrows within the RangeControl component will correctly trigger an onChange event, however subsequent mouseMove events above said RangeControl arrow elements will continue to trigger onChange events.
It looks like the RangeControl component ultimate renders an input of type number. I've looked and it doesn't seem that this is default behavior on Safari.
### Step-by-step reproduction instructions
1. Add any block that either uses a RangeControl component within its edit function or implements the component within the block's settings inspector. In my testing I've used the `core/spacer` block .
2. In Safari, manipulate the `Height in Pixels` within the Block tab in the settings panel by clicking the arrows to the right of the number within the input.
3. Observe the attributes are updated successfully
4. Move the mouse up and down while remaining in the dimensions of the two arrows.
5. Observe the attributes are continued to change without clicking anything.
### Screenshots, screen recording, code snippet
_No response_
### Environment info
- Safari , Wordpress 5.8
### Please confirm that you have searched existing issues in the repo.
Yes
### Please confirm that you have tested with all plugins deactivated except Gutenberg.
Yes
|
non_defect
|
ranecontrol block triggers onchange event during mousemove event when focused in safari description when working in safari browser desktop clicking the arrows within the rangecontrol component will correctly trigger an onchange event however subsequent mousemove events above said rangecontrol arrow elements will continue to trigger onchange events it looks like the rangecontrol component ultimate renders an input of type number i ve looked and it doesn t seem that this is default behavior on safari step by step reproduction instructions add any block that either uses a rangecontrol component within its edit function or implements the component within the block s settings inspector in my testing i ve used the core spacer block in safari manipulate the height in pixels within the block tab in the settings panel by clicking the arrows to the right of the number within the input observe the attributes are updated successfully move the mouse up and down while remaining in the dimensions of the two arrows observe the attributes are continued to change without clicking anything screenshots screen recording code snippet no response environment info safari wordpress please confirm that you have searched existing issues in the repo yes please confirm that you have tested with all plugins deactivated except gutenberg yes
| 0
|
40,446
| 10,526,048,329
|
IssuesEvent
|
2019-09-30 16:14:15
|
invertase/react-native-firebase
|
https://api.github.com/repos/invertase/react-native-firebase
|
closed
|
[IOS] Admob - You attempted to use a firebase module that's not installed natively on your iOS project by calling firebase.admob()
|
Blocked: User Follow Up Impact: Build Error Platform: iOS Resolution: Needs Repro Service: AdMob Type: Stale Version: 5.x.x
|
<!---
Hello there you awesome person;
Please note that the issue list of this repo is exclusively for bug reports;
1) For feature requests please visit our [Feature Request Board](https://boards.invertase.io/react-native-firebase).
2) For questions and support please use our Discord chat: https://discord.gg/C9aK28N or Stack Overflow: https://stackoverflow.com/questions/tagged/react-native-firebase
3) If this is a setup issue then please make sure you've correctly followed the setup guides, most setup issues such as 'duplicate dex files', 'default app has not been initialized' etc are all down to an incorrect setup as the guides haven't been correctly followed.
-->
<!-- NOTE: You can change any of the `[ ]` to `[x]` to mark an option(s) as selected -->
<!-- PLEASE DO NOT REMOVE ANY SECTIONS FROM THIS ISSUE TEMPLATE -->
<!-- Leave them as they are even if they're irrelevant to your issue -->
## Issue
<!-- Please describe your issue here --^ and provide as much detail as you can. -->
<!-- Include code snippets that show your usages of the library in the context of your project. -->
<!-- Snippets that also show how and where the library is imported in JS are useful to debug issues relating to importing or methods not found issues -->
I am having an issue on iOS with the adMob pod.
On Android everything works fine.
On iOS I get the following error when trying to call firebase.admob() :
```
Error: You attempted to use a firebase module that's not installed natively on your iOS project by calling firebase.admob().
Ensure you have the required Firebase iOS SDK pod for this module included in your Podfile, in this instance confirm you've added "pod 'Firebase/AdMob'" to your Podfile
See http://invertase.link/ios for full setup instructions.
This error is located at:
in App (at renderApplication.js:34)
in RCTView (at View.js:45)
in View (at AppContainer.js:98)
in RCTView (at View.js:45)
in View (at AppContainer.js:115)
in AppContainer (at renderApplication.js:33)
initialiseNativeModule
native.js:43:20
ModuleBase
ModuleBase.js:26:27
AdMob
index.js:21:4
<unknown>
apps.js:66:74
getModule
apps.js:186:20
App
App.js:98:21
App
[native code]:0
<unknown>
createClassProxy.js:98:23
instantiate
createClassProxy.js:96:6
constructClassInstance
ReactNativeRenderer-dev.js:8864:26
updateClassComponent
ReactNativeRenderer-dev.js:11508:6
performUnitOfWork
ReactNativeRenderer-dev.js:16075:21
workLoop
ReactNativeRenderer-dev.js:16115:41
renderRoot
ReactNativeRenderer-dev.js:16219:15
performWorkOnRoot
ReactNativeRenderer-dev.js:17198:17
performWork
ReactNativeRenderer-dev.js:17099:24
performSyncWork
ReactNativeRenderer-dev.js:17060:14
requestWork
ReactNativeRenderer-dev.js:16925:19
scheduleWork
ReactNativeRenderer-dev.js:16724:16
scheduleRootUpdate
ReactNativeRenderer-dev.js:17413:15
render
ReactNativeRenderer-dev.js:18191:20
renderApplication
renderApplication.js:59:34
run
AppRegistry.js:101:10
runApplication
AppRegistry.js:195:26
__callFunction
MessageQueue.js:366:47
<unknown>
MessageQueue.js:106:26
__guard
MessageQueue.js:314:10
callFunctionReturnFlushedQueue
MessageQueue.js:105:17
callFunctionReturnFlushedQueue
[native code]:0
```
I've tried re-do the initial setup guide on my project. And adding this pod:
`pod 'RNFirebase', :path => '../node_modules/react-native-firebase/ios'`
but I still have this issue
---
## Project Files
<!-- Provide the contents of key project files which will help to debug -->
<!-- For Example: -->
<!-- - iOS: `Podfile` contents. -->
<!-- - Android: `android/build.gradle` contents. -->
<!-- - Android: `android/app/build.gradle` contents. -->
<!-- - Android: `AndroidManifest.xml` contents. -->
<!-- ADD THE CONTENTS OF THE FILES IN THE PROVIDED CODE BLOCKS BELOW -->
### iOS
<details><summary>Click To Expand</summary>
<p>
#### `ios/Podfile`:
- [ ] I'm not using Pods
- [x] I'm using Pods and my Podfile looks like:
```
platform :ios, '9.0'
target 'xxx' do
# Uncomment the next line if you're using Swift or would like to use dynamic frameworks
# use_frameworks!
# Pods for xxx
pod 'React', :path => '../node_modules/react-native'
pod 'yoga', :path => '../node_modules/react-native/ReactCommon/yoga'
pod 'Firebase/Core', '~> 5.15.0'
pod 'Firebase/Database', '~> 5.15.0'
pod 'Firebase/AdMob', '~> 5.15.0'
pod 'RNCAsyncStorage', :path => '../node_modules/@react-native-community/async-storage'
target 'xxx' do
inherit! :search_paths
# Pods for testing
end
target 'xxx' do
inherit! :search_paths
# Pods for testing
end
end
```
`podfile.lock`:
```
PODS:
- Firebase/AdMob (5.15.0):
- Firebase/Core
- Google-Mobile-Ads-SDK (~> 7.37)
- Firebase/Core (5.15.0):
- Firebase/CoreOnly
- FirebaseAnalytics (= 5.4.0)
- Firebase/CoreOnly (5.15.0):
- FirebaseCore (= 5.1.10)
- Firebase/Database (5.15.0):
- Firebase/CoreOnly
- FirebaseDatabase (= 5.0.4)
- FirebaseAnalytics (5.4.0):
- FirebaseCore (~> 5.1)
- FirebaseInstanceID (~> 3.3)
- GoogleAppMeasurement (= 5.4.0)
- GoogleUtilities/AppDelegateSwizzler (~> 5.2)
- GoogleUtilities/MethodSwizzler (~> 5.2)
- GoogleUtilities/Network (~> 5.2)
- "GoogleUtilities/NSData+zlib (~> 5.2)"
- nanopb (~> 0.3)
- FirebaseAuthInterop (1.0.0)
- FirebaseCore (5.1.10):
- GoogleUtilities/Logger (~> 5.2)
- FirebaseDatabase (5.0.4):
- FirebaseAuthInterop (~> 1.0)
- FirebaseCore (~> 5.1)
- leveldb-library (~> 1.18)
- FirebaseInstanceID (3.3.0):
- FirebaseCore (~> 5.1)
- GoogleUtilities/Environment (~> 5.3)
- GoogleUtilities/UserDefaults (~> 5.3)
- Google-Mobile-Ads-SDK (7.41.0)
- GoogleAppMeasurement (5.4.0):
- GoogleUtilities/AppDelegateSwizzler (~> 5.2)
- GoogleUtilities/MethodSwizzler (~> 5.2)
- GoogleUtilities/Network (~> 5.2)
- "GoogleUtilities/NSData+zlib (~> 5.2)"
- nanopb (~> 0.3)
- GoogleUtilities/AppDelegateSwizzler (5.6.0):
- GoogleUtilities/Environment
- GoogleUtilities/Logger
- GoogleUtilities/Network
- GoogleUtilities/Environment (5.6.0)
- GoogleUtilities/Logger (5.6.0):
- GoogleUtilities/Environment
- GoogleUtilities/MethodSwizzler (5.6.0):
- GoogleUtilities/Logger
- GoogleUtilities/Network (5.6.0):
- GoogleUtilities/Logger
- "GoogleUtilities/NSData+zlib"
- GoogleUtilities/Reachability
- "GoogleUtilities/NSData+zlib (5.6.0)"
- GoogleUtilities/Reachability (5.6.0):
- GoogleUtilities/Logger
- GoogleUtilities/UserDefaults (5.6.0):
- GoogleUtilities/Logger
- leveldb-library (1.20)
- nanopb (0.3.901):
- nanopb/decode (= 0.3.901)
- nanopb/encode (= 0.3.901)
- nanopb/decode (0.3.901)
- nanopb/encode (0.3.901)
- React (0.58.6):
- React/Core (= 0.58.6)
- React/Core (0.58.6):
- yoga (= 0.58.6.React)
- RNCAsyncStorage (1.4.0):
- React
- yoga (0.58.6.React)
DEPENDENCIES:
- Firebase/AdMob (~> 5.15.0)
- Firebase/Core (~> 5.15.0)
- Firebase/Database (~> 5.15.0)
- React (from `../node_modules/react-native`)
- "RNCAsyncStorage (from `../node_modules/@react-native-community/async-storage`)"
- yoga (from `../node_modules/react-native/ReactCommon/yoga`)
SPEC REPOS:
https://github.com/cocoapods/specs.git:
- Firebase
- FirebaseAnalytics
- FirebaseAuthInterop
- FirebaseCore
- FirebaseDatabase
- FirebaseInstanceID
- Google-Mobile-Ads-SDK
- GoogleAppMeasurement
- GoogleUtilities
- leveldb-library
- nanopb
EXTERNAL SOURCES:
React:
:path: "../node_modules/react-native"
RNCAsyncStorage:
:path: "../node_modules/@react-native-community/async-storage"
yoga:
:path: "../node_modules/react-native/ReactCommon/yoga"
SPEC CHECKSUMS:
Firebase: 8bb9268bff82374f2cbaaabb143e725743c316ae
FirebaseAnalytics: c06f9d70577d79074214700a71fd5d39de5550fb
FirebaseAuthInterop: 0ffa57668be100582bb7643d4fcb7615496c41fc
FirebaseCore: 35747502d9e8c6ee217385ad04446c7c2aaf9c5c
FirebaseDatabase: 0621689f77528d62b47e1c06ca737c4c19275d1a
FirebaseInstanceID: e2fa4cb35ef5558c200f7f0ad8a53e212215f93e
Google-Mobile-Ads-SDK: 101ce89b94fdf60ec64ba7b6e9b9d812dc765936
GoogleAppMeasurement: 98b71f5e04142793729a5ef23e5b96651ff4b70f
GoogleUtilities: f44b068df02195da10409f415bd8c98439cd70fa
leveldb-library: 08cba283675b7ed2d99629a4bc5fd052cd2bb6a5
nanopb: 2901f78ea1b7b4015c860c2fdd1ea2fee1a18d48
React: 130b87b2d5e2baac646954282cab87be986d98fc
RNCAsyncStorage: b82dc6e5b39a625d70e3b3492bff75c0de94ba71
yoga: 32d7ef1081951e9a35a4c72a7be797598b138a48
PODFILE CHECKSUM: 7c7dc6ad557e143e89cd532086f991b949c19580
COCOAPODS: 1.6.1
```
#### `AppDelegate.m`:
```objc
// N/A
```
</p>
</details>
---
### Android
<details><summary>Click To Expand</summary>
<p>
#### `android/build.gradle`:
```groovy
// N/A
```
#### `android/app/build.gradle`:
```groovy
// N/A
```
#### `android/settings.gradle`:
```groovy
// N/A
```
#### `MainApplication.java`:
```java
// N/A
```
#### `AndroidManifest.xml`:
```xml
<!-- N/A -->
```
</p>
</details>
---
## Environment
<details><summary>Click To Expand</summary>
<p>
**`react-native info` output:**
<!-- Please run `react-native info` on your terminal and paste the contents into the code block below -->
```
React Native Environment Info:
System:
OS: macOS 10.14.2
CPU: (4) x64 Intel(R) Core(TM) i5-4278U CPU @ 2.60GHz
Memory: 1.17 GB / 8.00 GB
Shell: 3.2.57 - /bin/bash
Binaries:
Node: 11.6.0 - /usr/local/bin/node
Yarn: 1.12.3 - /usr/local/bin/yarn
npm: 6.5.0 - /usr/local/bin/npm
Watchman: 4.9.0 - /usr/local/bin/watchman
SDKs:
iOS SDK:
Platforms: iOS 12.1, macOS 10.14, tvOS 12.1, watchOS 5.1
Android SDK:
API Levels: 27, 28
Build Tools: 27.0.3, 28.0.3
System Images: android-18 | Google APIs Intel x86 Atom, android-19 | Google APIs Intel x86 Atom, android-22 | Google APIs Intel x86 Atom, android-26 | Google APIs Intel x86 Atom, android-28 | Google APIs Intel x86 Atom
IDEs:
Android Studio: 3.2 AI-181.5540.7.32.5056338
Xcode: 10.1/10B61 - /usr/bin/xcodebuild
npmPackages:
react: 16.8.3 => 16.8.3
react-native: 0.58.6 => 0.58.6
npmGlobalPackages:
react-native-app-id: 0.0.5
react-native-cli: 2.0.1
```
<!-- change `[ ]` to `[x]` to select an option(s) -->
- **Platform that you're experiencing the issue on**:
- [x] iOS
- [ ] Android
- [ ] **iOS** but have not tested behavior on Android
- [ ] **Android** but have not tested behavior on iOS
- [ ] Both
- **`Firebase` module(s) you're using that has the issue:**
- `Admob`
- **Are you using `TypeScript`?**
- `N`
</p>
</details>
<!-- Thanks for reading this far down ❤️ -->
<!-- High quality, detailed issues are much easier to triage for maintainers -->
<!-- For bonus points, if you put a 🔥 (:fire:) emojii at the start of the issue title we'll know -->
<!-- that you took the time to fill this out correctly, or, at least read this far -->
---
Think `react-native-firebase` is great? Please consider supporting all of the project maintainers and contributors by donating via our [Open Collective](https://opencollective.com/react-native-firebase/donate) where all contributors can submit expenses. [[Learn More]](https://invertase.io/oss/react-native-firebase/contributing/donations-expenses)
- 👉 Check out [`React Native Firebase`](https://twitter.com/rnfirebase) and [`Invertase`](https://twitter.com/invertaseio) on Twitter for updates on the library.
|
1.0
|
[IOS] Admob - You attempted to use a firebase module that's not installed natively on your iOS project by calling firebase.admob() - <!---
Hello there you awesome person;
Please note that the issue list of this repo is exclusively for bug reports;
1) For feature requests please visit our [Feature Request Board](https://boards.invertase.io/react-native-firebase).
2) For questions and support please use our Discord chat: https://discord.gg/C9aK28N or Stack Overflow: https://stackoverflow.com/questions/tagged/react-native-firebase
3) If this is a setup issue then please make sure you've correctly followed the setup guides, most setup issues such as 'duplicate dex files', 'default app has not been initialized' etc are all down to an incorrect setup as the guides haven't been correctly followed.
-->
<!-- NOTE: You can change any of the `[ ]` to `[x]` to mark an option(s) as selected -->
<!-- PLEASE DO NOT REMOVE ANY SECTIONS FROM THIS ISSUE TEMPLATE -->
<!-- Leave them as they are even if they're irrelevant to your issue -->
## Issue
<!-- Please describe your issue here --^ and provide as much detail as you can. -->
<!-- Include code snippets that show your usages of the library in the context of your project. -->
<!-- Snippets that also show how and where the library is imported in JS are useful to debug issues relating to importing or methods not found issues -->
I am having an issue on iOS with the adMob pod.
On Android everything works fine.
On iOS I get the following error when trying to call firebase.admob() :
```
Error: You attempted to use a firebase module that's not installed natively on your iOS project by calling firebase.admob().
Ensure you have the required Firebase iOS SDK pod for this module included in your Podfile, in this instance confirm you've added "pod 'Firebase/AdMob'" to your Podfile
See http://invertase.link/ios for full setup instructions.
This error is located at:
in App (at renderApplication.js:34)
in RCTView (at View.js:45)
in View (at AppContainer.js:98)
in RCTView (at View.js:45)
in View (at AppContainer.js:115)
in AppContainer (at renderApplication.js:33)
initialiseNativeModule
native.js:43:20
ModuleBase
ModuleBase.js:26:27
AdMob
index.js:21:4
<unknown>
apps.js:66:74
getModule
apps.js:186:20
App
App.js:98:21
App
[native code]:0
<unknown>
createClassProxy.js:98:23
instantiate
createClassProxy.js:96:6
constructClassInstance
ReactNativeRenderer-dev.js:8864:26
updateClassComponent
ReactNativeRenderer-dev.js:11508:6
performUnitOfWork
ReactNativeRenderer-dev.js:16075:21
workLoop
ReactNativeRenderer-dev.js:16115:41
renderRoot
ReactNativeRenderer-dev.js:16219:15
performWorkOnRoot
ReactNativeRenderer-dev.js:17198:17
performWork
ReactNativeRenderer-dev.js:17099:24
performSyncWork
ReactNativeRenderer-dev.js:17060:14
requestWork
ReactNativeRenderer-dev.js:16925:19
scheduleWork
ReactNativeRenderer-dev.js:16724:16
scheduleRootUpdate
ReactNativeRenderer-dev.js:17413:15
render
ReactNativeRenderer-dev.js:18191:20
renderApplication
renderApplication.js:59:34
run
AppRegistry.js:101:10
runApplication
AppRegistry.js:195:26
__callFunction
MessageQueue.js:366:47
<unknown>
MessageQueue.js:106:26
__guard
MessageQueue.js:314:10
callFunctionReturnFlushedQueue
MessageQueue.js:105:17
callFunctionReturnFlushedQueue
[native code]:0
```
I've tried re-do the initial setup guide on my project. And adding this pod:
`pod 'RNFirebase', :path => '../node_modules/react-native-firebase/ios'`
but I still have this issue
---
## Project Files
<!-- Provide the contents of key project files which will help to debug -->
<!-- For Example: -->
<!-- - iOS: `Podfile` contents. -->
<!-- - Android: `android/build.gradle` contents. -->
<!-- - Android: `android/app/build.gradle` contents. -->
<!-- - Android: `AndroidManifest.xml` contents. -->
<!-- ADD THE CONTENTS OF THE FILES IN THE PROVIDED CODE BLOCKS BELOW -->
### iOS
<details><summary>Click To Expand</summary>
<p>
#### `ios/Podfile`:
- [ ] I'm not using Pods
- [x] I'm using Pods and my Podfile looks like:
```
platform :ios, '9.0'
target 'xxx' do
# Uncomment the next line if you're using Swift or would like to use dynamic frameworks
# use_frameworks!
# Pods for xxx
pod 'React', :path => '../node_modules/react-native'
pod 'yoga', :path => '../node_modules/react-native/ReactCommon/yoga'
pod 'Firebase/Core', '~> 5.15.0'
pod 'Firebase/Database', '~> 5.15.0'
pod 'Firebase/AdMob', '~> 5.15.0'
pod 'RNCAsyncStorage', :path => '../node_modules/@react-native-community/async-storage'
target 'xxx' do
inherit! :search_paths
# Pods for testing
end
target 'xxx' do
inherit! :search_paths
# Pods for testing
end
end
```
`podfile.lock`:
```
PODS:
- Firebase/AdMob (5.15.0):
- Firebase/Core
- Google-Mobile-Ads-SDK (~> 7.37)
- Firebase/Core (5.15.0):
- Firebase/CoreOnly
- FirebaseAnalytics (= 5.4.0)
- Firebase/CoreOnly (5.15.0):
- FirebaseCore (= 5.1.10)
- Firebase/Database (5.15.0):
- Firebase/CoreOnly
- FirebaseDatabase (= 5.0.4)
- FirebaseAnalytics (5.4.0):
- FirebaseCore (~> 5.1)
- FirebaseInstanceID (~> 3.3)
- GoogleAppMeasurement (= 5.4.0)
- GoogleUtilities/AppDelegateSwizzler (~> 5.2)
- GoogleUtilities/MethodSwizzler (~> 5.2)
- GoogleUtilities/Network (~> 5.2)
- "GoogleUtilities/NSData+zlib (~> 5.2)"
- nanopb (~> 0.3)
- FirebaseAuthInterop (1.0.0)
- FirebaseCore (5.1.10):
- GoogleUtilities/Logger (~> 5.2)
- FirebaseDatabase (5.0.4):
- FirebaseAuthInterop (~> 1.0)
- FirebaseCore (~> 5.1)
- leveldb-library (~> 1.18)
- FirebaseInstanceID (3.3.0):
- FirebaseCore (~> 5.1)
- GoogleUtilities/Environment (~> 5.3)
- GoogleUtilities/UserDefaults (~> 5.3)
- Google-Mobile-Ads-SDK (7.41.0)
- GoogleAppMeasurement (5.4.0):
- GoogleUtilities/AppDelegateSwizzler (~> 5.2)
- GoogleUtilities/MethodSwizzler (~> 5.2)
- GoogleUtilities/Network (~> 5.2)
- "GoogleUtilities/NSData+zlib (~> 5.2)"
- nanopb (~> 0.3)
- GoogleUtilities/AppDelegateSwizzler (5.6.0):
- GoogleUtilities/Environment
- GoogleUtilities/Logger
- GoogleUtilities/Network
- GoogleUtilities/Environment (5.6.0)
- GoogleUtilities/Logger (5.6.0):
- GoogleUtilities/Environment
- GoogleUtilities/MethodSwizzler (5.6.0):
- GoogleUtilities/Logger
- GoogleUtilities/Network (5.6.0):
- GoogleUtilities/Logger
- "GoogleUtilities/NSData+zlib"
- GoogleUtilities/Reachability
- "GoogleUtilities/NSData+zlib (5.6.0)"
- GoogleUtilities/Reachability (5.6.0):
- GoogleUtilities/Logger
- GoogleUtilities/UserDefaults (5.6.0):
- GoogleUtilities/Logger
- leveldb-library (1.20)
- nanopb (0.3.901):
- nanopb/decode (= 0.3.901)
- nanopb/encode (= 0.3.901)
- nanopb/decode (0.3.901)
- nanopb/encode (0.3.901)
- React (0.58.6):
- React/Core (= 0.58.6)
- React/Core (0.58.6):
- yoga (= 0.58.6.React)
- RNCAsyncStorage (1.4.0):
- React
- yoga (0.58.6.React)
DEPENDENCIES:
- Firebase/AdMob (~> 5.15.0)
- Firebase/Core (~> 5.15.0)
- Firebase/Database (~> 5.15.0)
- React (from `../node_modules/react-native`)
- "RNCAsyncStorage (from `../node_modules/@react-native-community/async-storage`)"
- yoga (from `../node_modules/react-native/ReactCommon/yoga`)
SPEC REPOS:
https://github.com/cocoapods/specs.git:
- Firebase
- FirebaseAnalytics
- FirebaseAuthInterop
- FirebaseCore
- FirebaseDatabase
- FirebaseInstanceID
- Google-Mobile-Ads-SDK
- GoogleAppMeasurement
- GoogleUtilities
- leveldb-library
- nanopb
EXTERNAL SOURCES:
React:
:path: "../node_modules/react-native"
RNCAsyncStorage:
:path: "../node_modules/@react-native-community/async-storage"
yoga:
:path: "../node_modules/react-native/ReactCommon/yoga"
SPEC CHECKSUMS:
Firebase: 8bb9268bff82374f2cbaaabb143e725743c316ae
FirebaseAnalytics: c06f9d70577d79074214700a71fd5d39de5550fb
FirebaseAuthInterop: 0ffa57668be100582bb7643d4fcb7615496c41fc
FirebaseCore: 35747502d9e8c6ee217385ad04446c7c2aaf9c5c
FirebaseDatabase: 0621689f77528d62b47e1c06ca737c4c19275d1a
FirebaseInstanceID: e2fa4cb35ef5558c200f7f0ad8a53e212215f93e
Google-Mobile-Ads-SDK: 101ce89b94fdf60ec64ba7b6e9b9d812dc765936
GoogleAppMeasurement: 98b71f5e04142793729a5ef23e5b96651ff4b70f
GoogleUtilities: f44b068df02195da10409f415bd8c98439cd70fa
leveldb-library: 08cba283675b7ed2d99629a4bc5fd052cd2bb6a5
nanopb: 2901f78ea1b7b4015c860c2fdd1ea2fee1a18d48
React: 130b87b2d5e2baac646954282cab87be986d98fc
RNCAsyncStorage: b82dc6e5b39a625d70e3b3492bff75c0de94ba71
yoga: 32d7ef1081951e9a35a4c72a7be797598b138a48
PODFILE CHECKSUM: 7c7dc6ad557e143e89cd532086f991b949c19580
COCOAPODS: 1.6.1
```
#### `AppDelegate.m`:
```objc
// N/A
```
</p>
</details>
---
### Android
<details><summary>Click To Expand</summary>
<p>
#### `android/build.gradle`:
```groovy
// N/A
```
#### `android/app/build.gradle`:
```groovy
// N/A
```
#### `android/settings.gradle`:
```groovy
// N/A
```
#### `MainApplication.java`:
```java
// N/A
```
#### `AndroidManifest.xml`:
```xml
<!-- N/A -->
```
</p>
</details>
---
## Environment
<details><summary>Click To Expand</summary>
<p>
**`react-native info` output:**
<!-- Please run `react-native info` on your terminal and paste the contents into the code block below -->
```
React Native Environment Info:
System:
OS: macOS 10.14.2
CPU: (4) x64 Intel(R) Core(TM) i5-4278U CPU @ 2.60GHz
Memory: 1.17 GB / 8.00 GB
Shell: 3.2.57 - /bin/bash
Binaries:
Node: 11.6.0 - /usr/local/bin/node
Yarn: 1.12.3 - /usr/local/bin/yarn
npm: 6.5.0 - /usr/local/bin/npm
Watchman: 4.9.0 - /usr/local/bin/watchman
SDKs:
iOS SDK:
Platforms: iOS 12.1, macOS 10.14, tvOS 12.1, watchOS 5.1
Android SDK:
API Levels: 27, 28
Build Tools: 27.0.3, 28.0.3
System Images: android-18 | Google APIs Intel x86 Atom, android-19 | Google APIs Intel x86 Atom, android-22 | Google APIs Intel x86 Atom, android-26 | Google APIs Intel x86 Atom, android-28 | Google APIs Intel x86 Atom
IDEs:
Android Studio: 3.2 AI-181.5540.7.32.5056338
Xcode: 10.1/10B61 - /usr/bin/xcodebuild
npmPackages:
react: 16.8.3 => 16.8.3
react-native: 0.58.6 => 0.58.6
npmGlobalPackages:
react-native-app-id: 0.0.5
react-native-cli: 2.0.1
```
<!-- change `[ ]` to `[x]` to select an option(s) -->
- **Platform that you're experiencing the issue on**:
- [x] iOS
- [ ] Android
- [ ] **iOS** but have not tested behavior on Android
- [ ] **Android** but have not tested behavior on iOS
- [ ] Both
- **`Firebase` module(s) you're using that has the issue:**
- `Admob`
- **Are you using `TypeScript`?**
- `N`
</p>
</details>
<!-- Thanks for reading this far down ❤️ -->
<!-- High quality, detailed issues are much easier to triage for maintainers -->
<!-- For bonus points, if you put a 🔥 (:fire:) emojii at the start of the issue title we'll know -->
<!-- that you took the time to fill this out correctly, or, at least read this far -->
---
Think `react-native-firebase` is great? Please consider supporting all of the project maintainers and contributors by donating via our [Open Collective](https://opencollective.com/react-native-firebase/donate) where all contributors can submit expenses. [[Learn More]](https://invertase.io/oss/react-native-firebase/contributing/donations-expenses)
- 👉 Check out [`React Native Firebase`](https://twitter.com/rnfirebase) and [`Invertase`](https://twitter.com/invertaseio) on Twitter for updates on the library.
|
non_defect
|
admob you attempted to use a firebase module that s not installed natively on your ios project by calling firebase admob hello there you awesome person please note that the issue list of this repo is exclusively for bug reports for feature requests please visit our for questions and support please use our discord chat or stack overflow if this is a setup issue then please make sure you ve correctly followed the setup guides most setup issues such as duplicate dex files default app has not been initialized etc are all down to an incorrect setup as the guides haven t been correctly followed issue i am having an issue on ios with the admob pod on android everything works fine on ios i get the following error when trying to call firebase admob error you attempted to use a firebase module that s not installed natively on your ios project by calling firebase admob ensure you have the required firebase ios sdk pod for this module included in your podfile in this instance confirm you ve added pod firebase admob to your podfile see for full setup instructions this error is located at in app at renderapplication js in rctview at view js in view at appcontainer js in rctview at view js in view at appcontainer js in appcontainer at renderapplication js initialisenativemodule native js modulebase modulebase js admob index js apps js getmodule apps js app app js app createclassproxy js instantiate createclassproxy js constructclassinstance reactnativerenderer dev js updateclasscomponent reactnativerenderer dev js performunitofwork reactnativerenderer dev js workloop reactnativerenderer dev js renderroot reactnativerenderer dev js performworkonroot reactnativerenderer dev js performwork reactnativerenderer dev js performsyncwork reactnativerenderer dev js requestwork reactnativerenderer dev js schedulework reactnativerenderer dev js schedulerootupdate reactnativerenderer dev js render reactnativerenderer dev js renderapplication renderapplication js run appregistry js runapplication appregistry js callfunction messagequeue js messagequeue js guard messagequeue js callfunctionreturnflushedqueue messagequeue js callfunctionreturnflushedqueue i ve tried re do the initial setup guide on my project and adding this pod pod rnfirebase path node modules react native firebase ios but i still have this issue project files ios click to expand ios podfile i m not using pods i m using pods and my podfile looks like platform ios target xxx do uncomment the next line if you re using swift or would like to use dynamic frameworks use frameworks pods for xxx pod react path node modules react native pod yoga path node modules react native reactcommon yoga pod firebase core pod firebase database pod firebase admob pod rncasyncstorage path node modules react native community async storage target xxx do inherit search paths pods for testing end target xxx do inherit search paths pods for testing end end podfile lock pods firebase admob firebase core google mobile ads sdk firebase core firebase coreonly firebaseanalytics firebase coreonly firebasecore firebase database firebase coreonly firebasedatabase firebaseanalytics firebasecore firebaseinstanceid googleappmeasurement googleutilities appdelegateswizzler googleutilities methodswizzler googleutilities network googleutilities nsdata zlib nanopb firebaseauthinterop firebasecore googleutilities logger firebasedatabase firebaseauthinterop firebasecore leveldb library firebaseinstanceid firebasecore googleutilities environment googleutilities userdefaults google mobile ads sdk googleappmeasurement googleutilities appdelegateswizzler googleutilities methodswizzler googleutilities network googleutilities nsdata zlib nanopb googleutilities appdelegateswizzler googleutilities environment googleutilities logger googleutilities network googleutilities environment googleutilities logger googleutilities environment googleutilities methodswizzler googleutilities logger googleutilities network googleutilities logger googleutilities nsdata zlib googleutilities reachability googleutilities nsdata zlib googleutilities reachability googleutilities logger googleutilities userdefaults googleutilities logger leveldb library nanopb nanopb decode nanopb encode nanopb decode nanopb encode react react core react core yoga react rncasyncstorage react yoga react dependencies firebase admob firebase core firebase database react from node modules react native rncasyncstorage from node modules react native community async storage yoga from node modules react native reactcommon yoga spec repos firebase firebaseanalytics firebaseauthinterop firebasecore firebasedatabase firebaseinstanceid google mobile ads sdk googleappmeasurement googleutilities leveldb library nanopb external sources react path node modules react native rncasyncstorage path node modules react native community async storage yoga path node modules react native reactcommon yoga spec checksums firebase firebaseanalytics firebaseauthinterop firebasecore firebasedatabase firebaseinstanceid google mobile ads sdk googleappmeasurement googleutilities leveldb library nanopb react rncasyncstorage yoga podfile checksum cocoapods appdelegate m objc n a android click to expand android build gradle groovy n a android app build gradle groovy n a android settings gradle groovy n a mainapplication java java n a androidmanifest xml xml environment click to expand react native info output react native environment info system os macos cpu intel r core tm cpu memory gb gb shell bin bash binaries node usr local bin node yarn usr local bin yarn npm usr local bin npm watchman usr local bin watchman sdks ios sdk platforms ios macos tvos watchos android sdk api levels build tools system images android google apis intel atom android google apis intel atom android google apis intel atom android google apis intel atom android google apis intel atom ides android studio ai xcode usr bin xcodebuild npmpackages react react native npmglobalpackages react native app id react native cli platform that you re experiencing the issue on ios android ios but have not tested behavior on android android but have not tested behavior on ios both firebase module s you re using that has the issue admob are you using typescript n think react native firebase is great please consider supporting all of the project maintainers and contributors by donating via our where all contributors can submit expenses 👉 check out and on twitter for updates on the library
| 0
|
42,785
| 11,072,699,045
|
IssuesEvent
|
2019-12-12 10:48:59
|
rticommunity/rticonnextdds-examples
|
https://api.github.com/repos/rticommunity/rticonnextdds-examples
|
closed
|
Cut & paste issue in build example CMake file
|
bug build
|
<!-- :warning: Please, try to follow the template -->
### Information
- **RTI Product**: Connext DDS Pro
- **Version**: 6.0
- **Operating system**:
- **Compiler**:
- **Compiler version**:
- **Additional information**:
### What is the current behavior?
A customer pointed out an error in an example CMake file in this repo. The file is rticonnextdds-examples/examples/connext_dds/build_systems/cmake/CMakeLists.txt. The error appears in lines 60-62:
# Create the target for the publisher executable
add_executable(HelloWorld_subscriber
"${CMAKE_CURRENT_SOURCE_DIR}/HelloWorld_publisher.c"
${HelloWorld_C_GENERATED_SOURCES}
)
### Steps to reproduce the issue
### What is the expected behavior?
### Suggested solutions
It appears that the comment should be corrected to read "subscriber" and that the source file name two lines later should be "HelloWorld_subscriber.c"
|
1.0
|
Cut & paste issue in build example CMake file - <!-- :warning: Please, try to follow the template -->
### Information
- **RTI Product**: Connext DDS Pro
- **Version**: 6.0
- **Operating system**:
- **Compiler**:
- **Compiler version**:
- **Additional information**:
### What is the current behavior?
A customer pointed out an error in an example CMake file in this repo. The file is rticonnextdds-examples/examples/connext_dds/build_systems/cmake/CMakeLists.txt. The error appears in lines 60-62:
# Create the target for the publisher executable
add_executable(HelloWorld_subscriber
"${CMAKE_CURRENT_SOURCE_DIR}/HelloWorld_publisher.c"
${HelloWorld_C_GENERATED_SOURCES}
)
### Steps to reproduce the issue
### What is the expected behavior?
### Suggested solutions
It appears that the comment should be corrected to read "subscriber" and that the source file name two lines later should be "HelloWorld_subscriber.c"
|
non_defect
|
cut paste issue in build example cmake file information rti product connext dds pro version operating system compiler compiler version additional information what is the current behavior a customer pointed out an error in an example cmake file in this repo the file is rticonnextdds examples examples connext dds build systems cmake cmakelists txt the error appears in lines create the target for the publisher executable add executable helloworld subscriber cmake current source dir helloworld publisher c helloworld c generated sources steps to reproduce the issue what is the expected behavior suggested solutions it appears that the comment should be corrected to read subscriber and that the source file name two lines later should be helloworld subscriber c
| 0
|
630,479
| 20,110,814,943
|
IssuesEvent
|
2022-02-07 14:56:27
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.t-online.de - design is broken
|
browser-firefox priority-important engine-gecko
|
<!-- @browser: Firefox 100.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:100.0) Gecko/20100101 Firefox/100.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/99243 -->
**URL**: https://www.t-online.de/
**Browser / Version**: Firefox 100.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Images not loaded
**Steps to Reproduce**:
pictures are not visible. website checked with chrome and edge, all pictures visible.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/2/277cd791-75ef-4298-861a-37d61b43b5dd.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220206065831</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/2/a6e7566b-affd-4ab3-8ec0-9d7880814570)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.t-online.de - design is broken - <!-- @browser: Firefox 100.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:100.0) Gecko/20100101 Firefox/100.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/99243 -->
**URL**: https://www.t-online.de/
**Browser / Version**: Firefox 100.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Images not loaded
**Steps to Reproduce**:
pictures are not visible. website checked with chrome and edge, all pictures visible.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/2/277cd791-75ef-4298-861a-37d61b43b5dd.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220206065831</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/2/a6e7566b-affd-4ab3-8ec0-9d7880814570)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
design is broken url browser version firefox operating system windows tested another browser yes chrome problem type design is broken description images not loaded steps to reproduce pictures are not visible website checked with chrome and edge all pictures visible view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
47,886
| 13,066,330,997
|
IssuesEvent
|
2020-07-30 21:28:13
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
weighting 'from_simprod' function tries to load missing values (Trac #1356)
|
Migrated from Trac combo reconstruction defect
|
icecube.weighting.weighting.from_simprod (weighting.py L578)
after connecting to the simprod database, the function tries to load 'NUGEN::elogmin', 'NUGEN::elogmax', and 'NUGEN::injectionradius', which do not exist in new neutrino generator tables.
Migrated from https://code.icecube.wisc.edu/ticket/1356
```json
{
"status": "closed",
"changetime": "2015-09-18T18:29:44",
"description": "icecube.weighting.weighting.from_simprod (weighting.py L578)\nafter connecting to the simprod database, the function tries to load 'NUGEN::elogmin', 'NUGEN::elogmax', and 'NUGEN::injectionradius', which do not exist in new neutrino generator tables.",
"reporter": "ddouglas",
"cc": "",
"resolution": "fixed",
"_ts": "1442600984726412",
"component": "combo reconstruction",
"summary": "weighting 'from_simprod' function tries to load missing values",
"priority": "normal",
"keywords": "weighting, simprod",
"time": "2015-09-18T18:15:17",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
|
1.0
|
weighting 'from_simprod' function tries to load missing values (Trac #1356) - icecube.weighting.weighting.from_simprod (weighting.py L578)
after connecting to the simprod database, the function tries to load 'NUGEN::elogmin', 'NUGEN::elogmax', and 'NUGEN::injectionradius', which do not exist in new neutrino generator tables.
Migrated from https://code.icecube.wisc.edu/ticket/1356
```json
{
"status": "closed",
"changetime": "2015-09-18T18:29:44",
"description": "icecube.weighting.weighting.from_simprod (weighting.py L578)\nafter connecting to the simprod database, the function tries to load 'NUGEN::elogmin', 'NUGEN::elogmax', and 'NUGEN::injectionradius', which do not exist in new neutrino generator tables.",
"reporter": "ddouglas",
"cc": "",
"resolution": "fixed",
"_ts": "1442600984726412",
"component": "combo reconstruction",
"summary": "weighting 'from_simprod' function tries to load missing values",
"priority": "normal",
"keywords": "weighting, simprod",
"time": "2015-09-18T18:15:17",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
|
defect
|
weighting from simprod function tries to load missing values trac icecube weighting weighting from simprod weighting py after connecting to the simprod database the function tries to load nugen elogmin nugen elogmax and nugen injectionradius which do not exist in new neutrino generator tables migrated from json status closed changetime description icecube weighting weighting from simprod weighting py nafter connecting to the simprod database the function tries to load nugen elogmin nugen elogmax and nugen injectionradius which do not exist in new neutrino generator tables reporter ddouglas cc resolution fixed ts component combo reconstruction summary weighting from simprod function tries to load missing values priority normal keywords weighting simprod time milestone owner jvansanten type defect
| 1
|
15,744
| 3,973,961,368
|
IssuesEvent
|
2016-05-04 20:28:18
|
LigaData/Kamanja
|
https://api.github.com/repos/LigaData/Kamanja
|
closed
|
suggestion: document $kamanja help in a table
|
Bug (Documentation) Needs Discussion P3 Verify
|
Hello,
As a suggestion in the documentation, when you type the following line at the command line, the output is so long. Perhaps the documentation could create some kind of table to help keep things more clear.
$ kamanja help
When using search in the documentation for "kamanja help", I never came up with what appeard to be b documentation on this help message, or on what every line meant.
I attached a possible example table. You may want to re-arrange commands in columns.
[Kamanja command notes.xlsx](https://github.com/ligaDATA/Kamanja/files/122268/Kamanja.command.notes.xlsx)
|
1.0
|
suggestion: document $kamanja help in a table - Hello,
As a suggestion in the documentation, when you type the following line at the command line, the output is so long. Perhaps the documentation could create some kind of table to help keep things more clear.
$ kamanja help
When using search in the documentation for "kamanja help", I never came up with what appeard to be b documentation on this help message, or on what every line meant.
I attached a possible example table. You may want to re-arrange commands in columns.
[Kamanja command notes.xlsx](https://github.com/ligaDATA/Kamanja/files/122268/Kamanja.command.notes.xlsx)
|
non_defect
|
suggestion document kamanja help in a table hello as a suggestion in the documentation when you type the following line at the command line the output is so long perhaps the documentation could create some kind of table to help keep things more clear kamanja help when using search in the documentation for kamanja help i never came up with what appeard to be b documentation on this help message or on what every line meant i attached a possible example table you may want to re arrange commands in columns
| 0
|
31,947
| 6,666,585,789
|
IssuesEvent
|
2017-10-03 08:55:57
|
primefaces/primeng
|
https://api.github.com/repos/primefaces/primeng
|
closed
|
Filtered listbox toggle all fails with new items
|
defect
|
### There is no guarantee in receiving a response in GitHub Issue Tracker, If you'd like to secure our response, you may consider *PrimeNG PRO Support* where support is provided within 4 business hours
**I'm submitting a ...** (check one with "x")
```
[x ] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
Please fork the plunkr below and create a case demonstrating your bug report. Issues without a plunkr have much less possibility to be reviewed.
http://plnkr.co/edit/2qeE3EaVRu5AB3ZMSbd5?p=preview
**Current behavior**
1. Type the letter u (or arbitrary text) into the input filter.
2. Delete the text in the input search filter.
3. click the 'Add Unit' button to add another unit to the list.
4. use the 'select all' checkbox. Not all the items will be checked.
**Expected behavior**
Expecting all visible list items to be checked.
**What is the motivation / use case for changing the behavior?**
I cannot imagine this is expected behavior.
|
1.0
|
Filtered listbox toggle all fails with new items - ### There is no guarantee in receiving a response in GitHub Issue Tracker, If you'd like to secure our response, you may consider *PrimeNG PRO Support* where support is provided within 4 business hours
**I'm submitting a ...** (check one with "x")
```
[x ] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
Please fork the plunkr below and create a case demonstrating your bug report. Issues without a plunkr have much less possibility to be reviewed.
http://plnkr.co/edit/2qeE3EaVRu5AB3ZMSbd5?p=preview
**Current behavior**
1. Type the letter u (or arbitrary text) into the input filter.
2. Delete the text in the input search filter.
3. click the 'Add Unit' button to add another unit to the list.
4. use the 'select all' checkbox. Not all the items will be checked.
**Expected behavior**
Expecting all visible list items to be checked.
**What is the motivation / use case for changing the behavior?**
I cannot imagine this is expected behavior.
|
defect
|
filtered listbox toggle all fails with new items there is no guarantee in receiving a response in github issue tracker if you d like to secure our response you may consider primeng pro support where support is provided within business hours i m submitting a check one with x bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see plunkr case bug reports please fork the plunkr below and create a case demonstrating your bug report issues without a plunkr have much less possibility to be reviewed current behavior type the letter u or arbitrary text into the input filter delete the text in the input search filter click the add unit button to add another unit to the list use the select all checkbox not all the items will be checked expected behavior expecting all visible list items to be checked what is the motivation use case for changing the behavior i cannot imagine this is expected behavior
| 1
|
54,623
| 23,319,087,018
|
IssuesEvent
|
2022-08-08 14:51:45
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
[Discuss] Improve long URL handling in Kibana
|
Team:Core loe:hours Team:AppServicesSv impact:low
|
To improve how long URLs are treated in Kibana, please consider below points for discussion:
1. Come up with a way to measure the actual maximum URL length for a specific browser session.
1. Do not enforce arbitrary limits on maximum length of URLs in Kibana.
1. Allow developers to use the maximum length of URLs provided by each browser.
1. For any warnings or errors use the actual detected maximum URL length supported by that specific browser the user is using instead of hard-coding some URL length values.
1. Remove the `state:storeInSessionStorage` UI Setting. Instead, for each user session (each browser) detect the actual maximum URL length supported by that browser. `state:storeInSessionStorage` is applied to all users and all their devices, but each browser (even for the same user) may have different maximum URL length limits.
|
1.0
|
[Discuss] Improve long URL handling in Kibana - To improve how long URLs are treated in Kibana, please consider below points for discussion:
1. Come up with a way to measure the actual maximum URL length for a specific browser session.
1. Do not enforce arbitrary limits on maximum length of URLs in Kibana.
1. Allow developers to use the maximum length of URLs provided by each browser.
1. For any warnings or errors use the actual detected maximum URL length supported by that specific browser the user is using instead of hard-coding some URL length values.
1. Remove the `state:storeInSessionStorage` UI Setting. Instead, for each user session (each browser) detect the actual maximum URL length supported by that browser. `state:storeInSessionStorage` is applied to all users and all their devices, but each browser (even for the same user) may have different maximum URL length limits.
|
non_defect
|
improve long url handling in kibana to improve how long urls are treated in kibana please consider below points for discussion come up with a way to measure the actual maximum url length for a specific browser session do not enforce arbitrary limits on maximum length of urls in kibana allow developers to use the maximum length of urls provided by each browser for any warnings or errors use the actual detected maximum url length supported by that specific browser the user is using instead of hard coding some url length values remove the state storeinsessionstorage ui setting instead for each user session each browser detect the actual maximum url length supported by that browser state storeinsessionstorage is applied to all users and all their devices but each browser even for the same user may have different maximum url length limits
| 0
|
821,823
| 30,838,282,879
|
IssuesEvent
|
2023-08-02 08:55:50
|
graasp/graasp-app-excalidraw
|
https://api.github.com/repos/graasp/graasp-app-excalidraw
|
closed
|
Upgrade for refactored Graasp
|
dependencies feature priority
|
The app must be adapted to work with the new and refactored version of Graasp.
|
1.0
|
Upgrade for refactored Graasp - The app must be adapted to work with the new and refactored version of Graasp.
|
non_defect
|
upgrade for refactored graasp the app must be adapted to work with the new and refactored version of graasp
| 0
|
77,663
| 27,100,433,725
|
IssuesEvent
|
2023-02-15 08:12:04
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Calling iterator() on a stream causes eager evaluation
|
T: Defect C: Functionality P: Medium E: All Editions
|
### Expected behavior
My rough expectation is that the `iterator()` to be non-eager (as described in the [docs here](https://docs.oracle.com/javase/8/docs/api/java/util/stream/package-summary.html#StreamOps) ). However, it looks like calling it eagerly evaluates the stream.
Here's a small example:
**No call to iterator, directly consuming stream:**
```
long start = System.currentTimeMillis();
tx.dsl()
.select(STUFF.ID).from(STUFF).limit(1000)
.fetchSize(1)
.stream()
.forEach(__ -> {
System.out.println("First result after: " + (System.currentTimeMillis() - start));
});
```
This starts printing results as fast as the DB can vend them. A few milliseconds or so.
### Actual behavior
The stream is eagerly consumed before allowing iteration. Here's an example:
```
long start = System.currentTimeMillis();
tx.dsl()
.select(STUFF.ID).from(STUFF).limit(1000)
.fetchSize(1)
.stream()
.iterator() // hangs out here while fetching the entire data set
.next());
System.out.println("First result after: " + (System.currentTimeMillis() - start));
```
Is this expected behavior or a bug? `interator()` being an exception to the usual rule for terminal operators is pretty understated in the docs. So, I'm not sure if JOOQ is expected to honor it or not.
### Steps to reproduce the problem
```
long start = System.currentTimeMillis();
tx.dsl()
.select(STUFF.ID).from(STUFF).limit(1000)
.fetchSize(1)
.stream()
.iterator() // hangs out here while fetching the entire data set
.next());
System.out.println("First result after: " + (System.currentTimeMillis() - start));
```
### jOOQ Version
JOOQ 3.14
### Database product and version
Postgresql 12.xx
### Java Version
OpenJDK 8
### OS Version
MacOS
### JDBC driver name and version (include name if unofficial driver)
org.postgresql_postgresql-42.3.6
|
1.0
|
Calling iterator() on a stream causes eager evaluation - ### Expected behavior
My rough expectation is that the `iterator()` to be non-eager (as described in the [docs here](https://docs.oracle.com/javase/8/docs/api/java/util/stream/package-summary.html#StreamOps) ). However, it looks like calling it eagerly evaluates the stream.
Here's a small example:
**No call to iterator, directly consuming stream:**
```
long start = System.currentTimeMillis();
tx.dsl()
.select(STUFF.ID).from(STUFF).limit(1000)
.fetchSize(1)
.stream()
.forEach(__ -> {
System.out.println("First result after: " + (System.currentTimeMillis() - start));
});
```
This starts printing results as fast as the DB can vend them. A few milliseconds or so.
### Actual behavior
The stream is eagerly consumed before allowing iteration. Here's an example:
```
long start = System.currentTimeMillis();
tx.dsl()
.select(STUFF.ID).from(STUFF).limit(1000)
.fetchSize(1)
.stream()
.iterator() // hangs out here while fetching the entire data set
.next());
System.out.println("First result after: " + (System.currentTimeMillis() - start));
```
Is this expected behavior or a bug? `interator()` being an exception to the usual rule for terminal operators is pretty understated in the docs. So, I'm not sure if JOOQ is expected to honor it or not.
### Steps to reproduce the problem
```
long start = System.currentTimeMillis();
tx.dsl()
.select(STUFF.ID).from(STUFF).limit(1000)
.fetchSize(1)
.stream()
.iterator() // hangs out here while fetching the entire data set
.next());
System.out.println("First result after: " + (System.currentTimeMillis() - start));
```
### jOOQ Version
JOOQ 3.14
### Database product and version
Postgresql 12.xx
### Java Version
OpenJDK 8
### OS Version
MacOS
### JDBC driver name and version (include name if unofficial driver)
org.postgresql_postgresql-42.3.6
|
defect
|
calling iterator on a stream causes eager evaluation expected behavior my rough expectation is that the iterator to be non eager as described in the however it looks like calling it eagerly evaluates the stream here s a small example no call to iterator directly consuming stream long start system currenttimemillis tx dsl select stuff id from stuff limit fetchsize stream foreach system out println first result after system currenttimemillis start this starts printing results as fast as the db can vend them a few milliseconds or so actual behavior the stream is eagerly consumed before allowing iteration here s an example long start system currenttimemillis tx dsl select stuff id from stuff limit fetchsize stream iterator hangs out here while fetching the entire data set next system out println first result after system currenttimemillis start is this expected behavior or a bug interator being an exception to the usual rule for terminal operators is pretty understated in the docs so i m not sure if jooq is expected to honor it or not steps to reproduce the problem long start system currenttimemillis tx dsl select stuff id from stuff limit fetchsize stream iterator hangs out here while fetching the entire data set next system out println first result after system currenttimemillis start jooq version jooq database product and version postgresql xx java version openjdk os version macos jdbc driver name and version include name if unofficial driver org postgresql postgresql
| 1
|
63,671
| 17,856,088,636
|
IssuesEvent
|
2021-09-05 03:29:09
|
thomasrussellmurphy/istyle-verilog-formatter
|
https://api.github.com/repos/thomasrussellmurphy/istyle-verilog-formatter
|
closed
|
Formatter does not handle comments in method declaration properly
|
Priority-Medium Type-Defect auto-migrated
|
```
What steps will reproduce the problem?
1. run `iStyle --indent=tab --style=kr --suffix=".old~" --pad=all DE0_TOP.v`
2. Examine comments in module declaration
3. Find that they are changed
What is the expected output? What do you see instead?
The expected output is no change in the file as it has already been formatted.
However, in the module declaration, the text like `CLOCK_50, // 50 MHz` becomes
`CLOCK_50, // 50 MHz`. That is, there is a space added between the comma and
the comment.
What version of the product are you using? On what operating system?
iStyle 1.20 on MacOS 10.9.2
Please provide any additional information below.
```
Original issue reported on code.google.com by `thomasmu...@gmail.com` on 11 Apr 2014 at 1:49
Attachments:
- [DE0_TOP.v](https://storage.googleapis.com/google-code-attachments/istyle-verilog-formatter/issue-1/comment-0/DE0_TOP.v)
|
1.0
|
Formatter does not handle comments in method declaration properly - ```
What steps will reproduce the problem?
1. run `iStyle --indent=tab --style=kr --suffix=".old~" --pad=all DE0_TOP.v`
2. Examine comments in module declaration
3. Find that they are changed
What is the expected output? What do you see instead?
The expected output is no change in the file as it has already been formatted.
However, in the module declaration, the text like `CLOCK_50, // 50 MHz` becomes
`CLOCK_50, // 50 MHz`. That is, there is a space added between the comma and
the comment.
What version of the product are you using? On what operating system?
iStyle 1.20 on MacOS 10.9.2
Please provide any additional information below.
```
Original issue reported on code.google.com by `thomasmu...@gmail.com` on 11 Apr 2014 at 1:49
Attachments:
- [DE0_TOP.v](https://storage.googleapis.com/google-code-attachments/istyle-verilog-formatter/issue-1/comment-0/DE0_TOP.v)
|
defect
|
formatter does not handle comments in method declaration properly what steps will reproduce the problem run istyle indent tab style kr suffix old pad all top v examine comments in module declaration find that they are changed what is the expected output what do you see instead the expected output is no change in the file as it has already been formatted however in the module declaration the text like clock mhz becomes clock mhz that is there is a space added between the comma and the comment what version of the product are you using on what operating system istyle on macos please provide any additional information below original issue reported on code google com by thomasmu gmail com on apr at attachments
| 1
|
17,505
| 3,010,640,396
|
IssuesEvent
|
2015-07-28 14:14:04
|
AtlasOfLivingAustralia/biocache-store
|
https://api.github.com/repos/AtlasOfLivingAustralia/biocache-store
|
closed
|
Incorrect taxon matching for SAM records (ignoring higher taxa hint from collectory)
|
bug priority-medium status-new type-defect
|
_From @nickdos on August 19, 2014 13:26_
*migrated from:* https://code.google.com/p/ala/issues/detail?id=722
*date:* Sun Jun 29 17:55:32 2014
*author:* nickdos
---
Alexis identified some records being matched to non-animal higher taxa despite the collectory specifying animalia as the higher taxon hint for this collection.
[http://biocache.ala.org.au/occurrences/5bcd1ec2-211c-4bf6-9993-414d9e815898](http://biocache.ala.org.au/occurrences/5bcd1ec2-211c-4bf6-9993-414d9e815898)
and
[http://biocache.ala.org.au/occurrences/d6fb923e-6c2c-4937-8e2a-08b9187e252b](http://biocache.ala.org.au/occurrences/d6fb923e-6c2c-4937-8e2a-08b9187e252b)
[http://biocache.ala.org.au/occurrences/100c8df7-de72-432c-90ce-9dbe4bf5113d](http://biocache.ala.org.au/occurrences/100c8df7-de72-432c-90ce-9dbe4bf5113d)
[http://biocache.ala.org.au/occurrences/ed6cb62d-3f8d-4cd9-900e-a8d5d7e13abd](http://biocache.ala.org.au/occurrences/ed6cb62d-3f8d-4cd9-900e-a8d5d7e13abd)
which appear to be matching to genus via a fuzzy match despite having a family specified which appears to have been ignored.
_Copied from original issue: AtlasOfLivingAustralia/biocache-hubs#85_
|
1.0
|
Incorrect taxon matching for SAM records (ignoring higher taxa hint from collectory) - _From @nickdos on August 19, 2014 13:26_
*migrated from:* https://code.google.com/p/ala/issues/detail?id=722
*date:* Sun Jun 29 17:55:32 2014
*author:* nickdos
---
Alexis identified some records being matched to non-animal higher taxa despite the collectory specifying animalia as the higher taxon hint for this collection.
[http://biocache.ala.org.au/occurrences/5bcd1ec2-211c-4bf6-9993-414d9e815898](http://biocache.ala.org.au/occurrences/5bcd1ec2-211c-4bf6-9993-414d9e815898)
and
[http://biocache.ala.org.au/occurrences/d6fb923e-6c2c-4937-8e2a-08b9187e252b](http://biocache.ala.org.au/occurrences/d6fb923e-6c2c-4937-8e2a-08b9187e252b)
[http://biocache.ala.org.au/occurrences/100c8df7-de72-432c-90ce-9dbe4bf5113d](http://biocache.ala.org.au/occurrences/100c8df7-de72-432c-90ce-9dbe4bf5113d)
[http://biocache.ala.org.au/occurrences/ed6cb62d-3f8d-4cd9-900e-a8d5d7e13abd](http://biocache.ala.org.au/occurrences/ed6cb62d-3f8d-4cd9-900e-a8d5d7e13abd)
which appear to be matching to genus via a fuzzy match despite having a family specified which appears to have been ignored.
_Copied from original issue: AtlasOfLivingAustralia/biocache-hubs#85_
|
defect
|
incorrect taxon matching for sam records ignoring higher taxa hint from collectory from nickdos on august migrated from date sun jun author nickdos alexis identified some records being matched to non animal higher taxa despite the collectory specifying animalia as the higher taxon hint for this collection and which appear to be matching to genus via a fuzzy match despite having a family specified which appears to have been ignored copied from original issue atlasoflivingaustralia biocache hubs
| 1
|
3,426
| 2,610,062,545
|
IssuesEvent
|
2015-02-26 18:18:23
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
黄岩治疗不育哪里最好
|
auto-migrated Priority-Medium Type-Defect
|
```
黄岩治疗不育哪里最好【台州五洲生殖医院】24小时健康咨询
热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市
椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、1
18、198及椒江一金清公交车直达枫南小区,乘坐107、105、109、
112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:38
|
1.0
|
黄岩治疗不育哪里最好 - ```
黄岩治疗不育哪里最好【台州五洲生殖医院】24小时健康咨询
热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市
椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、1
18、198及椒江一金清公交车直达枫南小区,乘坐107、105、109、
112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:38
|
defect
|
黄岩治疗不育哪里最好 黄岩治疗不育哪里最好【台州五洲生殖医院】 热线 微信号tzwzszyy 医院地址 台州市 (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 1
|
27,165
| 4,891,432,222
|
IssuesEvent
|
2016-11-18 16:40:21
|
BugBuster1701/visitors
|
https://api.github.com/repos/BugBuster1701/visitors
|
closed
|
Duplicate entry in tl_search Problem bei Nachrichtenbeitrag Zählung
|
2 - Working <= 5 Defect frontend
|
Die `$objPage->id` zu ändern, um die Beiträge statt die Seite zu zählen, ist wohl nicht so günstig.
Entweder am Ende die Originale ID wieder herstellen oder eigene Variable verwenden.
|
1.0
|
Duplicate entry in tl_search Problem bei Nachrichtenbeitrag Zählung - Die `$objPage->id` zu ändern, um die Beiträge statt die Seite zu zählen, ist wohl nicht so günstig.
Entweder am Ende die Originale ID wieder herstellen oder eigene Variable verwenden.
|
defect
|
duplicate entry in tl search problem bei nachrichtenbeitrag zählung die objpage id zu ändern um die beiträge statt die seite zu zählen ist wohl nicht so günstig entweder am ende die originale id wieder herstellen oder eigene variable verwenden
| 1
|
437,061
| 12,559,621,609
|
IssuesEvent
|
2020-06-07 19:31:10
|
kubeflow/examples
|
https://api.github.com/repos/kubeflow/examples
|
closed
|
Periodic notebook tests aren't running; can't get credentials for auto-deployed cluster
|
area/engprod kind/bug lifecycle/stale priority/p0
|
Looking at
https://k8s-testgrid.appspot.com/sig-big-data#kubeflow-periodic-examples
Looks like the periodic tests aren't running.
Example workflow
http://testing-argo.kubeflow.org/workflows/kubeflow-test-infra/kubeflow-periodic-examples-notebooks-v1-4418-ad41?tab=workflow&nodeId=kubeflow-periodic-examples-notebooks-v1-4418-ad41-3089756459
Looks like get-credentials step is failing.
There was a problem recently with being out of quota.
kubeflow/testing#616
But that should be fixed but its possible we don't have any valid auto deployed clusters.
No endpoints are listed in:
https://kf-ci-v1.endpoints.kubeflow-ci.cloud.goog/auto_deploy/
/assign @gabrielwen
|
1.0
|
Periodic notebook tests aren't running; can't get credentials for auto-deployed cluster - Looking at
https://k8s-testgrid.appspot.com/sig-big-data#kubeflow-periodic-examples
Looks like the periodic tests aren't running.
Example workflow
http://testing-argo.kubeflow.org/workflows/kubeflow-test-infra/kubeflow-periodic-examples-notebooks-v1-4418-ad41?tab=workflow&nodeId=kubeflow-periodic-examples-notebooks-v1-4418-ad41-3089756459
Looks like get-credentials step is failing.
There was a problem recently with being out of quota.
kubeflow/testing#616
But that should be fixed but its possible we don't have any valid auto deployed clusters.
No endpoints are listed in:
https://kf-ci-v1.endpoints.kubeflow-ci.cloud.goog/auto_deploy/
/assign @gabrielwen
|
non_defect
|
periodic notebook tests aren t running can t get credentials for auto deployed cluster looking at looks like the periodic tests aren t running example workflow looks like get credentials step is failing there was a problem recently with being out of quota kubeflow testing but that should be fixed but its possible we don t have any valid auto deployed clusters no endpoints are listed in assign gabrielwen
| 0
|
123,852
| 12,220,108,465
|
IssuesEvent
|
2020-05-02 00:01:36
|
vtex-apps/io-documentation
|
https://api.github.com/repos/vtex-apps/io-documentation
|
opened
|
vtex-apps/store-notifications has no documentation yet
|
no-documentation
|
[vtex-apps/store-notifications](https://github.com/vtex-apps/store-notifications) hasn't created any README file yet or is not using Docs Builder
|
1.0
|
vtex-apps/store-notifications has no documentation yet - [vtex-apps/store-notifications](https://github.com/vtex-apps/store-notifications) hasn't created any README file yet or is not using Docs Builder
|
non_defect
|
vtex apps store notifications has no documentation yet hasn t created any readme file yet or is not using docs builder
| 0
|
565,878
| 16,771,878,919
|
IssuesEvent
|
2021-06-14 15:40:57
|
mozilla/addons-server
|
https://api.github.com/repos/mozilla/addons-server
|
closed
|
Missing UI to change add-on visibility when the add-on no longer has listed versions
|
component: devhub priority: p3
|
### Describe the problem and steps to reproduce it:
Addon with only one build set to invisible. Deleted the build and uploaded new one.
### What happened?
After delete last (invisible) build and uploading new one, I can't change the visibility status of new builds. Alson while uploading the new build the page warned me that the addon was invisible but there was no previous build (because a deleted it) and didn't ask me the new status.
### What did you expect to happen?
When uploading the new build should be asked the visibility status. When uploaded should be able to change visbility staus through the interface.
### Anything else we should know?
(Please include a link to the page, screenshots and any relevant files.)
|
1.0
|
Missing UI to change add-on visibility when the add-on no longer has listed versions - ### Describe the problem and steps to reproduce it:
Addon with only one build set to invisible. Deleted the build and uploaded new one.
### What happened?
After delete last (invisible) build and uploading new one, I can't change the visibility status of new builds. Alson while uploading the new build the page warned me that the addon was invisible but there was no previous build (because a deleted it) and didn't ask me the new status.
### What did you expect to happen?
When uploading the new build should be asked the visibility status. When uploaded should be able to change visbility staus through the interface.
### Anything else we should know?
(Please include a link to the page, screenshots and any relevant files.)
|
non_defect
|
missing ui to change add on visibility when the add on no longer has listed versions describe the problem and steps to reproduce it addon with only one build set to invisible deleted the build and uploaded new one what happened after delete last invisible build and uploading new one i can t change the visibility status of new builds alson while uploading the new build the page warned me that the addon was invisible but there was no previous build because a deleted it and didn t ask me the new status what did you expect to happen when uploading the new build should be asked the visibility status when uploaded should be able to change visbility staus through the interface anything else we should know please include a link to the page screenshots and any relevant files
| 0
|
53,171
| 13,261,088,300
|
IssuesEvent
|
2020-08-20 19:16:11
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
Primary field wrong for datasets 9122/9112 (Trac #889)
|
Migrated from Trac defect iceprod
|
The CORSIKA-ice-top dataset 9166 and its parent (9122) have CORSIKA::primary set to Fe, which is a lie. It should be corrected to H.
You can eventually get to a set with the field set to H (9112), but by following the chain of input_dataset, but it's unclear where to stop, as the next set in the chain (5048) is also wrong.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/889">https://code.icecube.wisc.edu/projects/icecube/ticket/889</a>, reported by jvansantenand owned by jgonzalez</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-10-07T14:46:40",
"_ts": "1444229200470219",
"description": "The CORSIKA-ice-top dataset 9166 and its parent (9122) have CORSIKA::primary set to Fe, which is a lie. It should be corrected to H.\n\nYou can eventually get to a set with the field set to H (9112), but by following the chain of input_dataset, but it's unclear where to stop, as the next set in the chain (5048) is also wrong.",
"reporter": "jvansanten",
"cc": "frank.mcnally",
"resolution": "fixed",
"time": "2015-03-03T08:26:29",
"component": "iceprod",
"summary": "Primary field wrong for datasets 9122/9112",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jgonzalez",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Primary field wrong for datasets 9122/9112 (Trac #889) - The CORSIKA-ice-top dataset 9166 and its parent (9122) have CORSIKA::primary set to Fe, which is a lie. It should be corrected to H.
You can eventually get to a set with the field set to H (9112), but by following the chain of input_dataset, but it's unclear where to stop, as the next set in the chain (5048) is also wrong.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/889">https://code.icecube.wisc.edu/projects/icecube/ticket/889</a>, reported by jvansantenand owned by jgonzalez</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-10-07T14:46:40",
"_ts": "1444229200470219",
"description": "The CORSIKA-ice-top dataset 9166 and its parent (9122) have CORSIKA::primary set to Fe, which is a lie. It should be corrected to H.\n\nYou can eventually get to a set with the field set to H (9112), but by following the chain of input_dataset, but it's unclear where to stop, as the next set in the chain (5048) is also wrong.",
"reporter": "jvansanten",
"cc": "frank.mcnally",
"resolution": "fixed",
"time": "2015-03-03T08:26:29",
"component": "iceprod",
"summary": "Primary field wrong for datasets 9122/9112",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jgonzalez",
"type": "defect"
}
```
</p>
</details>
|
defect
|
primary field wrong for datasets trac the corsika ice top dataset and its parent have corsika primary set to fe which is a lie it should be corrected to h you can eventually get to a set with the field set to h but by following the chain of input dataset but it s unclear where to stop as the next set in the chain is also wrong migrated from json status closed changetime ts description the corsika ice top dataset and its parent have corsika primary set to fe which is a lie it should be corrected to h n nyou can eventually get to a set with the field set to h but by following the chain of input dataset but it s unclear where to stop as the next set in the chain is also wrong reporter jvansanten cc frank mcnally resolution fixed time component iceprod summary primary field wrong for datasets priority normal keywords milestone owner jgonzalez type defect
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.