Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
679,878
23,248,022,672
IssuesEvent
2022-08-03 22:38:42
chaotic-aur/packages
https://api.github.com/repos/chaotic-aur/packages
closed
[Request] gnome-text-editor
request:new-pkg priority:low
### Link to the package(s) in the AUR https://aur.archlinux.org/packages/gnome-text-editor ### Utility this package has for you GNOME's next default text editor. ### Do you consider the package(s) to be useful for every Chaotic-AUR user? No, but for a great amount. ### Do you consider the package to be useful for feature testing/preview? - [ ] Yes ### Have you tested if the package builds in a clean chroot? - [ ] Yes ### Does the package's license allow redistributing it? YES! ### Have you searched the issues to ensure this request is unique? - [X] YES! ### Have you read the README to ensure this package is not banned? - [X] YES! ### More information There's `-git` version already, but people may prefer released versions instead of main branch.
1.0
[Request] gnome-text-editor - ### Link to the package(s) in the AUR https://aur.archlinux.org/packages/gnome-text-editor ### Utility this package has for you GNOME's next default text editor. ### Do you consider the package(s) to be useful for every Chaotic-AUR user? No, but for a great amount. ### Do you consider the package to be useful for feature testing/preview? - [ ] Yes ### Have you tested if the package builds in a clean chroot? - [ ] Yes ### Does the package's license allow redistributing it? YES! ### Have you searched the issues to ensure this request is unique? - [X] YES! ### Have you read the README to ensure this package is not banned? - [X] YES! ### More information There's `-git` version already, but people may prefer released versions instead of main branch.
priority
gnome text editor link to the package s in the aur utility this package has for you gnome s next default text editor do you consider the package s to be useful for every chaotic aur user no but for a great amount do you consider the package to be useful for feature testing preview yes have you tested if the package builds in a clean chroot yes does the package s license allow redistributing it yes have you searched the issues to ensure this request is unique yes have you read the readme to ensure this package is not banned yes more information there s git version already but people may prefer released versions instead of main branch
1
2,638
3,931,665,430
IssuesEvent
2016-04-25 13:22:38
hyperledger/fabric
https://api.github.com/repos/hyperledger/fabric
closed
non-validating peer fails with 'Not Implemented' when trying to sign HELLO msg
blocker security
The call in core/Peer below fails with 'Not Implemented' error. if viper.GetBool("security.enabled") { sig, err := p.secHelper.Sign(msg.Payload) if err != nil { return fmt.Errorf("Error signing Openchain Message: %s", err) } // Set the signature in the message msg.Signature = sig } return nil
True
non-validating peer fails with 'Not Implemented' when trying to sign HELLO msg - The call in core/Peer below fails with 'Not Implemented' error. if viper.GetBool("security.enabled") { sig, err := p.secHelper.Sign(msg.Payload) if err != nil { return fmt.Errorf("Error signing Openchain Message: %s", err) } // Set the signature in the message msg.Signature = sig } return nil
non_priority
non validating peer fails with not implemented when trying to sign hello msg the call in core peer below fails with not implemented error if viper getbool security enabled sig err p sechelper sign msg payload if err nil return fmt errorf error signing openchain message s err set the signature in the message msg signature sig return nil
0
19,750
4,442,032,385
IssuesEvent
2016-08-19 11:51:05
coala-analyzer/coala
https://api.github.com/repos/coala-analyzer/coala
closed
all languages prototype - possibly remove or reorder
area/documentation difficulty/low
From [Supported Languages](https://github.com/coala-analyzer/coala-bears/wiki/Supported-languages) I wanted to learn more so I clicked on the first link under Home on the right. It took me to [all languages prototype](https://github.com/coala-analyzer/coala-bears/wiki/all-languages-prototype) which had the same information in a different layout from what I was looking at. I then noticed that the order on the right was ordered in a way where the flow would be better starting at the bottom and moving up the links. Consider reordering
1.0
all languages prototype - possibly remove or reorder - From [Supported Languages](https://github.com/coala-analyzer/coala-bears/wiki/Supported-languages) I wanted to learn more so I clicked on the first link under Home on the right. It took me to [all languages prototype](https://github.com/coala-analyzer/coala-bears/wiki/all-languages-prototype) which had the same information in a different layout from what I was looking at. I then noticed that the order on the right was ordered in a way where the flow would be better starting at the bottom and moving up the links. Consider reordering
non_priority
all languages prototype possibly remove or reorder from i wanted to learn more so i clicked on the first link under home on the right it took me to which had the same information in a different layout from what i was looking at i then noticed that the order on the right was ordered in a way where the flow would be better starting at the bottom and moving up the links consider reordering
0
69,417
7,133,829,120
IssuesEvent
2018-01-22 18:46:05
letsencrypt/boulder
https://api.github.com/repos/letsencrypt/boulder
opened
Remove config-next gating on some unittests
kind/tests
Some of our unittests were gated on config-next because they relied on migrations from _db-next. Now that all of those migrations have been moved to _db, we should remove the gating on the unittests.
1.0
Remove config-next gating on some unittests - Some of our unittests were gated on config-next because they relied on migrations from _db-next. Now that all of those migrations have been moved to _db, we should remove the gating on the unittests.
non_priority
remove config next gating on some unittests some of our unittests were gated on config next because they relied on migrations from db next now that all of those migrations have been moved to db we should remove the gating on the unittests
0
181,859
6,664,824,669
IssuesEvent
2017-10-02 21:38:33
MorpheusXAUT/slotlist-backend
https://api.github.com/repos/MorpheusXAUT/slotlist-backend
opened
Add number of open slots to public mission data
enhancement priority/normal
Buur suggested adding the number of remaining open slots might be useful in the mission overview list. ----- ### Tasks - [ ] Add number of open slots to data returned in mission list
1.0
Add number of open slots to public mission data - Buur suggested adding the number of remaining open slots might be useful in the mission overview list. ----- ### Tasks - [ ] Add number of open slots to data returned in mission list
priority
add number of open slots to public mission data buur suggested adding the number of remaining open slots might be useful in the mission overview list tasks add number of open slots to data returned in mission list
1
329,179
28,181,389,704
IssuesEvent
2023-04-04 02:57:06
MultiProject23/Shop_SemiProject
https://api.github.com/repos/MultiProject23/Shop_SemiProject
closed
[BE] 도메인 테스트 코드 작성
✅ test 🌱 back-end
# 개요 도메인 계층의 테스트 코드를 작성한다. ## 할 일 - `review` 작성 - `member` 작성 - `cart` 작성
1.0
[BE] 도메인 테스트 코드 작성 - # 개요 도메인 계층의 테스트 코드를 작성한다. ## 할 일 - `review` 작성 - `member` 작성 - `cart` 작성
non_priority
 도메인 테스트 코드 작성 개요 도메인 계층의 테스트 코드를 작성한다 할 일 review 작성 member 작성 cart 작성
0
295,663
25,492,860,281
IssuesEvent
2022-11-27 09:54:32
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test: TestSqlLiteLogic_testindexorderby_nosort1000slt_good_0_test failed
C-test-failure O-robot branch-master
pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test.TestSqlLiteLogic_testindexorderby_nosort1000slt_good_0_test [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_SQLiteLogicTestsBazel/7709095?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_SQLiteLogicTestsBazel/7709095?buildTab=artifacts#/) on master @ [1a6e9f885baa124d5ff2996adb966ea15a1a9b2b](https://github.com/cockroachdb/cockroach/commits/1a6e9f885baa124d5ff2996adb966ea15a1a9b2b): Fatal error: ``` panic: test timed out after 1h59m55s ``` Stack: ``` goroutine 17067675 [running]: testing.(*M).startAlarm.func1() GOROOT/src/testing/testing.go:2036 +0x8e created by time.goFunc GOROOT/src/time/sleep.go:176 +0x32 ``` <details><summary>Log preceding fatal error</summary> <p> ``` [09:44:10] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1388 statements [09:44:12] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1393 statements [09:44:15] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1398 statements [09:44:17] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1401 statements [09:44:19] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1404 statements [09:44:21] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1408 statements [09:44:23] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1413 statements [09:44:26] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1418 statements [09:44:28] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1423 statements [09:44:31] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1426 statements [09:44:33] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1430 statements [09:44:35] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1435 statements [09:44:38] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1441 statements [09:44:40] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1451 statements [09:44:42] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1455 statements [09:44:45] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1460 statements [09:44:48] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1465 statements [09:44:51] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1470 statements [09:44:53] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1474 statements [09:44:55] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1478 statements [09:44:57] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1482 statements [09:45:00] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1487 statements [09:45:02] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1490 statements [09:45:04] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1495 statements [09:45:07] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1500 statements [09:45:09] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1508 statements [09:45:12] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1515 statements [09:45:14] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1518 statements [09:45:16] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1522 statements [09:45:19] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1527 statements [09:45:21] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1532 statements [09:45:23] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1537 statements [09:45:25] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1542 statements [09:45:27] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1548 statements [09:45:30] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1555 statements [09:45:32] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1562 statements [09:45:34] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1576 statements [09:45:36] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1591 statements [09:45:38] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1606 statements [09:45:40] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1616 statements [09:45:42] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1628 statements [09:45:45] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1633 statements [09:45:47] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1640 statements [09:45:49] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1646 statements [09:45:51] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1653 statements [09:45:53] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1660 statements [09:45:55] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1667 statements [09:45:58] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1674 statements [09:46:00] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1684 statements [09:46:02] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1701 statements ``` </p> </details> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSqlLiteLogic_testindexorderby_nosort1000slt_good_0_test.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test: TestSqlLiteLogic_testindexorderby_nosort1000slt_good_0_test failed - pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test.TestSqlLiteLogic_testindexorderby_nosort1000slt_good_0_test [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_SQLiteLogicTestsBazel/7709095?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_SQLiteLogicTestsBazel/7709095?buildTab=artifacts#/) on master @ [1a6e9f885baa124d5ff2996adb966ea15a1a9b2b](https://github.com/cockroachdb/cockroach/commits/1a6e9f885baa124d5ff2996adb966ea15a1a9b2b): Fatal error: ``` panic: test timed out after 1h59m55s ``` Stack: ``` goroutine 17067675 [running]: testing.(*M).startAlarm.func1() GOROOT/src/testing/testing.go:2036 +0x8e created by time.goFunc GOROOT/src/time/sleep.go:176 +0x32 ``` <details><summary>Log preceding fatal error</summary> <p> ``` [09:44:10] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1388 statements [09:44:12] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1393 statements [09:44:15] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1398 statements [09:44:17] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1401 statements [09:44:19] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1404 statements [09:44:21] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1408 statements [09:44:23] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1413 statements [09:44:26] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1418 statements [09:44:28] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1423 statements [09:44:31] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1426 statements [09:44:33] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1430 statements [09:44:35] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1435 statements [09:44:38] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1441 statements [09:44:40] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1451 statements [09:44:42] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1455 statements [09:44:45] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1460 statements [09:44:48] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1465 statements [09:44:51] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1470 statements [09:44:53] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1474 statements [09:44:55] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1478 statements [09:44:57] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1482 statements [09:45:00] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1487 statements [09:45:02] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1490 statements [09:45:04] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1495 statements [09:45:07] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1500 statements [09:45:09] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1508 statements [09:45:12] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1515 statements [09:45:14] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1518 statements [09:45:16] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1522 statements [09:45:19] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1527 statements [09:45:21] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1532 statements [09:45:23] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1537 statements [09:45:25] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1542 statements [09:45:27] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1548 statements [09:45:30] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1555 statements [09:45:32] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1562 statements [09:45:34] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1576 statements [09:45:36] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1591 statements [09:45:38] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1606 statements [09:45:40] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1616 statements [09:45:42] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1628 statements [09:45:45] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1633 statements [09:45:47] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1640 statements [09:45:49] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1646 statements [09:45:51] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1653 statements [09:45:53] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1660 statements [09:45:55] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1667 statements [09:45:58] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1674 statements [09:46:00] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1684 statements [09:46:02] --- progress: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/3327/execroot/com_github_cockroachdb_cockroach/bazel-out/k8-fastbuild/bin/pkg/sql/sqlitelogictest/tests/fakedist/fakedist_test_/fakedist_test.runfiles/com_github_cockroachdb_cockroach/external/com_github_cockroachdb_sqllogictest/test/index/orderby_nosort/1000/slt_good_0.test: 1701 statements ``` </p> </details> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestSqlLiteLogic_testindexorderby_nosort1000slt_good_0_test.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_priority
pkg sql sqlitelogictest tests fakedist fakedist test testsqllitelogic testindexorderby good test failed pkg sql sqlitelogictest tests fakedist fakedist test testsqllitelogic testindexorderby good test with on master fatal error panic test timed out after stack goroutine testing m startalarm goroot src testing testing go created by time gofunc goroot src time sleep go log preceding fatal error progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements progress home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql sqlitelogictest tests fakedist fakedist test fakedist test runfiles com github cockroachdb cockroach external com github cockroachdb sqllogictest test index orderby nosort slt good test statements help see also cc cockroachdb sql queries
0
238,955
7,784,847,760
IssuesEvent
2018-06-06 14:22:10
salesagility/SuiteCRM
https://api.github.com/repos/salesagility/SuiteCRM
closed
Single Opt In status not being saved
High Priority Resolved: Next Release bug category:emails
<!--- Provide a general summary of the issue in the **Title** above --> <!--- Before you open an issue, please check if a similar issue already exists or has been closed before. ---> <!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. ---> #### Issue <!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug --> Single opt in status not being saved when a record is created or update on the system. If a record is being updated, for example in the contacts module. and an email address is added or already exists, and the user selects the 'Opt In' check box, this status will not save. This issue also therefor impacts the ability to send Confirm Opt In bulk action emails within the list views. #### Expected Behavior <!--- Tell us what should happen --> When the 'Opt In' checkbox is selected, this status should save and remain persistent, and also be visible with the detail view, list view, edit view etc. #### Actual Behavior <!--- Tell us what happens instead --> <!--- Also please check relevant logs (suitecrm.log, php error.log etc.) --> The status for being 'Opt In' is not saved, and therefor the record is not stored as being opt in, and within the edit view, the box is unchecked as before. #### Possible Fix <!--- Not obligatory, but suggest a fix or reason for the bug --> #### Steps to Reproduce <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug include code to reproduce, if relevant --> 1. Have the system set up with confirmed opt in functionality, 1.1. Navigate to the admin panel, 1.2. Select 'Email Settings', 1.3. Under the 'Email Options' panel, use the drop down box for 'Opt In Settings' and select Confirmed Opt In, 1.4. Save the settings, 2. Create a record or modify an existing record inside the Contacts module, can also be done with Targets and Leads, 3. Beside the 'Email Address' field, select 'Opted In'. Save this record, 4. See that in the detail view, it is not shown that the record is Opt In, 5. Navigate to the list view. See that there is no indication of this record being 'Opt In', 6. If you navigate back to the edit view, the 'Opt In' checkbox will be unchecked. #### Context <!--- How has this bug affected you? What were you trying to accomplish? --> <!--- If you feel this should be a low/medium/high priority then please state so --> This impacts the 'Opt In functionality within the system. The bulk actions 'Send confirm Opt In' will not be possible, any way to find opt in records will not be possible. #### Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * SuiteCRM Version used: 7.10.4 * Browser name and version (e.g. Chrome Version 51.0.2704.63 (64-bit)): Chromium * Environment name and version (e.g. MySQL, PHP 7): PHP 7.1 * Operating System and version (e.g Ubuntu 16.04): Ubuntu 16.04
1.0
Single Opt In status not being saved - <!--- Provide a general summary of the issue in the **Title** above --> <!--- Before you open an issue, please check if a similar issue already exists or has been closed before. ---> <!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. ---> #### Issue <!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug --> Single opt in status not being saved when a record is created or update on the system. If a record is being updated, for example in the contacts module. and an email address is added or already exists, and the user selects the 'Opt In' check box, this status will not save. This issue also therefor impacts the ability to send Confirm Opt In bulk action emails within the list views. #### Expected Behavior <!--- Tell us what should happen --> When the 'Opt In' checkbox is selected, this status should save and remain persistent, and also be visible with the detail view, list view, edit view etc. #### Actual Behavior <!--- Tell us what happens instead --> <!--- Also please check relevant logs (suitecrm.log, php error.log etc.) --> The status for being 'Opt In' is not saved, and therefor the record is not stored as being opt in, and within the edit view, the box is unchecked as before. #### Possible Fix <!--- Not obligatory, but suggest a fix or reason for the bug --> #### Steps to Reproduce <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug include code to reproduce, if relevant --> 1. Have the system set up with confirmed opt in functionality, 1.1. Navigate to the admin panel, 1.2. Select 'Email Settings', 1.3. Under the 'Email Options' panel, use the drop down box for 'Opt In Settings' and select Confirmed Opt In, 1.4. Save the settings, 2. Create a record or modify an existing record inside the Contacts module, can also be done with Targets and Leads, 3. Beside the 'Email Address' field, select 'Opted In'. Save this record, 4. See that in the detail view, it is not shown that the record is Opt In, 5. Navigate to the list view. See that there is no indication of this record being 'Opt In', 6. If you navigate back to the edit view, the 'Opt In' checkbox will be unchecked. #### Context <!--- How has this bug affected you? What were you trying to accomplish? --> <!--- If you feel this should be a low/medium/high priority then please state so --> This impacts the 'Opt In functionality within the system. The bulk actions 'Send confirm Opt In' will not be possible, any way to find opt in records will not be possible. #### Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * SuiteCRM Version used: 7.10.4 * Browser name and version (e.g. Chrome Version 51.0.2704.63 (64-bit)): Chromium * Environment name and version (e.g. MySQL, PHP 7): PHP 7.1 * Operating System and version (e.g Ubuntu 16.04): Ubuntu 16.04
priority
single opt in status not being saved issue single opt in status not being saved when a record is created or update on the system if a record is being updated for example in the contacts module and an email address is added or already exists and the user selects the opt in check box this status will not save this issue also therefor impacts the ability to send confirm opt in bulk action emails within the list views expected behavior when the opt in checkbox is selected this status should save and remain persistent and also be visible with the detail view list view edit view etc actual behavior the status for being opt in is not saved and therefor the record is not stored as being opt in and within the edit view the box is unchecked as before possible fix steps to reproduce have the system set up with confirmed opt in functionality navigate to the admin panel select email settings under the email options panel use the drop down box for opt in settings and select confirmed opt in save the settings create a record or modify an existing record inside the contacts module can also be done with targets and leads beside the email address field select opted in save this record see that in the detail view it is not shown that the record is opt in navigate to the list view see that there is no indication of this record being opt in if you navigate back to the edit view the opt in checkbox will be unchecked context this impacts the opt in functionality within the system the bulk actions send confirm opt in will not be possible any way to find opt in records will not be possible your environment suitecrm version used browser name and version e g chrome version bit chromium environment name and version e g mysql php php operating system and version e g ubuntu ubuntu
1
602,091
18,450,178,323
IssuesEvent
2021-10-15 09:32:43
gambitph/Stackable
https://api.github.com/repos/gambitph/Stackable
closed
Changing block alignment adds space below the image in horizontal Card (Backend)
bug high priority [version] V3 [block] Card
<!-- Before posting, make sure that: 1. you are running the latest version of Stackable, and 2. you have searched whether your issue has already been reported --> Steps: 1. Add a v3 Card > horizontal layout 2. upload an image 3. change the block alignment to align center / wide / full 4. see bug in backend (this is ok in frontend) https://user-images.githubusercontent.com/28699204/137095128-4b512cea-911a-493c-8505-faab9329376e.mov
1.0
Changing block alignment adds space below the image in horizontal Card (Backend) - <!-- Before posting, make sure that: 1. you are running the latest version of Stackable, and 2. you have searched whether your issue has already been reported --> Steps: 1. Add a v3 Card > horizontal layout 2. upload an image 3. change the block alignment to align center / wide / full 4. see bug in backend (this is ok in frontend) https://user-images.githubusercontent.com/28699204/137095128-4b512cea-911a-493c-8505-faab9329376e.mov
priority
changing block alignment adds space below the image in horizontal card backend before posting make sure that you are running the latest version of stackable and you have searched whether your issue has already been reported steps add a card horizontal layout upload an image change the block alignment to align center wide full see bug in backend this is ok in frontend
1
449,635
12,972,633,488
IssuesEvent
2020-07-21 12:51:50
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
reopened
[Devportal] UX issue when editing an application
Priority/Normal Store Type/Bug Type/React-UI
### Description: ![appdevporta](https://user-images.githubusercontent.com/3313885/86484465-e5a66f00-bd73-11ea-8cd4-ee1a80dc632d.gif) I think this is happening because we have used same edit page for application create, application edit from app listing page and edit from individual application details UI ### Steps to reproduce: ### Affected Product Version: <!-- Members can use Affected/*** labels --> ### Environment details (with versions): - OS: - Client: - Env (Docker/K8s): --- ### Optional Fields #### Related Issues: <!-- Any related issues from this/other repositories--> #### Suggested Labels: <!--Only to be used by non-members--> #### Suggested Assignees: <!--Only to be used by non-members-->
1.0
[Devportal] UX issue when editing an application - ### Description: ![appdevporta](https://user-images.githubusercontent.com/3313885/86484465-e5a66f00-bd73-11ea-8cd4-ee1a80dc632d.gif) I think this is happening because we have used same edit page for application create, application edit from app listing page and edit from individual application details UI ### Steps to reproduce: ### Affected Product Version: <!-- Members can use Affected/*** labels --> ### Environment details (with versions): - OS: - Client: - Env (Docker/K8s): --- ### Optional Fields #### Related Issues: <!-- Any related issues from this/other repositories--> #### Suggested Labels: <!--Only to be used by non-members--> #### Suggested Assignees: <!--Only to be used by non-members-->
priority
ux issue when editing an application description i think this is happening because we have used same edit page for application create application edit from app listing page and edit from individual application details ui steps to reproduce affected product version environment details with versions os client env docker optional fields related issues suggested labels suggested assignees
1
677,205
23,154,869,230
IssuesEvent
2022-07-29 12:02:33
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[YSQL] Dropping a primary key constraint is not supported
kind/bug area/ysql priority/high
Jira Link: [DB-1677](https://yugabyte.atlassian.net/browse/DB-1677) Attempting dropping a primary key constraint results in an error: ``` Error: ERROR: dropping a primary key constraint is not yet supported [Failed SQL: ALTER TABLE public.TABLE_NAME DROP CONSTRAINT CONSTRAINT_2] ``` This is explicitly disabled in the code: https://github.com/yugabyte/yugabyte-db/blob/1a8f1611365a6c442d523483fd12e56f28009190/src/postgres/src/backend/commands/tablecmds.c#L10500. From what I hear, the reason is that the primary key is tied to the table's storage layout and this is remotely related to `ALTER TABLE adding a primary key` https://github.com/yugabyte/yugabyte-db/issues/1104. This seems like a reversal of https://github.com/yugabyte/yugabyte-db/commit/fd257e7faaf701439cdaf1ff936cc46fc0f8927d.
1.0
[YSQL] Dropping a primary key constraint is not supported - Jira Link: [DB-1677](https://yugabyte.atlassian.net/browse/DB-1677) Attempting dropping a primary key constraint results in an error: ``` Error: ERROR: dropping a primary key constraint is not yet supported [Failed SQL: ALTER TABLE public.TABLE_NAME DROP CONSTRAINT CONSTRAINT_2] ``` This is explicitly disabled in the code: https://github.com/yugabyte/yugabyte-db/blob/1a8f1611365a6c442d523483fd12e56f28009190/src/postgres/src/backend/commands/tablecmds.c#L10500. From what I hear, the reason is that the primary key is tied to the table's storage layout and this is remotely related to `ALTER TABLE adding a primary key` https://github.com/yugabyte/yugabyte-db/issues/1104. This seems like a reversal of https://github.com/yugabyte/yugabyte-db/commit/fd257e7faaf701439cdaf1ff936cc46fc0f8927d.
priority
dropping a primary key constraint is not supported jira link attempting dropping a primary key constraint results in an error error error dropping a primary key constraint is not yet supported this is explicitly disabled in the code from what i hear the reason is that the primary key is tied to the table s storage layout and this is remotely related to alter table adding a primary key this seems like a reversal of
1
350,114
10,478,471,840
IssuesEvent
2019-09-24 00:10:34
BCcampus/edehr
https://api.github.com/repos/BCcampus/edehr
closed
Generate external resources pages from input sheet
Effort - Medium Epic - Content Priority - High ~Feature
The Inputs spreadsheet contains the definitive information for the EHR app and its 39 pages of data, tables, forms and dialogs. Each page may have a form or one or more tables. These need to be displayed and, if the user is permitted, edited or added too. Each table has a related modal dialog. The task is to organize the spreadsheet so the raw information can be copied and pasted into a configuration file. Then we run a generating script that composes the page routing, menu system (nav panel), components, data structures, dialogs, tables, etc. See the files in the makeEHR subfolder).
1.0
Generate external resources pages from input sheet - The Inputs spreadsheet contains the definitive information for the EHR app and its 39 pages of data, tables, forms and dialogs. Each page may have a form or one or more tables. These need to be displayed and, if the user is permitted, edited or added too. Each table has a related modal dialog. The task is to organize the spreadsheet so the raw information can be copied and pasted into a configuration file. Then we run a generating script that composes the page routing, menu system (nav panel), components, data structures, dialogs, tables, etc. See the files in the makeEHR subfolder).
priority
generate external resources pages from input sheet the inputs spreadsheet contains the definitive information for the ehr app and its pages of data tables forms and dialogs each page may have a form or one or more tables these need to be displayed and if the user is permitted edited or added too each table has a related modal dialog the task is to organize the spreadsheet so the raw information can be copied and pasted into a configuration file then we run a generating script that composes the page routing menu system nav panel components data structures dialogs tables etc see the files in the makeehr subfolder
1
363,284
10,740,827,400
IssuesEvent
2019-10-29 18:58:04
jenkins-x/jx
https://api.github.com/repos/jenkins-x/jx
closed
jx repo BDD tests on jenkins should use tillerless
area/quality backlog kind/enhancement priority/important-longterm
right now they are using tiller which isn't our default
1.0
jx repo BDD tests on jenkins should use tillerless - right now they are using tiller which isn't our default
priority
jx repo bdd tests on jenkins should use tillerless right now they are using tiller which isn t our default
1
623,988
19,684,408,939
IssuesEvent
2022-01-11 20:18:30
GoogleChrome/lighthouse
https://api.github.com/repos/GoogleChrome/lighthouse
closed
Send reports to Lighthouse CI Server
pending-close needs-priority
**Feature request summary** Send reports generated programmatically by Lighthouse to a remote instance of Lighthouse CI Server. **What is the motivation or use case for changing this?** We are working with [playwright-lighthouse](https://github.com/abhinaba-ghosh/playwright-lighthouse) and we would like to throw the generated reports to a Lighthouse CI server. The aim is to have them published and to compare them between commits. **How is this beneficial to Lighthouse?** Lighthouse already generates HTML reports based on the test execution. It would be better to publish them and take advantage of the awesome Lighthouse CI Server features.
1.0
Send reports to Lighthouse CI Server - **Feature request summary** Send reports generated programmatically by Lighthouse to a remote instance of Lighthouse CI Server. **What is the motivation or use case for changing this?** We are working with [playwright-lighthouse](https://github.com/abhinaba-ghosh/playwright-lighthouse) and we would like to throw the generated reports to a Lighthouse CI server. The aim is to have them published and to compare them between commits. **How is this beneficial to Lighthouse?** Lighthouse already generates HTML reports based on the test execution. It would be better to publish them and take advantage of the awesome Lighthouse CI Server features.
priority
send reports to lighthouse ci server feature request summary send reports generated programmatically by lighthouse to a remote instance of lighthouse ci server what is the motivation or use case for changing this we are working with and we would like to throw the generated reports to a lighthouse ci server the aim is to have them published and to compare them between commits how is this beneficial to lighthouse lighthouse already generates html reports based on the test execution it would be better to publish them and take advantage of the awesome lighthouse ci server features
1
9,570
24,845,833,233
IssuesEvent
2022-10-26 15:48:17
MicrosoftDocs/architecture-center
https://api.github.com/repos/MicrosoftDocs/architecture-center
closed
DataProducts link is not working
doc-bug assigned-to-author triaged architecture-center/svc example-scenario/subsvc Pri2
[Enter feedback here] When clicking the link for 'data product' mentioned in the line 'Establish a data product architecture', it opens up a page with '404- Content not found'. ![image](https://user-images.githubusercontent.com/96313557/146596031-acd64bc5-7c37-4f25-bad9-b9906a030dca.png) --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 5f10895e-4e1d-1cdd-2d57-32880e0cc45a * Version Independent ID: 3518d0a5-ffd1-dfa7-6d6b-de2a9ef170d6 * Content: [Analytics end-to-end with Azure Synapse - Azure Example Scenarios](https://docs.microsoft.com/en-us/azure/architecture/example-scenario/dataplate2e/data-platform-end-to-end) * Content Source: [docs/example-scenario/dataplate2e/data-platform-end-to-end.yml](https://github.com/microsoftdocs/architecture-center/blob/main/docs/example-scenario/dataplate2e/data-platform-end-to-end.yml) * Service: **architecture-center** * Sub-service: **example-scenario** * GitHub Login: @fabragaMS * Microsoft Alias: **fabraga**
1.0
DataProducts link is not working - [Enter feedback here] When clicking the link for 'data product' mentioned in the line 'Establish a data product architecture', it opens up a page with '404- Content not found'. ![image](https://user-images.githubusercontent.com/96313557/146596031-acd64bc5-7c37-4f25-bad9-b9906a030dca.png) --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 5f10895e-4e1d-1cdd-2d57-32880e0cc45a * Version Independent ID: 3518d0a5-ffd1-dfa7-6d6b-de2a9ef170d6 * Content: [Analytics end-to-end with Azure Synapse - Azure Example Scenarios](https://docs.microsoft.com/en-us/azure/architecture/example-scenario/dataplate2e/data-platform-end-to-end) * Content Source: [docs/example-scenario/dataplate2e/data-platform-end-to-end.yml](https://github.com/microsoftdocs/architecture-center/blob/main/docs/example-scenario/dataplate2e/data-platform-end-to-end.yml) * Service: **architecture-center** * Sub-service: **example-scenario** * GitHub Login: @fabragaMS * Microsoft Alias: **fabraga**
non_priority
dataproducts link is not working when clicking the link for data product mentioned in the line establish a data product architecture it opens up a page with content not found document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service example scenario github login fabragams microsoft alias fabraga
0
189,770
22,047,114,973
IssuesEvent
2022-05-30 03:55:34
Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2022-0492
https://api.github.com/repos/Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2022-0492
closed
CVE-2020-14331 (Medium) detected in linuxlinux-4.19.88 - autoclosed
security vulnerability
## CVE-2020-14331 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.88</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2022-0492/commit/8d2169763c8858bce8d07fbb569f01ef9b30383b">8d2169763c8858bce8d07fbb569f01ef9b30383b</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/linux-4.19.72/drivers/video/console/vgacon.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/linux-4.19.72/drivers/video/console/vgacon.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the Linux kernel’s implementation of the invert video code on VGA consoles when a local attacker attempts to resize the console, calling an ioctl VT_RESIZE, which causes an out-of-bounds write to occur. This flaw allows a local user with access to the VGA console to crash the system, potentially escalating their privileges on the system. The highest threat from this vulnerability is to data confidentiality and integrity as well as system availability. <p>Publish Date: 2020-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14331>CVE-2020-14331</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14331">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14331</a></p> <p>Release Date: 2020-09-15</p> <p>Fix Resolution: v5.9-rc1,v4.14.194,v4.19.139,v5.4.58,v5.7.15</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-14331 (Medium) detected in linuxlinux-4.19.88 - autoclosed - ## CVE-2020-14331 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.88</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2022-0492/commit/8d2169763c8858bce8d07fbb569f01ef9b30383b">8d2169763c8858bce8d07fbb569f01ef9b30383b</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/linux-4.19.72/drivers/video/console/vgacon.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/linux-4.19.72/drivers/video/console/vgacon.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the Linux kernel’s implementation of the invert video code on VGA consoles when a local attacker attempts to resize the console, calling an ioctl VT_RESIZE, which causes an out-of-bounds write to occur. This flaw allows a local user with access to the VGA console to crash the system, potentially escalating their privileges on the system. The highest threat from this vulnerability is to data confidentiality and integrity as well as system availability. <p>Publish Date: 2020-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14331>CVE-2020-14331</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14331">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14331</a></p> <p>Release Date: 2020-09-15</p> <p>Fix Resolution: v5.9-rc1,v4.14.194,v4.19.139,v5.4.58,v5.7.15</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files linux drivers video console vgacon c linux drivers video console vgacon c vulnerability details a flaw was found in the linux kernel’s implementation of the invert video code on vga consoles when a local attacker attempts to resize the console calling an ioctl vt resize which causes an out of bounds write to occur this flaw allows a local user with access to the vga console to crash the system potentially escalating their privileges on the system the highest threat from this vulnerability is to data confidentiality and integrity as well as system availability publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
649,813
21,326,873,124
IssuesEvent
2022-04-18 00:46:00
napari/napari
https://api.github.com/repos/napari/napari
closed
Additive blending mode requires black canvas (broken in light theme)
bug severity-high priority-medium
## 🐛 Bug Originally posted at napari-aicsimageio: https://github.com/AllenCellModeling/napari-aicsimageio/issues/22 When using Light theme in Napari, the napari-aicsimageio plugin results in wierd behavior: drag-n-dropping a LIF results in a blank image, regardless of colormap, etc. However, switching to Dark theme shows that the image was properly loaded. Light theme: <img width="1546" alt="image" src="https://user-images.githubusercontent.com/76622105/132992458-01d18391-6c62-42a8-b7b7-973d40402597.png"> There is no error message. Note that the tiny preview icon in the layer list does properly show the image. Dark theme: <img width="1546" alt="image" src="https://user-images.githubusercontent.com/76622105/132992463-0ce65f4b-63b2-4283-a1aa-30a1202c3d14.png"> ## To Reproduce Drag-and-drop a LIF onto napari with napari-aicsimageio v 0.3.3 or 0.3.5 in napari 4.10 or 4.11. Change theme between Light and Dark. Here's the smallest LIF I have handy: https://www.dropbox.com/s/7201v0ae0f3iiy8/2021040_H33342_L929_24w_A3_tilescan.lif?dl=0 ## Expected behavior Image should be visible regardless of theme. Perhaps the colormap needs to be adjusted for optimal viewing, but it should not be a blank canvas. ## Environment ``` napari: 0.4.11 Platform: macOS-11.5.2-arm64-arm-64bit System: MacOS 11.5.2 Python: 3.9.6 | packaged by conda-forge | (default, Jul 6 2021, 08:51:19) [Clang 11.1.0 ] Qt: 5.15.2 PyQt5: 5.15.4 NumPy: 1.21.0 SciPy: 1.7.0 Dask: 2021.06.2 VisPy: 0.7.0 OpenGL: - GL version: 2.1 Metal - 71.7.1 - MAX_TEXTURE_SIZE: 16384 Screens: - screen 1: resolution 1680x1050, scale 2.0 Plugins: - Measurements (Plot profile): 0.1.5 - Measurements (skimage regionprops): 0.1.6 - Utilities (skimage regionprops): 0.1.6 - aicsimageio-out-of-memory: 0.3.5 - animation: 0.0.2 - clEsperanto: 0.10.0 - console: 0.0.4 - ome-types: 0.2.6 - scikit-image: 0.4.11 - svg: 0.1.5 ``` (Note: I also tested 4.10 and v. 0.3.3 of the plugin)
1.0
Additive blending mode requires black canvas (broken in light theme) - ## 🐛 Bug Originally posted at napari-aicsimageio: https://github.com/AllenCellModeling/napari-aicsimageio/issues/22 When using Light theme in Napari, the napari-aicsimageio plugin results in wierd behavior: drag-n-dropping a LIF results in a blank image, regardless of colormap, etc. However, switching to Dark theme shows that the image was properly loaded. Light theme: <img width="1546" alt="image" src="https://user-images.githubusercontent.com/76622105/132992458-01d18391-6c62-42a8-b7b7-973d40402597.png"> There is no error message. Note that the tiny preview icon in the layer list does properly show the image. Dark theme: <img width="1546" alt="image" src="https://user-images.githubusercontent.com/76622105/132992463-0ce65f4b-63b2-4283-a1aa-30a1202c3d14.png"> ## To Reproduce Drag-and-drop a LIF onto napari with napari-aicsimageio v 0.3.3 or 0.3.5 in napari 4.10 or 4.11. Change theme between Light and Dark. Here's the smallest LIF I have handy: https://www.dropbox.com/s/7201v0ae0f3iiy8/2021040_H33342_L929_24w_A3_tilescan.lif?dl=0 ## Expected behavior Image should be visible regardless of theme. Perhaps the colormap needs to be adjusted for optimal viewing, but it should not be a blank canvas. ## Environment ``` napari: 0.4.11 Platform: macOS-11.5.2-arm64-arm-64bit System: MacOS 11.5.2 Python: 3.9.6 | packaged by conda-forge | (default, Jul 6 2021, 08:51:19) [Clang 11.1.0 ] Qt: 5.15.2 PyQt5: 5.15.4 NumPy: 1.21.0 SciPy: 1.7.0 Dask: 2021.06.2 VisPy: 0.7.0 OpenGL: - GL version: 2.1 Metal - 71.7.1 - MAX_TEXTURE_SIZE: 16384 Screens: - screen 1: resolution 1680x1050, scale 2.0 Plugins: - Measurements (Plot profile): 0.1.5 - Measurements (skimage regionprops): 0.1.6 - Utilities (skimage regionprops): 0.1.6 - aicsimageio-out-of-memory: 0.3.5 - animation: 0.0.2 - clEsperanto: 0.10.0 - console: 0.0.4 - ome-types: 0.2.6 - scikit-image: 0.4.11 - svg: 0.1.5 ``` (Note: I also tested 4.10 and v. 0.3.3 of the plugin)
priority
additive blending mode requires black canvas broken in light theme 🐛 bug originally posted at napari aicsimageio when using light theme in napari the napari aicsimageio plugin results in wierd behavior drag n dropping a lif results in a blank image regardless of colormap etc however switching to dark theme shows that the image was properly loaded light theme img width alt image src there is no error message note that the tiny preview icon in the layer list does properly show the image dark theme img width alt image src to reproduce drag and drop a lif onto napari with napari aicsimageio v or in napari or change theme between light and dark here s the smallest lif i have handy expected behavior image should be visible regardless of theme perhaps the colormap needs to be adjusted for optimal viewing but it should not be a blank canvas environment napari platform macos arm system macos python packaged by conda forge default jul qt numpy scipy dask vispy opengl gl version metal max texture size screens screen resolution scale plugins measurements plot profile measurements skimage regionprops utilities skimage regionprops aicsimageio out of memory animation clesperanto console ome types scikit image svg note i also tested and v of the plugin
1
113,060
24,357,717,161
IssuesEvent
2022-10-03 08:58:15
microsoft/AL
https://api.github.com/repos/microsoft/AL
closed
Extensibility protected var error while publishing dependency apps
ships-in-future-update static-code-analysis AppSourceCop accepted
Creating protected var variables in extension objects can cause breaking changes or can cause an error during the creation of a package: 'The request for path /BC200/dev/apps?SchemaUpdateMode=forcesync&DependencyPublishingOption=default failed with code UnprocessableEntity. Reason: Failure while emitting method. Object:'TableExtension "Warehouse Journal Line A101"' Method:'OnValidate()' Index was outside the bounds of the array.' In attached example I have created two apps App_A en App_B. App_A depends on the Base app and App_B depends on App_A. The Base table object "Warehouse Journal Line" has a protected var Item. In App_A in my "Warehouse Journal Line" extension I have also created a protected var Item. This app can be published fine. In App_B there is also a "Warehouse Journal Line" extension which compiles fine and is using an Item var from either the Base or App_A app. The compiler throws above mentioned error when trying to publish App_B. The solution would obviously be to also use a suffix in App_A on protected var variables and rename Item to ItemA100. This situation is currently not acknowledged by any of the Cops. What concerns me that if somebody (like me) asks for a change in the Base app and asks for the transfer of an existing variable to the protected var section it might break a lot of dependency apps. [ProtectedVar Issue.zip](https://github.com/microsoft/ALAppExtensions/files/9525860/ProtectedVar.Issue.zip)
1.0
Extensibility protected var error while publishing dependency apps - Creating protected var variables in extension objects can cause breaking changes or can cause an error during the creation of a package: 'The request for path /BC200/dev/apps?SchemaUpdateMode=forcesync&DependencyPublishingOption=default failed with code UnprocessableEntity. Reason: Failure while emitting method. Object:'TableExtension "Warehouse Journal Line A101"' Method:'OnValidate()' Index was outside the bounds of the array.' In attached example I have created two apps App_A en App_B. App_A depends on the Base app and App_B depends on App_A. The Base table object "Warehouse Journal Line" has a protected var Item. In App_A in my "Warehouse Journal Line" extension I have also created a protected var Item. This app can be published fine. In App_B there is also a "Warehouse Journal Line" extension which compiles fine and is using an Item var from either the Base or App_A app. The compiler throws above mentioned error when trying to publish App_B. The solution would obviously be to also use a suffix in App_A on protected var variables and rename Item to ItemA100. This situation is currently not acknowledged by any of the Cops. What concerns me that if somebody (like me) asks for a change in the Base app and asks for the transfer of an existing variable to the protected var section it might break a lot of dependency apps. [ProtectedVar Issue.zip](https://github.com/microsoft/ALAppExtensions/files/9525860/ProtectedVar.Issue.zip)
non_priority
extensibility protected var error while publishing dependency apps creating protected var variables in extension objects can cause breaking changes or can cause an error during the creation of a package the request for path dev apps schemaupdatemode forcesync dependencypublishingoption default failed with code unprocessableentity reason failure while emitting method object tableextension warehouse journal line method onvalidate index was outside the bounds of the array in attached example i have created two apps app a en app b app a depends on the base app and app b depends on app a the base table object warehouse journal line has a protected var item in app a in my warehouse journal line extension i have also created a protected var item this app can be published fine in app b there is also a warehouse journal line extension which compiles fine and is using an item var from either the base or app a app the compiler throws above mentioned error when trying to publish app b the solution would obviously be to also use a suffix in app a on protected var variables and rename item to this situation is currently not acknowledged by any of the cops what concerns me that if somebody like me asks for a change in the base app and asks for the transfer of an existing variable to the protected var section it might break a lot of dependency apps
0
92,473
18,871,617,072
IssuesEvent
2021-11-13 09:18:39
beattosetto/beattosetto
https://api.github.com/repos/beattosetto/beattosetto
closed
Demo player on beatmap card improvement
frontend type:ui-ux type:code-quality type:reliability
This demo player on beatmap card is using the very old version of JQuery. It's need to be updated. If it's improved I think it's should fix the proflem like #131 too.
1.0
Demo player on beatmap card improvement - This demo player on beatmap card is using the very old version of JQuery. It's need to be updated. If it's improved I think it's should fix the proflem like #131 too.
non_priority
demo player on beatmap card improvement this demo player on beatmap card is using the very old version of jquery it s need to be updated if it s improved i think it s should fix the proflem like too
0
80,204
7,742,048,593
IssuesEvent
2018-05-29 08:20:25
eth-cscs/reframe
https://api.github.com/repos/eth-cscs/reframe
opened
Fix prog env settings for Kesch
enhancement prio: normal regression test
We need to revert the changes for `PrgEnv-pgi` on Kesch and make available also alternative the `PrgEnv-pgi` modules. To summarise, the following must be changed: 1. `PrgEnv-pgi` environment on Kesch must use the default `PrgEnv-pgi` module. 2. Provide additional `PrgEnv-pgi` environments on Kesch (AJ's version, PGI 18.4) 3. Adapt `valid_prog_environs` on Kesch using wildcards.
1.0
Fix prog env settings for Kesch - We need to revert the changes for `PrgEnv-pgi` on Kesch and make available also alternative the `PrgEnv-pgi` modules. To summarise, the following must be changed: 1. `PrgEnv-pgi` environment on Kesch must use the default `PrgEnv-pgi` module. 2. Provide additional `PrgEnv-pgi` environments on Kesch (AJ's version, PGI 18.4) 3. Adapt `valid_prog_environs` on Kesch using wildcards.
non_priority
fix prog env settings for kesch we need to revert the changes for prgenv pgi on kesch and make available also alternative the prgenv pgi modules to summarise the following must be changed prgenv pgi environment on kesch must use the default prgenv pgi module provide additional prgenv pgi environments on kesch aj s version pgi adapt valid prog environs on kesch using wildcards
0
67,760
3,281,885,507
IssuesEvent
2015-10-28 01:08:40
sceene/sprint-planning
https://api.github.com/repos/sceene/sprint-planning
closed
2. ACTIVITY SCRUB
HIGH PRIORITY
## **IMPORTANT!!** **CHECK #UPDATES CHANNEL IN SLACK FOR ANYTHING THAT RELATES TO ACTIVITY** **EVERYTHING IS PRIORITIZED AND IN ORDER.** - [x] Clicking another users avatar or username crashes app. [**KILLER BUG**] - [x] Nothing happens when clicking an alert from the iOS pull down notifications panel [**BUG**] - [x] Quotes are shown as photos in notifications [**BUG**] - [x] Display **ALL** the content in **ALL** the content containers for **ALL** content types - Don't fake these as the content is now available. - Make sure you're resizing the content images to 40 height, 40width @1x - Make sure the content containers are 36X36 @1x - Endpoints are here: https://github.com/sceene/scn-api-v1/issues/64 [**STEVE COMPLETED**] - Provide the chevron image to Steve so it can be added to the payload. - **This is the third week of this being incomplete. Make sure this is addressed.** - [x] When opening a notification from the lock screen – or drop down badge – display the content landing page that enables a user to click the back button – or swipe back – to the activity tab the notification is from. - **For example**: I get a like notification for my photo on lock screen. - I tap and am redirected to the photo landing page - I see a back button in the navigation icon and tap it - I am redirected back to the notification panel where I see the like as the top notification - [x] Remove the modal view for content landing page. - [x] Make Notifications the default tab (On the left) - [x] Activate the tab based on the type of notification when app is in foreground. - If users receives a like notification show the notifications panel. - If the user receives an alert activate the alert tab. - If the user receives both then show the tab notification badge - [ ] Content containers and images hiding – and reloading – some times when you scroll up and down the activity tab - [x] Resize the avatars and content in content containers and do NOT use the same size for all of the different screen sizes. - This can be solved by setting the base avatar to 40px height and 40px width then setting **dpr** value of 2, 3 etc based on the screen ratio you're targeting. - https://www.imgix.com/docs/reference/pixeldensity#param-dpr - @1x height: 40, width: 40 - @2x height: 80, width: 80 - @3x height: 120, width: 120 - [x] Move the tab badge over 5px to the right - [x] Tie in the modes to the settings. - Modes go under the Notifications & Alerts section in the settings - User will select modes then see a screen with three options that can be selected. - The selection will be a tick. - See comments below.
1.0
2. ACTIVITY SCRUB - ## **IMPORTANT!!** **CHECK #UPDATES CHANNEL IN SLACK FOR ANYTHING THAT RELATES TO ACTIVITY** **EVERYTHING IS PRIORITIZED AND IN ORDER.** - [x] Clicking another users avatar or username crashes app. [**KILLER BUG**] - [x] Nothing happens when clicking an alert from the iOS pull down notifications panel [**BUG**] - [x] Quotes are shown as photos in notifications [**BUG**] - [x] Display **ALL** the content in **ALL** the content containers for **ALL** content types - Don't fake these as the content is now available. - Make sure you're resizing the content images to 40 height, 40width @1x - Make sure the content containers are 36X36 @1x - Endpoints are here: https://github.com/sceene/scn-api-v1/issues/64 [**STEVE COMPLETED**] - Provide the chevron image to Steve so it can be added to the payload. - **This is the third week of this being incomplete. Make sure this is addressed.** - [x] When opening a notification from the lock screen – or drop down badge – display the content landing page that enables a user to click the back button – or swipe back – to the activity tab the notification is from. - **For example**: I get a like notification for my photo on lock screen. - I tap and am redirected to the photo landing page - I see a back button in the navigation icon and tap it - I am redirected back to the notification panel where I see the like as the top notification - [x] Remove the modal view for content landing page. - [x] Make Notifications the default tab (On the left) - [x] Activate the tab based on the type of notification when app is in foreground. - If users receives a like notification show the notifications panel. - If the user receives an alert activate the alert tab. - If the user receives both then show the tab notification badge - [ ] Content containers and images hiding – and reloading – some times when you scroll up and down the activity tab - [x] Resize the avatars and content in content containers and do NOT use the same size for all of the different screen sizes. - This can be solved by setting the base avatar to 40px height and 40px width then setting **dpr** value of 2, 3 etc based on the screen ratio you're targeting. - https://www.imgix.com/docs/reference/pixeldensity#param-dpr - @1x height: 40, width: 40 - @2x height: 80, width: 80 - @3x height: 120, width: 120 - [x] Move the tab badge over 5px to the right - [x] Tie in the modes to the settings. - Modes go under the Notifications & Alerts section in the settings - User will select modes then see a screen with three options that can be selected. - The selection will be a tick. - See comments below.
priority
activity scrub important check updates channel in slack for anything that relates to activity everything is prioritized and in order clicking another users avatar or username crashes app nothing happens when clicking an alert from the ios pull down notifications panel quotes are shown as photos in notifications display all the content in all the content containers for all content types don t fake these as the content is now available make sure you re resizing the content images to height make sure the content containers are endpoints are here provide the chevron image to steve so it can be added to the payload this is the third week of this being incomplete make sure this is addressed when opening a notification from the lock screen – or drop down badge – display the content landing page that enables a user to click the back button – or swipe back – to the activity tab the notification is from for example i get a like notification for my photo on lock screen i tap and am redirected to the photo landing page i see a back button in the navigation icon and tap it i am redirected back to the notification panel where i see the like as the top notification remove the modal view for content landing page make notifications the default tab on the left activate the tab based on the type of notification when app is in foreground if users receives a like notification show the notifications panel if the user receives an alert activate the alert tab if the user receives both then show the tab notification badge content containers and images hiding – and reloading – some times when you scroll up and down the activity tab resize the avatars and content in content containers and do not use the same size for all of the different screen sizes this can be solved by setting the base avatar to height and width then setting dpr value of etc based on the screen ratio you re targeting height width height width height width move the tab badge over to the right tie in the modes to the settings modes go under the notifications alerts section in the settings user will select modes then see a screen with three options that can be selected the selection will be a tick see comments below
1
194,169
6,892,063,615
IssuesEvent
2017-11-22 19:25:33
Peergos/Peergos
https://api.github.com/repos/Peergos/Peergos
closed
If i try to submit a follow request with a person i already follow, no error message is displayed
bug priority:Low
Currently this scenario does not display an error message.
1.0
If i try to submit a follow request with a person i already follow, no error message is displayed - Currently this scenario does not display an error message.
priority
if i try to submit a follow request with a person i already follow no error message is displayed currently this scenario does not display an error message
1
505,736
14,644,410,866
IssuesEvent
2020-12-25 23:21:24
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
opened
Copy paste images to text editor
feature: enhancement priority: medium
**Is your feature request related to a problem? Please describe.** Some users assume that copy-pasting works on the text editor, but after sending the message or posting it, the image is gone. **Describe the solution you'd like** Should be able to copy-paste images without using the media uploader. **Support ticket links** https://secure.helpscout.net/conversation/1373801519/115950/
1.0
Copy paste images to text editor - **Is your feature request related to a problem? Please describe.** Some users assume that copy-pasting works on the text editor, but after sending the message or posting it, the image is gone. **Describe the solution you'd like** Should be able to copy-paste images without using the media uploader. **Support ticket links** https://secure.helpscout.net/conversation/1373801519/115950/
priority
copy paste images to text editor is your feature request related to a problem please describe some users assume that copy pasting works on the text editor but after sending the message or posting it the image is gone describe the solution you d like should be able to copy paste images without using the media uploader support ticket links
1
365,432
10,781,293,760
IssuesEvent
2019-11-04 14:39:32
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
outlook.office.com - site is not usable
browser-fenix engine-gecko priority-critical
<!-- @browser: Firefox Mobile 70.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:70.0) Gecko/70.0 Firefox/70.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://outlook.office.com/mail/ **Browser / Version**: Firefox Mobile 70.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: reports error after loggin in **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
outlook.office.com - site is not usable - <!-- @browser: Firefox Mobile 70.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:70.0) Gecko/70.0 Firefox/70.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://outlook.office.com/mail/ **Browser / Version**: Firefox Mobile 70.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: reports error after loggin in **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
outlook office com site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description reports error after loggin in steps to reproduce browser configuration none from with ❤️
1
724,506
24,933,082,843
IssuesEvent
2022-10-31 13:15:05
AY2223S1-CS2103T-T09-2/tp
https://api.github.com/repos/AY2223S1-CS2103T-T09-2/tp
closed
[PE-D][Tester E] UI does not display the correct term.
priority.Medium
Describe the bug The UI shows *person listed* instead of *orders listed* ## To Reproduce Steps to reproduce the behaviour: 1. Input *filter-o os/Pending* ## Expected behaviour Success message will be (NUMBER OF) orders listed. ## Actual behaviour Success message is (NUMBER OF ) persons listed. ## Screenshots ![image.png](https://raw.githubusercontent.com/yeoyujie/ped/main/files/cd2ad1c8-9ba8-446c-857b-5dfe6e503e83.png) <!--session: 1666944362706-c0fb4544-4a59-4b80-94d1-13eff4e55507--> <!--Version: Web v3.4.4--> ------------- Labels: `severity.VeryLow` `type.FunctionalityBug` original: yeoyujie/ped#4
1.0
[PE-D][Tester E] UI does not display the correct term. - Describe the bug The UI shows *person listed* instead of *orders listed* ## To Reproduce Steps to reproduce the behaviour: 1. Input *filter-o os/Pending* ## Expected behaviour Success message will be (NUMBER OF) orders listed. ## Actual behaviour Success message is (NUMBER OF ) persons listed. ## Screenshots ![image.png](https://raw.githubusercontent.com/yeoyujie/ped/main/files/cd2ad1c8-9ba8-446c-857b-5dfe6e503e83.png) <!--session: 1666944362706-c0fb4544-4a59-4b80-94d1-13eff4e55507--> <!--Version: Web v3.4.4--> ------------- Labels: `severity.VeryLow` `type.FunctionalityBug` original: yeoyujie/ped#4
priority
ui does not display the correct term describe the bug the ui shows person listed instead of orders listed to reproduce steps to reproduce the behaviour input filter o os pending expected behaviour success message will be number of orders listed actual behaviour success message is number of persons listed screenshots labels severity verylow type functionalitybug original yeoyujie ped
1
669,971
22,647,928,633
IssuesEvent
2022-07-01 10:32:10
ntop/ntopng
https://api.github.com/repos/ntop/ntopng
closed
Implement alert search in Description
feature request priority ticket alerts
![Screen Shot 2021-10-07 at 12 34 32](https://user-images.githubusercontent.com/4493366/136370647-735159c0-779a-40bf-86c3-cf6c4dbc710a.png) in the above example it should be possible to filter "hello.ntop.org" os a substring (e.g. hello) in the description field
1.0
Implement alert search in Description - ![Screen Shot 2021-10-07 at 12 34 32](https://user-images.githubusercontent.com/4493366/136370647-735159c0-779a-40bf-86c3-cf6c4dbc710a.png) in the above example it should be possible to filter "hello.ntop.org" os a substring (e.g. hello) in the description field
priority
implement alert search in description in the above example it should be possible to filter hello ntop org os a substring e g hello in the description field
1
363,689
10,746,157,338
IssuesEvent
2019-10-30 10:29:21
AY1920S1-CS2103T-T12-2/main
https://api.github.com/repos/AY1920S1-CS2103T-T12-2/main
reopened
As an administrator I can view all relevant commands with a single character input...
priority.Medium type.Story
... so that I can efficiently manage the system without having the need to memorize or refer to the list of commands when typing.
1.0
As an administrator I can view all relevant commands with a single character input... - ... so that I can efficiently manage the system without having the need to memorize or refer to the list of commands when typing.
priority
as an administrator i can view all relevant commands with a single character input so that i can efficiently manage the system without having the need to memorize or refer to the list of commands when typing
1
69,545
13,263,039,617
IssuesEvent
2020-08-20 23:19:20
eoscostarica/lifebank
https://api.github.com/repos/eoscostarica/lifebank
closed
Fix 'Using latest is prone to errors if the image will ever update. Pin the version explicitly to a release tag' issue in webapp\Dockerfile
code-quality
[CodeFactor](https://www.codefactor.io/repository/github/eoscostarica/lifebank/overview/master) found an issue: Using latest is prone to errors if the image will ever update. Pin the version explicitly to a release tag It's currently on: [webapp\Dockerfile:36 ](https://www.codefactor.io/repository/github/eoscostarica/lifebank/source/master/webapp/Dockerfile#L36)
1.0
Fix 'Using latest is prone to errors if the image will ever update. Pin the version explicitly to a release tag' issue in webapp\Dockerfile - [CodeFactor](https://www.codefactor.io/repository/github/eoscostarica/lifebank/overview/master) found an issue: Using latest is prone to errors if the image will ever update. Pin the version explicitly to a release tag It's currently on: [webapp\Dockerfile:36 ](https://www.codefactor.io/repository/github/eoscostarica/lifebank/source/master/webapp/Dockerfile#L36)
non_priority
fix using latest is prone to errors if the image will ever update pin the version explicitly to a release tag issue in webapp dockerfile found an issue using latest is prone to errors if the image will ever update pin the version explicitly to a release tag it s currently on webapp dockerfile
0
1,911
2,867,193,019
IssuesEvent
2015-06-05 11:43:18
angular/angular
https://api.github.com/repos/angular/angular
opened
gulp serve.js.prod is not more building when a src file is updated
comp: build/pipeline effort: easy P1: urgent type: bug
seen with this file: angular/modules/benchmarks/src/naive_infinite_scroll/app.ts linked to the migration to ts ?
1.0
gulp serve.js.prod is not more building when a src file is updated - seen with this file: angular/modules/benchmarks/src/naive_infinite_scroll/app.ts linked to the migration to ts ?
non_priority
gulp serve js prod is not more building when a src file is updated seen with this file angular modules benchmarks src naive infinite scroll app ts linked to the migration to ts
0
91,386
11,499,308,400
IssuesEvent
2020-02-12 13:43:47
Stichting-CROW/imbor
https://api.github.com/repos/Stichting-CROW/imbor
closed
GWSW informatie wordt uit de transformatie gelaten
design decision
GWSW records zetten we niet over naar de ontologie (deze filteren we bij de transformatie er uit op basis van informatie uit de Access database. Dit is de relatie tussen de tabel [X_Informatiemodellen] en [X_ObjecttypesAttributen]. Deze gaat Jochem nog toevoegen (20191029) Bijproduct is een Excel sheet met in kolom A de URI's van de IMBOR objecttypen en de naam in GWSW _Er is besloten door de provincies dat alles m.b.t. riolering bij de GWSWS gelaten wordt. We willen hier niets zeggen over deze objeccttypen en eigenschappen. Vandaar dat we het geheel uit de ontologie laten om geen overlap te hebben._ *uitzoeken
1.0
GWSW informatie wordt uit de transformatie gelaten - GWSW records zetten we niet over naar de ontologie (deze filteren we bij de transformatie er uit op basis van informatie uit de Access database. Dit is de relatie tussen de tabel [X_Informatiemodellen] en [X_ObjecttypesAttributen]. Deze gaat Jochem nog toevoegen (20191029) Bijproduct is een Excel sheet met in kolom A de URI's van de IMBOR objecttypen en de naam in GWSW _Er is besloten door de provincies dat alles m.b.t. riolering bij de GWSWS gelaten wordt. We willen hier niets zeggen over deze objeccttypen en eigenschappen. Vandaar dat we het geheel uit de ontologie laten om geen overlap te hebben._ *uitzoeken
non_priority
gwsw informatie wordt uit de transformatie gelaten gwsw records zetten we niet over naar de ontologie deze filteren we bij de transformatie er uit op basis van informatie uit de access database dit is de relatie tussen de tabel en deze gaat jochem nog toevoegen bijproduct is een excel sheet met in kolom a de uri s van de imbor objecttypen en de naam in gwsw er is besloten door de provincies dat alles m b t riolering bij de gwsws gelaten wordt we willen hier niets zeggen over deze objeccttypen en eigenschappen vandaar dat we het geheel uit de ontologie laten om geen overlap te hebben uitzoeken
0
180,893
13,964,354,150
IssuesEvent
2020-10-25 17:49:15
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
sql: add randomized correctness testing for TPCH queries
A-testing C-enhancement
TPCH queries are interesting and produce tricky plans. We currently only test correctness on these queries by using the "standard" example placeholders provided by the test spec. We've found bugs using these queries already - but it seems like there's no guarantee that producing correct results for the standard example placeholders means that we'd produce correct results for all combinations of placeholders. This issue tracks adding infrastructure to generate random variations on the TPCH queries and verify their results against the vectorized engine, the non-vectorized engine, and PostgreSQL.
1.0
sql: add randomized correctness testing for TPCH queries - TPCH queries are interesting and produce tricky plans. We currently only test correctness on these queries by using the "standard" example placeholders provided by the test spec. We've found bugs using these queries already - but it seems like there's no guarantee that producing correct results for the standard example placeholders means that we'd produce correct results for all combinations of placeholders. This issue tracks adding infrastructure to generate random variations on the TPCH queries and verify their results against the vectorized engine, the non-vectorized engine, and PostgreSQL.
non_priority
sql add randomized correctness testing for tpch queries tpch queries are interesting and produce tricky plans we currently only test correctness on these queries by using the standard example placeholders provided by the test spec we ve found bugs using these queries already but it seems like there s no guarantee that producing correct results for the standard example placeholders means that we d produce correct results for all combinations of placeholders this issue tracks adding infrastructure to generate random variations on the tpch queries and verify their results against the vectorized engine the non vectorized engine and postgresql
0
32,943
7,624,313,475
IssuesEvent
2018-05-03 17:36:44
MicrosoftDocs/windows-driver-docs
https://api.github.com/repos/MicrosoftDocs/windows-driver-docs
closed
Device Configuration failed
help with customer code
Installing necessary components... Copying required files Task "Copying required files" completed successfully Configuring WDK Remote User Account Task "Configuring WDK Remote User Account" completed successfully Installing .NET Framework (possible reboot) Task "Installing .NET Framework (possible reboot)" completed successfully Installing VC Redist (x64) Task "Installing VC Redist (x64)" completed successfully Installing test automation (x86) Task "Installing test automation (x86)" completed successfully Installing test automation (x64) Task "Installing test automation (x64)" completed successfully Installing debuggers (x86) Task "Installing debuggers (x86)" completed successfully Installing debuggers (x64) Task "Installing debuggers (x64)" completed successfully Installing driver test framework Task "Installing driver test framework" completed successfully Registering logging components Task "Registering logging components" completed successfully Configuring TAEF test service to start automatically Task "Configuring TAEF test service to start automatically" completed successfully Configuring kernel debugger settings (possible reboot) Task "Configuring kernel debugger settings (possible reboot)" completed successfully Configuring computer settings (possible reboot) Task "Configuring computer settings (possible reboot)" completed successfully Creating system restore point ERROR: Task "Creating system restore point" failed to complete successfully. Look at the logs in the driver test group explorer for more details on the failure. Computer configuration log file://C:/Users/Silence/AppData/Roaming/Microsoft/WDKTestInfrastructure/ProvisioningLogs/Driver%20Test%20Computer%20Configuration%2020180404211822384.log Failed installing components --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 75819dbc-e20a-29f0-af02-5afc4628fd95 * Version Independent ID: fca932fe-08ba-a01c-30a7-2288342f6014 * Content: [Setting Up Kernel-Mode Debugging over a Network Cable in Visual Studio](https://docs.microsoft.com/en-us/windows-hardware/drivers/debugger/setting-up-a-network-debugging-connection-in-visual-studio) * Content Source: [windows-driver-docs-pr/debugger/setting-up-a-network-debugging-connection-in-visual-studio.md](https://github.com/MicrosoftDocs/windows-driver-docs/blob/staging/windows-driver-docs-pr/debugger/setting-up-a-network-debugging-connection-in-visual-studio.md) * Service: **unspecified** * GitHub Login: @DOMARS * Microsoft Alias: **domars**
1.0
Device Configuration failed - Installing necessary components... Copying required files Task "Copying required files" completed successfully Configuring WDK Remote User Account Task "Configuring WDK Remote User Account" completed successfully Installing .NET Framework (possible reboot) Task "Installing .NET Framework (possible reboot)" completed successfully Installing VC Redist (x64) Task "Installing VC Redist (x64)" completed successfully Installing test automation (x86) Task "Installing test automation (x86)" completed successfully Installing test automation (x64) Task "Installing test automation (x64)" completed successfully Installing debuggers (x86) Task "Installing debuggers (x86)" completed successfully Installing debuggers (x64) Task "Installing debuggers (x64)" completed successfully Installing driver test framework Task "Installing driver test framework" completed successfully Registering logging components Task "Registering logging components" completed successfully Configuring TAEF test service to start automatically Task "Configuring TAEF test service to start automatically" completed successfully Configuring kernel debugger settings (possible reboot) Task "Configuring kernel debugger settings (possible reboot)" completed successfully Configuring computer settings (possible reboot) Task "Configuring computer settings (possible reboot)" completed successfully Creating system restore point ERROR: Task "Creating system restore point" failed to complete successfully. Look at the logs in the driver test group explorer for more details on the failure. Computer configuration log file://C:/Users/Silence/AppData/Roaming/Microsoft/WDKTestInfrastructure/ProvisioningLogs/Driver%20Test%20Computer%20Configuration%2020180404211822384.log Failed installing components --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 75819dbc-e20a-29f0-af02-5afc4628fd95 * Version Independent ID: fca932fe-08ba-a01c-30a7-2288342f6014 * Content: [Setting Up Kernel-Mode Debugging over a Network Cable in Visual Studio](https://docs.microsoft.com/en-us/windows-hardware/drivers/debugger/setting-up-a-network-debugging-connection-in-visual-studio) * Content Source: [windows-driver-docs-pr/debugger/setting-up-a-network-debugging-connection-in-visual-studio.md](https://github.com/MicrosoftDocs/windows-driver-docs/blob/staging/windows-driver-docs-pr/debugger/setting-up-a-network-debugging-connection-in-visual-studio.md) * Service: **unspecified** * GitHub Login: @DOMARS * Microsoft Alias: **domars**
non_priority
device configuration failed installing necessary components copying required files task copying required files completed successfully configuring wdk remote user account task configuring wdk remote user account completed successfully installing net framework possible reboot task installing net framework possible reboot completed successfully installing vc redist task installing vc redist completed successfully installing test automation task installing test automation completed successfully installing test automation task installing test automation completed successfully installing debuggers task installing debuggers completed successfully installing debuggers task installing debuggers completed successfully installing driver test framework task installing driver test framework completed successfully registering logging components task registering logging components completed successfully configuring taef test service to start automatically task configuring taef test service to start automatically completed successfully configuring kernel debugger settings possible reboot task configuring kernel debugger settings possible reboot completed successfully configuring computer settings possible reboot task configuring computer settings possible reboot completed successfully creating system restore point error task creating system restore point failed to complete successfully look at the logs in the driver test group explorer for more details on the failure computer configuration log file c users silence appdata roaming microsoft wdktestinfrastructure provisioninglogs driver log failed installing components document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service unspecified github login domars microsoft alias domars
0
554,219
16,414,740,249
IssuesEvent
2021-05-19 04:20:12
HBRP/Homebound
https://api.github.com/repos/HBRP/Homebound
opened
Character - Time Progression System
feature medium priority
To facilitate players creating and playing with multiple characters, tie a progression system to the amount of character's they currently own and how much time they've individually spent on each character.
1.0
Character - Time Progression System - To facilitate players creating and playing with multiple characters, tie a progression system to the amount of character's they currently own and how much time they've individually spent on each character.
priority
character time progression system to facilitate players creating and playing with multiple characters tie a progression system to the amount of character s they currently own and how much time they ve individually spent on each character
1
266,554
23,245,540,240
IssuesEvent
2022-08-03 19:45:25
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Notebook API tests editor editing event
integration-test-failure notebook
``` 1) Notebook API tests editor editing event: Error: asPromise TIMEOUT reached at Timeout._onTimeout (extensions/vscode-api-tests/src/utils.ts:132:11) at listOnTimeout (node:internal/timers:557:17) at processTimers (node:internal/timers:500:7) ``` https://dev.azure.com/monacotools/Monaco/_build/results?buildId=174304&view=logs&j=9833cdaa-2e92-5f8d-56a8-239fd18fd628&t=9bc48eac-eb66-52c8-7250-d6490c5f3a15&l=677
1.0
Notebook API tests editor editing event - ``` 1) Notebook API tests editor editing event: Error: asPromise TIMEOUT reached at Timeout._onTimeout (extensions/vscode-api-tests/src/utils.ts:132:11) at listOnTimeout (node:internal/timers:557:17) at processTimers (node:internal/timers:500:7) ``` https://dev.azure.com/monacotools/Monaco/_build/results?buildId=174304&view=logs&j=9833cdaa-2e92-5f8d-56a8-239fd18fd628&t=9bc48eac-eb66-52c8-7250-d6490c5f3a15&l=677
non_priority
notebook api tests editor editing event notebook api tests editor editing event error aspromise timeout reached at timeout ontimeout extensions vscode api tests src utils ts at listontimeout node internal timers at processtimers node internal timers
0
168,406
6,370,822,144
IssuesEvent
2017-08-01 14:54:16
tardis-sn/tardis
https://api.github.com/repos/tardis-sn/tardis
closed
simple sets of stratified abundances
configuration feature request priority - low
It would be good to have a version of the abundance section that can deal with simple sets of multi zone abundances - say a "type:stratified" or something. Basically one would set up a bunch of abundance zones, each of which is specified by a set of numbers like: ``` start: 10000 km/s stop: 11000 km/s O: 0.19 Mg: 0.03 Si: 0.52 S: 0.19 Ar: 0.04 Ca: 0.03 ``` …so that the abundances are populated with those vales in that velocity range. The user can then set up a model with a handful of zones (perhaps up to a max of 10 or so) so that they can play with non-uniform abundances along one of the simple density profiles (i.e. a level of sophistication between our current "uniform" setting and the gory details of giving a fully tabulated set via an input file). Can this be incorporated as part of the next version of the config file?
1.0
simple sets of stratified abundances - It would be good to have a version of the abundance section that can deal with simple sets of multi zone abundances - say a "type:stratified" or something. Basically one would set up a bunch of abundance zones, each of which is specified by a set of numbers like: ``` start: 10000 km/s stop: 11000 km/s O: 0.19 Mg: 0.03 Si: 0.52 S: 0.19 Ar: 0.04 Ca: 0.03 ``` …so that the abundances are populated with those vales in that velocity range. The user can then set up a model with a handful of zones (perhaps up to a max of 10 or so) so that they can play with non-uniform abundances along one of the simple density profiles (i.e. a level of sophistication between our current "uniform" setting and the gory details of giving a fully tabulated set via an input file). Can this be incorporated as part of the next version of the config file?
priority
simple sets of stratified abundances it would be good to have a version of the abundance section that can deal with simple sets of multi zone abundances say a type stratified or something basically one would set up a bunch of abundance zones each of which is specified by a set of numbers like start km s stop km s o mg si s ar ca …so that the abundances are populated with those vales in that velocity range the user can then set up a model with a handful of zones perhaps up to a max of or so so that they can play with non uniform abundances along one of the simple density profiles i e a level of sophistication between our current uniform setting and the gory details of giving a fully tabulated set via an input file can this be incorporated as part of the next version of the config file
1
69,306
30,224,048,973
IssuesEvent
2023-07-05 22:01:29
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
Unexpected error message - The configuration of 'ExceptionMarshalling' and 'ExceptionMarshallingCustomType' on method 'Demo.MySafeHandle' is invalid.
area-System.Runtime.InteropServices source-generator
An unexpected error message is produced with the following. I'm not using or expecting any `ExceptionMarshalling` cases. ``` ...\runtime\src\samples\LibraryImportGeneratorSample\Program.cs(41,111): error SYSLIB1051: The configuration of 'ExceptionMarshalling' and 'ExceptionMarshallingCustomType' on method 'Demo.MySafeHandle' is invalid. h [...\runtime\src\samples\LibraryImportGeneratorSample\LibraryImportGeneratorSample.csproj] ...\runtime\src\samples\LibraryImportGeneratorSample\Microsoft.Interop.ComInterfaceGenerator\Microsoft.Interop.ComInterfaceGenerator\Demo.IType.cs(32,113): error CS8894: Cannot use 'MySafeHandle' as a parameter type on a method attributed with 'UnmanagedCallersOnly'. [...\runtime\src\samples\LibraryImportGeneratorSample\LibraryImportGeneratorSample.csproj] ``` ```csharp namspace Demo; public sealed class MySafeHandle : Microsoft.Win32.SafeHandles.SafeHandleZeroOrMinusOneIsInvalid { public MySafeHandle() : base(true) { } public MySafeHandle(IntPtr handle, bool ownsHandle) : base(ownsHandle) { } override protected bool ReleaseHandle() {} } [GeneratedComInterface] [Guid("BA0050D3-4BE1-4DFF-8130-705412C858E6")] public partial interface IEnum { void Foo([MarshalUsing(typeof(SafeHandleMarshaller<MySafeHandle>.ManagedToUnmanagedIn))] MySafeHandle h); } ```
1.0
Unexpected error message - The configuration of 'ExceptionMarshalling' and 'ExceptionMarshallingCustomType' on method 'Demo.MySafeHandle' is invalid. - An unexpected error message is produced with the following. I'm not using or expecting any `ExceptionMarshalling` cases. ``` ...\runtime\src\samples\LibraryImportGeneratorSample\Program.cs(41,111): error SYSLIB1051: The configuration of 'ExceptionMarshalling' and 'ExceptionMarshallingCustomType' on method 'Demo.MySafeHandle' is invalid. h [...\runtime\src\samples\LibraryImportGeneratorSample\LibraryImportGeneratorSample.csproj] ...\runtime\src\samples\LibraryImportGeneratorSample\Microsoft.Interop.ComInterfaceGenerator\Microsoft.Interop.ComInterfaceGenerator\Demo.IType.cs(32,113): error CS8894: Cannot use 'MySafeHandle' as a parameter type on a method attributed with 'UnmanagedCallersOnly'. [...\runtime\src\samples\LibraryImportGeneratorSample\LibraryImportGeneratorSample.csproj] ``` ```csharp namspace Demo; public sealed class MySafeHandle : Microsoft.Win32.SafeHandles.SafeHandleZeroOrMinusOneIsInvalid { public MySafeHandle() : base(true) { } public MySafeHandle(IntPtr handle, bool ownsHandle) : base(ownsHandle) { } override protected bool ReleaseHandle() {} } [GeneratedComInterface] [Guid("BA0050D3-4BE1-4DFF-8130-705412C858E6")] public partial interface IEnum { void Foo([MarshalUsing(typeof(SafeHandleMarshaller<MySafeHandle>.ManagedToUnmanagedIn))] MySafeHandle h); } ```
non_priority
unexpected error message the configuration of exceptionmarshalling and exceptionmarshallingcustomtype on method demo mysafehandle is invalid an unexpected error message is produced with the following i m not using or expecting any exceptionmarshalling cases runtime src samples libraryimportgeneratorsample program cs error the configuration of exceptionmarshalling and exceptionmarshallingcustomtype on method demo mysafehandle is invalid h runtime src samples libraryimportgeneratorsample microsoft interop cominterfacegenerator microsoft interop cominterfacegenerator demo itype cs error cannot use mysafehandle as a parameter type on a method attributed with unmanagedcallersonly csharp namspace demo public sealed class mysafehandle microsoft safehandles safehandlezeroorminusoneisinvalid public mysafehandle base true public mysafehandle intptr handle bool ownshandle base ownshandle override protected bool releasehandle public partial interface ienum void foo mysafehandle h
0
684,070
23,405,942,150
IssuesEvent
2022-08-12 12:51:10
googleapis/python-compute
https://api.github.com/repos/googleapis/python-compute
closed
tests.system.test_smoke.TestComputeSmoke: test_zero_values failed
type: bug priority: p1 api: compute flakybot: issue flakybot: flaky
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: ceb24f3e474c385f8c755881f42e825a13a1150d buildURL: [Build Status](https://source.cloud.google.com/results/invocations/736e3f2d-1feb-4efd-b925-13ecd667540d), [Sponge](http://sponge2/736e3f2d-1feb-4efd-b925-13ecd667540d) status: failed <details><summary>Test output</summary><br><pre>self = <test_smoke.TestComputeSmoke testMethod=test_zero_values> def test_zero_values(self): with self.assertRaises(expected_exception=TypeError) as ex: self.client.get(instance=self.name, zone=0) > self.assertIn( "0 has type int, but expected one of: bytes, unicode", str(ex.exception.args), ) E AssertionError: '0 has type int, but expected one of: bytes, unicode' not found in "('bad argument type for built-in operation',)" tests/system/test_smoke.py:77: AssertionError</pre></details>
1.0
tests.system.test_smoke.TestComputeSmoke: test_zero_values failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: ceb24f3e474c385f8c755881f42e825a13a1150d buildURL: [Build Status](https://source.cloud.google.com/results/invocations/736e3f2d-1feb-4efd-b925-13ecd667540d), [Sponge](http://sponge2/736e3f2d-1feb-4efd-b925-13ecd667540d) status: failed <details><summary>Test output</summary><br><pre>self = <test_smoke.TestComputeSmoke testMethod=test_zero_values> def test_zero_values(self): with self.assertRaises(expected_exception=TypeError) as ex: self.client.get(instance=self.name, zone=0) > self.assertIn( "0 has type int, but expected one of: bytes, unicode", str(ex.exception.args), ) E AssertionError: '0 has type int, but expected one of: bytes, unicode' not found in "('bad argument type for built-in operation',)" tests/system/test_smoke.py:77: AssertionError</pre></details>
priority
tests system test smoke testcomputesmoke test zero values failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output self def test zero values self with self assertraises expected exception typeerror as ex self client get instance self name zone self assertin has type int but expected one of bytes unicode str ex exception args e assertionerror has type int but expected one of bytes unicode not found in bad argument type for built in operation tests system test smoke py assertionerror
1
446,010
12,838,641,230
IssuesEvent
2020-07-07 17:50:23
vanilla/vanilla
https://api.github.com/repos/vanilla/vanilla
closed
Keystone - Dusk Theme Lacks Visible 'Ok/Cancel' Buttons When Adding Note To A Profile
Domain: Legacy Frontend Priority: Low Type: Bug
**Describe the bug** While using the Keystone Dusk theme option, if there is an attempt to add a note to a user's profile, the note editor will appear as expected but the "Ok" and "Cancel" buttons are invisible unless the cursor is hovering over them. **To Reproduce** 1. Switch a community to use the Keystone Dusk theme option 2. Navigate to a user profile and attempt to add a note 3. Observe the seeming lack of "Ok" and "Cancel" buttons under the editor 4. Observe that by hovering the cursor over a certain area, the buttons do appear and function as expected **Expected behavior** - "Ok" and "Cancel" buttons should always be visible **Screenshots** When opening note editor: ![image](https://user-images.githubusercontent.com/19805402/66486625-fe9a2200-ea78-11e9-9e90-b26c62fb5851.png) Hovering over "Ok" button: ![image](https://user-images.githubusercontent.com/19805402/66486680-15407900-ea79-11e9-9dc9-56bb7d99c6d3.png) **Other Info** - Issue does not occur with any other Keystone theme option
1.0
Keystone - Dusk Theme Lacks Visible 'Ok/Cancel' Buttons When Adding Note To A Profile - **Describe the bug** While using the Keystone Dusk theme option, if there is an attempt to add a note to a user's profile, the note editor will appear as expected but the "Ok" and "Cancel" buttons are invisible unless the cursor is hovering over them. **To Reproduce** 1. Switch a community to use the Keystone Dusk theme option 2. Navigate to a user profile and attempt to add a note 3. Observe the seeming lack of "Ok" and "Cancel" buttons under the editor 4. Observe that by hovering the cursor over a certain area, the buttons do appear and function as expected **Expected behavior** - "Ok" and "Cancel" buttons should always be visible **Screenshots** When opening note editor: ![image](https://user-images.githubusercontent.com/19805402/66486625-fe9a2200-ea78-11e9-9e90-b26c62fb5851.png) Hovering over "Ok" button: ![image](https://user-images.githubusercontent.com/19805402/66486680-15407900-ea79-11e9-9dc9-56bb7d99c6d3.png) **Other Info** - Issue does not occur with any other Keystone theme option
priority
keystone dusk theme lacks visible ok cancel buttons when adding note to a profile describe the bug while using the keystone dusk theme option if there is an attempt to add a note to a user s profile the note editor will appear as expected but the ok and cancel buttons are invisible unless the cursor is hovering over them to reproduce switch a community to use the keystone dusk theme option navigate to a user profile and attempt to add a note observe the seeming lack of ok and cancel buttons under the editor observe that by hovering the cursor over a certain area the buttons do appear and function as expected expected behavior ok and cancel buttons should always be visible screenshots when opening note editor hovering over ok button other info issue does not occur with any other keystone theme option
1
47,747
13,066,187,170
IssuesEvent
2020-07-30 21:10:17
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
testing - separate tests in to short/long groups (Trac #1113)
Migrated from Trac defect infrastructure
I'll cmake this up if lists for each group are provided Migrated from https://code.icecube.wisc.edu/ticket/1113 ```json { "status": "closed", "changetime": "2020-06-24T12:31:42", "description": "I'll cmake this up if lists for each group are provided", "reporter": "nega", "cc": "", "resolution": "worksforme", "_ts": "1593001902142004", "component": "infrastructure", "summary": "testing - separate tests in to short/long groups", "priority": "normal", "keywords": "", "time": "2015-08-14T16:40:21", "milestone": "Autumnal Equinox 2020", "owner": "olivas", "type": "defect" } ```
1.0
testing - separate tests in to short/long groups (Trac #1113) - I'll cmake this up if lists for each group are provided Migrated from https://code.icecube.wisc.edu/ticket/1113 ```json { "status": "closed", "changetime": "2020-06-24T12:31:42", "description": "I'll cmake this up if lists for each group are provided", "reporter": "nega", "cc": "", "resolution": "worksforme", "_ts": "1593001902142004", "component": "infrastructure", "summary": "testing - separate tests in to short/long groups", "priority": "normal", "keywords": "", "time": "2015-08-14T16:40:21", "milestone": "Autumnal Equinox 2020", "owner": "olivas", "type": "defect" } ```
non_priority
testing separate tests in to short long groups trac i ll cmake this up if lists for each group are provided migrated from json status closed changetime description i ll cmake this up if lists for each group are provided reporter nega cc resolution worksforme ts component infrastructure summary testing separate tests in to short long groups priority normal keywords time milestone autumnal equinox owner olivas type defect
0
753,453
26,347,167,625
IssuesEvent
2023-01-10 23:28:37
getindiekit/indiekit
https://api.github.com/repos/getindiekit/indiekit
closed
Syndication to Mastodon being triggered repeatedly for the same note
bug sponsor priority
## Describe the bug Successful POSTs to /syndicate are causing my first published post to be re-syndicated repeatedly. ## To reproduce Steps to reproduce the behaviour: 0. Configure an Indiekit server to syndicate to Mastodon. 1. Publish a few notes, syndicate each one using the "Syndicate" button in the Indiekit web UI. 2. Use RapidAPI (formerly PAW) to send an HTTP POST request to `https://<indiekit server>/syndicate?token=<valid token>` 3. The first syndicated post is re-syndicated. ## Expected behaviour Indiekit should syndicate each published post only once. Indiekit should not produce duplicate syndicated posts on Mastodon. ## Screenshots or logs Here's the contents of my database at the point where I am able to trigger this issue: <img width="1227" alt="CleanShot 2023-01-01 at 23 26 55@2x" src="https://user-images.githubusercontent.com/89772/210170564-c275b092-3e38-4246-a8da-3ef59cf464eb.png"> Here are three separate GitHub commits, all syndicating the same post: 1. [Original syndication](https://github.com/sentience/kevinyank.com/commit/1d9d6ceb84faf52af2b58aa2ed372b2a49cf485e) 2. [First duplicate syndication](https://github.com/sentience/kevinyank.com/commit/5734a7e6427f4a1de52db706135c3a9ed007881c) 3. [Second duplicate syndication](https://github.com/sentience/kevinyank.com/commit/4ece83e4e1573bb4fcd3bee4f52a1874ade68753) ## Environment - Indiekit version: v1.0.0-beta.1 - Browser: RapidAPI (previously known as Paw) - Server/host: Private Synology NAS - Content store URL: https://github.com/sentience/kevinyank.com/ - Publication URL: https://kevinyank.com/ ## Additional context Per [my comment](https://github.com/getindiekit/indiekit/issues/567#issuecomment-1368207792) on #567, I've been troubleshooting my Netlify outgoing deploy hook, which doesn't seem reliably to be triggering syndication from Indiekit. Strangely, POSTs originating from Netlify's outgoing deploy hook, which arrive with no `User-Agent` request header, receive an HTTP 200 response from Indiekit (91 bytes in size), and _do not_ seem to trigger syndication. Meanwhile, POSTs originating from my RapidAPI desktop HTTP client app, which arrive with a User-Agent header of `RapidAPI/4.0.0 (Macintosh; OS X/13.1.0) GCDHTTPRequest`, receive an HTTP 201 response from Indiekit (105 bytes in size) (indicating the creation of the syndicated post, I gather?). **Update:** I've ruled out the `User-Agent` header as a possible cause for the different behaviour. Sending a request from RapidAPI without the `User-Agent` header still causes the repeat syndication. ~The only other difference I'm seeing between the successful and unsuccessful requests is that the successful ones arrive with an `X-Forwarded-For` header (showing my computer's public IP address). I'm not sure why my local network requests from RapidAPI get an `X-Forwarded-For` while the ones hitting my Indiekit server from Netlify (via the same Nginx reverse proxy) do not. I'm guessing this has something to do with my Synology server's reverse proxy setup's handling of "loopback" requests originating from within my network. In any case, I don't know why this should be causing Indiekit to handle the requests differently.~ **Update:** I've eliminated the `X-Forwarded-For` header as a possible root cause. If I use RapidAPI to hit Indiekit directly via my local network and not via my Nginx reverse proxy, such that no `X-Forwarded-For` header is added to the request, Indiekit still performs a duplicate syndication. ~The `User-Agent` header at this point is still the most significant difference between the requests.~
1.0
Syndication to Mastodon being triggered repeatedly for the same note - ## Describe the bug Successful POSTs to /syndicate are causing my first published post to be re-syndicated repeatedly. ## To reproduce Steps to reproduce the behaviour: 0. Configure an Indiekit server to syndicate to Mastodon. 1. Publish a few notes, syndicate each one using the "Syndicate" button in the Indiekit web UI. 2. Use RapidAPI (formerly PAW) to send an HTTP POST request to `https://<indiekit server>/syndicate?token=<valid token>` 3. The first syndicated post is re-syndicated. ## Expected behaviour Indiekit should syndicate each published post only once. Indiekit should not produce duplicate syndicated posts on Mastodon. ## Screenshots or logs Here's the contents of my database at the point where I am able to trigger this issue: <img width="1227" alt="CleanShot 2023-01-01 at 23 26 55@2x" src="https://user-images.githubusercontent.com/89772/210170564-c275b092-3e38-4246-a8da-3ef59cf464eb.png"> Here are three separate GitHub commits, all syndicating the same post: 1. [Original syndication](https://github.com/sentience/kevinyank.com/commit/1d9d6ceb84faf52af2b58aa2ed372b2a49cf485e) 2. [First duplicate syndication](https://github.com/sentience/kevinyank.com/commit/5734a7e6427f4a1de52db706135c3a9ed007881c) 3. [Second duplicate syndication](https://github.com/sentience/kevinyank.com/commit/4ece83e4e1573bb4fcd3bee4f52a1874ade68753) ## Environment - Indiekit version: v1.0.0-beta.1 - Browser: RapidAPI (previously known as Paw) - Server/host: Private Synology NAS - Content store URL: https://github.com/sentience/kevinyank.com/ - Publication URL: https://kevinyank.com/ ## Additional context Per [my comment](https://github.com/getindiekit/indiekit/issues/567#issuecomment-1368207792) on #567, I've been troubleshooting my Netlify outgoing deploy hook, which doesn't seem reliably to be triggering syndication from Indiekit. Strangely, POSTs originating from Netlify's outgoing deploy hook, which arrive with no `User-Agent` request header, receive an HTTP 200 response from Indiekit (91 bytes in size), and _do not_ seem to trigger syndication. Meanwhile, POSTs originating from my RapidAPI desktop HTTP client app, which arrive with a User-Agent header of `RapidAPI/4.0.0 (Macintosh; OS X/13.1.0) GCDHTTPRequest`, receive an HTTP 201 response from Indiekit (105 bytes in size) (indicating the creation of the syndicated post, I gather?). **Update:** I've ruled out the `User-Agent` header as a possible cause for the different behaviour. Sending a request from RapidAPI without the `User-Agent` header still causes the repeat syndication. ~The only other difference I'm seeing between the successful and unsuccessful requests is that the successful ones arrive with an `X-Forwarded-For` header (showing my computer's public IP address). I'm not sure why my local network requests from RapidAPI get an `X-Forwarded-For` while the ones hitting my Indiekit server from Netlify (via the same Nginx reverse proxy) do not. I'm guessing this has something to do with my Synology server's reverse proxy setup's handling of "loopback" requests originating from within my network. In any case, I don't know why this should be causing Indiekit to handle the requests differently.~ **Update:** I've eliminated the `X-Forwarded-For` header as a possible root cause. If I use RapidAPI to hit Indiekit directly via my local network and not via my Nginx reverse proxy, such that no `X-Forwarded-For` header is added to the request, Indiekit still performs a duplicate syndication. ~The `User-Agent` header at this point is still the most significant difference between the requests.~
priority
syndication to mastodon being triggered repeatedly for the same note describe the bug successful posts to syndicate are causing my first published post to be re syndicated repeatedly to reproduce steps to reproduce the behaviour configure an indiekit server to syndicate to mastodon publish a few notes syndicate each one using the syndicate button in the indiekit web ui use rapidapi formerly paw to send an http post request to server syndicate token the first syndicated post is re syndicated expected behaviour indiekit should syndicate each published post only once indiekit should not produce duplicate syndicated posts on mastodon screenshots or logs here s the contents of my database at the point where i am able to trigger this issue img width alt cleanshot at src here are three separate github commits all syndicating the same post environment indiekit version beta browser rapidapi previously known as paw server host private synology nas content store url publication url additional context per on i ve been troubleshooting my netlify outgoing deploy hook which doesn t seem reliably to be triggering syndication from indiekit strangely posts originating from netlify s outgoing deploy hook which arrive with no user agent request header receive an http response from indiekit bytes in size and do not seem to trigger syndication meanwhile posts originating from my rapidapi desktop http client app which arrive with a user agent header of rapidapi macintosh os x gcdhttprequest receive an http response from indiekit bytes in size indicating the creation of the syndicated post i gather update i ve ruled out the user agent header as a possible cause for the different behaviour sending a request from rapidapi without the user agent header still causes the repeat syndication the only other difference i m seeing between the successful and unsuccessful requests is that the successful ones arrive with an x forwarded for header showing my computer s public ip address i m not sure why my local network requests from rapidapi get an x forwarded for while the ones hitting my indiekit server from netlify via the same nginx reverse proxy do not i m guessing this has something to do with my synology server s reverse proxy setup s handling of loopback requests originating from within my network in any case i don t know why this should be causing indiekit to handle the requests differently update i ve eliminated the x forwarded for header as a possible root cause if i use rapidapi to hit indiekit directly via my local network and not via my nginx reverse proxy such that no x forwarded for header is added to the request indiekit still performs a duplicate syndication the user agent header at this point is still the most significant difference between the requests
1
663,870
22,209,705,764
IssuesEvent
2022-06-07 17:59:53
dotnet/maui
https://api.github.com/repos/dotnet/maui
closed
WPF BlazorWebView TabStopOrder not working
area/blazor 🕸️ feature-blazor-windesktop discussed Priority:1
### Describe the bug I played arround with WPF and BlazorWebView, since I know Keyboard control is VERY importaint for powerusers I checked if the tab stop interaction works with the web view: As you see the tab order should be First the Button, then the Button inside the BlazorWebView and then the Combo Box which drops down over the webside. But the Website is always last, would it be possible to make the browser control to accept tab stop? PS: Have not yet checked how to make keyboard shortcut interactions. ``` <StackPanel Grid.Row="0" Orientation="Horizontal"> <TextBlock VerticalAlignment="Center" Text="ComboBox: " /> <ComboBox TabIndex="2" Margin="2,5,0,5" Width="80"> <ComboBoxItem>Item 1</ComboBoxItem> <ComboBoxItem>Item 2</ComboBoxItem> <ComboBoxItem>Item 3</ComboBoxItem> <ComboBoxItem>Item 4</ComboBoxItem> <ComboBoxItem>Item 5</ComboBoxItem> <ComboBoxItem>Item 6</ComboBoxItem> </ComboBox> <TextBlock VerticalAlignment="Center" Margin="5,0,0,0" Text="Button Tab Index: " /> <Button TabIndex="0" Margin="2,5,0,5">ButtonTabIdx3</Button> </StackPanel> <blazor:BlazorWebView TabIndex="1" IsTabStop="True" Grid.Row="1" HostPage="wwwroot/index.html" Services="{StaticResource services}"> <blazor:BlazorWebView.RootComponents> <blazor:RootComponent Selector="#app" ComponentType="{x:Type local:Counter}" /> </blazor:BlazorWebView.RootComponents> </blazor:BlazorWebView> ``` ### To Reproduce In this sample https://github.com/MichaelPeter/BlazorDesktopWPFTestTabStop/blob/master/README.md I made it reproducable ### Exceptions (if any) <!-- Include the exception you get when facing this issue --> ### Further technical details - ASP.NET Core version - Include the output of `dotnet --info` - The IDE (VS / VS Code/ VS4Mac) you're running on, and its version
1.0
WPF BlazorWebView TabStopOrder not working - ### Describe the bug I played arround with WPF and BlazorWebView, since I know Keyboard control is VERY importaint for powerusers I checked if the tab stop interaction works with the web view: As you see the tab order should be First the Button, then the Button inside the BlazorWebView and then the Combo Box which drops down over the webside. But the Website is always last, would it be possible to make the browser control to accept tab stop? PS: Have not yet checked how to make keyboard shortcut interactions. ``` <StackPanel Grid.Row="0" Orientation="Horizontal"> <TextBlock VerticalAlignment="Center" Text="ComboBox: " /> <ComboBox TabIndex="2" Margin="2,5,0,5" Width="80"> <ComboBoxItem>Item 1</ComboBoxItem> <ComboBoxItem>Item 2</ComboBoxItem> <ComboBoxItem>Item 3</ComboBoxItem> <ComboBoxItem>Item 4</ComboBoxItem> <ComboBoxItem>Item 5</ComboBoxItem> <ComboBoxItem>Item 6</ComboBoxItem> </ComboBox> <TextBlock VerticalAlignment="Center" Margin="5,0,0,0" Text="Button Tab Index: " /> <Button TabIndex="0" Margin="2,5,0,5">ButtonTabIdx3</Button> </StackPanel> <blazor:BlazorWebView TabIndex="1" IsTabStop="True" Grid.Row="1" HostPage="wwwroot/index.html" Services="{StaticResource services}"> <blazor:BlazorWebView.RootComponents> <blazor:RootComponent Selector="#app" ComponentType="{x:Type local:Counter}" /> </blazor:BlazorWebView.RootComponents> </blazor:BlazorWebView> ``` ### To Reproduce In this sample https://github.com/MichaelPeter/BlazorDesktopWPFTestTabStop/blob/master/README.md I made it reproducable ### Exceptions (if any) <!-- Include the exception you get when facing this issue --> ### Further technical details - ASP.NET Core version - Include the output of `dotnet --info` - The IDE (VS / VS Code/ VS4Mac) you're running on, and its version
priority
wpf blazorwebview tabstoporder not working describe the bug i played arround with wpf and blazorwebview since i know keyboard control is very importaint for powerusers i checked if the tab stop interaction works with the web view as you see the tab order should be first the button then the button inside the blazorwebview and then the combo box which drops down over the webside but the website is always last would it be possible to make the browser control to accept tab stop ps have not yet checked how to make keyboard shortcut interactions item item item item item item to reproduce in this sample i made it reproducable exceptions if any include the exception you get when facing this issue further technical details asp net core version include the output of dotnet info the ide vs vs code you re running on and its version
1
376,240
11,140,068,936
IssuesEvent
2019-12-21 11:14:12
wilcommerce/Wilcommerce.Catalog
https://api.github.com/repos/wilcommerce/Wilcommerce.Catalog
opened
Category Set as visibile doesn't set values properly
priority:high status:backlog type:bug
In the Category entity, SetAsVisibile(DateTime, DateTime) method should be changed accepting nullable values.
1.0
Category Set as visibile doesn't set values properly - In the Category entity, SetAsVisibile(DateTime, DateTime) method should be changed accepting nullable values.
priority
category set as visibile doesn t set values properly in the category entity setasvisibile datetime datetime method should be changed accepting nullable values
1
3,282
2,754,160,201
IssuesEvent
2015-04-25 11:41:29
energia/Energia
https://api.github.com/repos/energia/Energia
opened
Provide Guidelines to Add Support for New MCU
Board Others IDE Documentation IDE Energia Type Feature request
There are numerous tickets related to `Add Support for New MCU`. As Energia –and its team– can't obviously add support for all the MCUs and variants from the extensive Texas Instruments portfolio, why not provide guidelines to help enthusiasts to do so? More specifically: * Structure of the folders * List of files impacted * Conventions
1.0
Provide Guidelines to Add Support for New MCU - There are numerous tickets related to `Add Support for New MCU`. As Energia –and its team– can't obviously add support for all the MCUs and variants from the extensive Texas Instruments portfolio, why not provide guidelines to help enthusiasts to do so? More specifically: * Structure of the folders * List of files impacted * Conventions
non_priority
provide guidelines to add support for new mcu there are numerous tickets related to add support for new mcu as energia –and its team– can t obviously add support for all the mcus and variants from the extensive texas instruments portfolio why not provide guidelines to help enthusiasts to do so more specifically structure of the folders list of files impacted conventions
0
595,718
18,073,803,848
IssuesEvent
2021-09-21 07:33:27
googleapis/gaxios
https://api.github.com/repos/googleapis/gaxios
opened
🥁 configuration options: should merge URL parameters with the params option failed
type: bug priority: p1 flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 0666c1d980d948e1759683b750cf6811c9aace68 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/0f6a0074-4780-47f5-8899-589413740226), [Sponge](http://sponge2/0f6a0074-4780-47f5-8899-589413740226) status: failed <details><summary>Test output</summary><br><pre>request to https://example.com/?james=beckwith&montgomery=scott&james=kirk failed, reason: Nock: No match for request { "method": "GET", "url": "https://example.com/", "headers": { "accept": [ "application/json" ], "user-agent": [ "node-fetch/1.0 (+https://github.com/bitinn/node-fetch)" ], "accept-encoding": [ "gzip,deflate" ], "connection": [ "close" ] } } FetchError: request to https://example.com/?james=beckwith&montgomery=scott&james=kirk failed, reason: Nock: No match for request { "method": "GET", "url": "https://example.com/", "headers": { "accept": [ "application/json" ], "user-agent": [ "node-fetch/1.0 (+https://github.com/bitinn/node-fetch)" ], "accept-encoding": [ "gzip,deflate" ], "connection": [ "close" ] } } at OverriddenClientRequest.<anonymous> (node_modules/node-fetch/lib/index.js:1494:11) at Socket.<anonymous> (node_modules/propagate/index.js:64:17) at process.nextTick (node_modules/nock/lib/socket.js:100:14) at process._tickCallback (internal/process/next_tick.js:61:11)</pre></details>
1.0
🥁 configuration options: should merge URL parameters with the params option failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 0666c1d980d948e1759683b750cf6811c9aace68 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/0f6a0074-4780-47f5-8899-589413740226), [Sponge](http://sponge2/0f6a0074-4780-47f5-8899-589413740226) status: failed <details><summary>Test output</summary><br><pre>request to https://example.com/?james=beckwith&montgomery=scott&james=kirk failed, reason: Nock: No match for request { "method": "GET", "url": "https://example.com/", "headers": { "accept": [ "application/json" ], "user-agent": [ "node-fetch/1.0 (+https://github.com/bitinn/node-fetch)" ], "accept-encoding": [ "gzip,deflate" ], "connection": [ "close" ] } } FetchError: request to https://example.com/?james=beckwith&montgomery=scott&james=kirk failed, reason: Nock: No match for request { "method": "GET", "url": "https://example.com/", "headers": { "accept": [ "application/json" ], "user-agent": [ "node-fetch/1.0 (+https://github.com/bitinn/node-fetch)" ], "accept-encoding": [ "gzip,deflate" ], "connection": [ "close" ] } } at OverriddenClientRequest.<anonymous> (node_modules/node-fetch/lib/index.js:1494:11) at Socket.<anonymous> (node_modules/propagate/index.js:64:17) at process.nextTick (node_modules/nock/lib/socket.js:100:14) at process._tickCallback (internal/process/next_tick.js:61:11)</pre></details>
priority
🥁 configuration options should merge url parameters with the params option failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output request to failed reason nock no match for request method get url headers accept application json user agent node fetch accept encoding gzip deflate connection close fetcherror request to failed reason nock no match for request method get url headers accept application json user agent node fetch accept encoding gzip deflate connection close at overriddenclientrequest node modules node fetch lib index js at socket node modules propagate index js at process nexttick node modules nock lib socket js at process tickcallback internal process next tick js
1
535,677
15,696,290,487
IssuesEvent
2021-03-26 01:40:15
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
[AsyncAPI] Topics page menu is not highlighted when in the Topics page
API-M 4.0.0 Feature/AsyncAPIs Priority/Highest Type/Bug
### Description: When in the Topics page of the Publisher UI, the Topic menu item is not highlighted. ### Affected Product Version: APIM 4.0.0-alpha
1.0
[AsyncAPI] Topics page menu is not highlighted when in the Topics page - ### Description: When in the Topics page of the Publisher UI, the Topic menu item is not highlighted. ### Affected Product Version: APIM 4.0.0-alpha
priority
topics page menu is not highlighted when in the topics page description when in the topics page of the publisher ui the topic menu item is not highlighted affected product version apim alpha
1
21,261
28,434,726,608
IssuesEvent
2023-04-15 06:48:43
GoogleCloudPlatform/pgadapter
https://api.github.com/repos/GoogleCloudPlatform/pgadapter
closed
Test failure for random date tests
type: process priority: p3
``` AbortedMockServerTest.testRandomResults:876->assertEqual:931 expected:<1582-10-12> but was:<1582-10-22> ``` The random results tests can fail if a random date during the julian/gregorian cut-off is selected.
1.0
Test failure for random date tests - ``` AbortedMockServerTest.testRandomResults:876->assertEqual:931 expected:<1582-10-12> but was:<1582-10-22> ``` The random results tests can fail if a random date during the julian/gregorian cut-off is selected.
non_priority
test failure for random date tests abortedmockservertest testrandomresults assertequal expected but was the random results tests can fail if a random date during the julian gregorian cut off is selected
0
189,310
6,796,430,196
IssuesEvent
2017-11-01 18:59:14
Luigi1000/SOEN-341_Project_Team-A
https://api.github.com/repos/Luigi1000/SOEN-341_Project_Team-A
closed
Search also based on the main category and cityName
feature: header FrontEnd high priority
add categories option in drop down list in search bar add cityName option in in drop down list in search bar #95
1.0
Search also based on the main category and cityName - add categories option in drop down list in search bar add cityName option in in drop down list in search bar #95
priority
search also based on the main category and cityname add categories option in drop down list in search bar add cityname option in in drop down list in search bar
1
343,022
10,324,457,566
IssuesEvent
2019-09-01 09:19:29
wso2/product-is
https://api.github.com/repos/wso2/product-is
opened
Bug in token endpoint invalid_request error_description
Affected/5.9.0-Alpha Component/OAuth Priority/High Severity/Major Type/Bug
When sending an invalid request to the token endpoint to get a token. The error_description is in the wrong format. ``` { "error_description": "OAuthProblemException{error='unsupported_grant_type', description='Unsupported grant_type value', uri='null', state='null', scope='null', redirectUri='null', responseStatus=0, parameters={}}", "error": "invalid_request" } ``` To reproduce. Create an OAuth app and call the following cURL with the client credentials. ``` curl -X POST https://localhost:9443/oauth2/token -H 'Authorization: Basic <base64(clientid:secret)>' -d 'grant_type=grant_type=password' ```
1.0
Bug in token endpoint invalid_request error_description - When sending an invalid request to the token endpoint to get a token. The error_description is in the wrong format. ``` { "error_description": "OAuthProblemException{error='unsupported_grant_type', description='Unsupported grant_type value', uri='null', state='null', scope='null', redirectUri='null', responseStatus=0, parameters={}}", "error": "invalid_request" } ``` To reproduce. Create an OAuth app and call the following cURL with the client credentials. ``` curl -X POST https://localhost:9443/oauth2/token -H 'Authorization: Basic <base64(clientid:secret)>' -d 'grant_type=grant_type=password' ```
priority
bug in token endpoint invalid request error description when sending an invalid request to the token endpoint to get a token the error description is in the wrong format error description oauthproblemexception error unsupported grant type description unsupported grant type value uri null state null scope null redirecturi null responsestatus parameters error invalid request to reproduce create an oauth app and call the following curl with the client credentials curl x post h authorization basic d grant type grant type password
1
86,481
10,500,566,027
IssuesEvent
2019-09-26 10:48:07
containous/traefik
https://api.github.com/repos/containous/traefik
closed
Dashboard does not initialize with --network="host"
area/api area/documentation area/provider/docker area/webui kind/enhancement
<!-- DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS. The issue tracker is for reporting bugs and feature requests only. For end-user related support questions, please refer to one of the following: - the Traefik community forum: https://community.containo.us/ --> ### Do you want to request a *feature* or report a *bug*? Bug ### Did you try using a 1.7.x configuration for the version 2.0? - [ ] Yes - [x] No <!-- If you just checked the "Yes" box, be aware that this is probably not a bug. The configurations between 1.X and 2.X are NOT compatible. Please have a look here https://docs.traefik.io/v2.0/getting-started/configuration-overview/. --> ### What did you do? ``` docker run --rm \ --name="traefik" \ --label="traefik.enable=true" \ --label="traefik.http.routers.traefik.entrypoints=http" \ --label="traefik.http.routers.traefik.rule=Host(\`traefik.localhost\`)" \ --label="traefik.http.routers.traefik.service=api@internal" \ --volume="/var/run/docker.sock:/var/run/docker.sock" \ --network="host" \ traefik:2.0 \ "--api=true" \ "--api.dashboard=true" \ "--log.level=DEBUG" \ "--providers.docker=true" \ "--providers.docker.exposedbydefault=false" \ "--entrypoints.http.address=:80" ``` <!-- HOW TO WRITE A GOOD BUG REPORT? - Respect the issue template as much as possible. - The title should be short and descriptive. - Explain the conditions which led you to report this issue: the context. - The context should lead to something, an idea or a problem that you’re facing. - Remain clear and concise. - Format your messages to help the reader focus on what matters and understand the structure of your message, use Markdown syntax https://help.github.com/articles/github-flavored-markdown --> ### What did you expect to see? Dashboard available at `http://traefik.localhost:80`. ### What did you see instead? ``` time="2019-09-22T20:48:54Z" level=info msg="Configuration loaded from flags." time="2019-09-22T20:48:54Z" level=info msg="Traefik version 2.0.0 built on 2019-09-16T17:35:17Z" time="2019-09-22T20:48:54Z" level=debug msg="Static configuration loaded {\"global\":{\"checkNewVersion\":true},\"serversTransport\":{\"maxIdleConnsPerHost\":200},\"entryPoints\":{\"http\":{\"address\":\":80\",\"transport\":{\"lifeCycle\":{\"graceTimeOut\":10000000000},\"respondingTimeouts\":{\"idleTimeout\":180000000000}},\"forwardedHeaders\":{}}},\"providers\":{\"providersThrottleDuration\":2000000000,\"docker\":{\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"swarmModeRefreshSeconds\":15000000000}},\"api\":{\"dashboard\":true},\"log\":{\"level\":\"DEBUG\",\"format\":\"common\"}}" time="2019-09-22T20:48:54Z" level=info msg="\nStats collection is disabled.\nHelp us improve Traefik by turning this feature on :)\nMore details on: https://docs.traefik.io/v2.0/contributing/data-collection/\n" time="2019-09-22T20:48:54Z" level=debug msg="No default certificate, generating one" time="2019-09-22T20:48:54Z" level=info msg="Starting provider aggregator.ProviderAggregator {}" time="2019-09-22T20:48:54Z" level=debug msg="Start TCP Server" entryPointName=http time="2019-09-22T20:48:54Z" level=info msg="Starting provider *docker.Provider {\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"swarmModeRefreshSeconds\":15000000000}" time="2019-09-22T20:48:54Z" level=debug msg="Provider connection established with docker 19.03.1 (API 1.40)" providerName=docker time="2019-09-22T20:48:54Z" level=error msg="port is missing" providerName=docker container=/traefik-08181dbd4297d897e52e150c0748d82e571c6542553ba1932150eef02262f526 time="2019-09-22T20:48:54Z" level=debug msg="Configuration received from provider docker: {\"http\":{},\"tcp\":{}}" providerName=docker time="2019-09-22T20:48:54Z" level=debug msg="No default certificate, generating one" ``` Remove `--network="host"` and Dashboard initializes as expected: ``` time="2019-09-22T20:49:46Z" level=info msg="Configuration loaded from flags." time="2019-09-22T20:49:46Z" level=info msg="Traefik version 2.0.0 built on 2019-09-16T17:35:17Z" time="2019-09-22T20:49:46Z" level=debug msg="Static configuration loaded {\"global\":{\"checkNewVersion\":true},\"serversTransport\":{\"maxIdleConnsPerHost\":200},\"entryPoints\":{\"http\":{\"address\":\":80\",\"transport\":{\"lifeCycle\":{\"graceTimeOut\":10000000000},\"respondingTimeouts\":{\"idleTimeout\":180000000000}},\"forwardedHeaders\":{}}},\"providers\":{\"providersThrottleDuration\":2000000000,\"docker\":{\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"swarmModeRefreshSeconds\":15000000000}},\"api\":{\"dashboard\":true},\"log\":{\"level\":\"DEBUG\",\"format\":\"common\"}}" time="2019-09-22T20:49:46Z" level=info msg="\nStats collection is disabled.\nHelp us improve Traefik by turning this feature on :)\nMore details on: https://docs.traefik.io/v2.0/contributing/data-collection/\n" time="2019-09-22T20:49:46Z" level=debug msg="No default certificate, generating one" time="2019-09-22T20:49:47Z" level=info msg="Starting provider aggregator.ProviderAggregator {}" time="2019-09-22T20:49:47Z" level=debug msg="Start TCP Server" entryPointName=http time="2019-09-22T20:49:47Z" level=info msg="Starting provider *docker.Provider {\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"swarmModeRefreshSeconds\":15000000000}" time="2019-09-22T20:49:47Z" level=debug msg="Provider connection established with docker 19.03.1 (API 1.40)" providerName=docker time="2019-09-22T20:49:47Z" level=debug msg="Configuration received from provider docker: {\"http\":{\"routers\":{\"traefik\":{\"entryPoints\":[\"http\"],\"service\":\"api@internal\",\"rule\":\"Host(`traefik.localhost`)\"}},\"services\":{\"/traefik\":{\"loadBalancer\":{\"servers\":[{\"url\":\"http://10.255.0.2:80\"}],\"passHostHeader\":true}}}},\"tcp\":{}}" providerName=docker time="2019-09-22T20:49:47Z" level=debug msg="Added outgoing tracing middleware api@internal" entryPointName=http middlewareName=tracing middlewareType=TracingForwarder routerName=traefik@docker time="2019-09-22T20:49:47Z" level=debug msg="Creating middleware" entryPointName=http middlewareType=Recovery middlewareName=traefik-internal-recovery time="2019-09-22T20:49:47Z" level=debug msg="No default certificate, generating one" ``` ### Output of `traefik version`: (_What version of Traefik are you using?_) <!-- For the Traefik Docker image: docker run [IMAGE] version ex: docker run traefik version For the alpine Traefik Docker image: docker run [IMAGE] traefik version ex: docker run traefik traefik version --> ``` Version: 2.0.0 Codename: montdor Go version: go1.13 Built: 2019-09-16T17:35:17Z OS/Arch: linux/amd64 ``` ### What is your environment & configuration (arguments, toml, provider, platform, ...)? ``` N/A ``` <!-- Add more configuration information here. --> ### If applicable, please paste the log output in DEBUG level (`--log.level=DEBUG` switch) ``` N/A ```
1.0
Dashboard does not initialize with --network="host" - <!-- DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS. The issue tracker is for reporting bugs and feature requests only. For end-user related support questions, please refer to one of the following: - the Traefik community forum: https://community.containo.us/ --> ### Do you want to request a *feature* or report a *bug*? Bug ### Did you try using a 1.7.x configuration for the version 2.0? - [ ] Yes - [x] No <!-- If you just checked the "Yes" box, be aware that this is probably not a bug. The configurations between 1.X and 2.X are NOT compatible. Please have a look here https://docs.traefik.io/v2.0/getting-started/configuration-overview/. --> ### What did you do? ``` docker run --rm \ --name="traefik" \ --label="traefik.enable=true" \ --label="traefik.http.routers.traefik.entrypoints=http" \ --label="traefik.http.routers.traefik.rule=Host(\`traefik.localhost\`)" \ --label="traefik.http.routers.traefik.service=api@internal" \ --volume="/var/run/docker.sock:/var/run/docker.sock" \ --network="host" \ traefik:2.0 \ "--api=true" \ "--api.dashboard=true" \ "--log.level=DEBUG" \ "--providers.docker=true" \ "--providers.docker.exposedbydefault=false" \ "--entrypoints.http.address=:80" ``` <!-- HOW TO WRITE A GOOD BUG REPORT? - Respect the issue template as much as possible. - The title should be short and descriptive. - Explain the conditions which led you to report this issue: the context. - The context should lead to something, an idea or a problem that you’re facing. - Remain clear and concise. - Format your messages to help the reader focus on what matters and understand the structure of your message, use Markdown syntax https://help.github.com/articles/github-flavored-markdown --> ### What did you expect to see? Dashboard available at `http://traefik.localhost:80`. ### What did you see instead? ``` time="2019-09-22T20:48:54Z" level=info msg="Configuration loaded from flags." time="2019-09-22T20:48:54Z" level=info msg="Traefik version 2.0.0 built on 2019-09-16T17:35:17Z" time="2019-09-22T20:48:54Z" level=debug msg="Static configuration loaded {\"global\":{\"checkNewVersion\":true},\"serversTransport\":{\"maxIdleConnsPerHost\":200},\"entryPoints\":{\"http\":{\"address\":\":80\",\"transport\":{\"lifeCycle\":{\"graceTimeOut\":10000000000},\"respondingTimeouts\":{\"idleTimeout\":180000000000}},\"forwardedHeaders\":{}}},\"providers\":{\"providersThrottleDuration\":2000000000,\"docker\":{\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"swarmModeRefreshSeconds\":15000000000}},\"api\":{\"dashboard\":true},\"log\":{\"level\":\"DEBUG\",\"format\":\"common\"}}" time="2019-09-22T20:48:54Z" level=info msg="\nStats collection is disabled.\nHelp us improve Traefik by turning this feature on :)\nMore details on: https://docs.traefik.io/v2.0/contributing/data-collection/\n" time="2019-09-22T20:48:54Z" level=debug msg="No default certificate, generating one" time="2019-09-22T20:48:54Z" level=info msg="Starting provider aggregator.ProviderAggregator {}" time="2019-09-22T20:48:54Z" level=debug msg="Start TCP Server" entryPointName=http time="2019-09-22T20:48:54Z" level=info msg="Starting provider *docker.Provider {\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"swarmModeRefreshSeconds\":15000000000}" time="2019-09-22T20:48:54Z" level=debug msg="Provider connection established with docker 19.03.1 (API 1.40)" providerName=docker time="2019-09-22T20:48:54Z" level=error msg="port is missing" providerName=docker container=/traefik-08181dbd4297d897e52e150c0748d82e571c6542553ba1932150eef02262f526 time="2019-09-22T20:48:54Z" level=debug msg="Configuration received from provider docker: {\"http\":{},\"tcp\":{}}" providerName=docker time="2019-09-22T20:48:54Z" level=debug msg="No default certificate, generating one" ``` Remove `--network="host"` and Dashboard initializes as expected: ``` time="2019-09-22T20:49:46Z" level=info msg="Configuration loaded from flags." time="2019-09-22T20:49:46Z" level=info msg="Traefik version 2.0.0 built on 2019-09-16T17:35:17Z" time="2019-09-22T20:49:46Z" level=debug msg="Static configuration loaded {\"global\":{\"checkNewVersion\":true},\"serversTransport\":{\"maxIdleConnsPerHost\":200},\"entryPoints\":{\"http\":{\"address\":\":80\",\"transport\":{\"lifeCycle\":{\"graceTimeOut\":10000000000},\"respondingTimeouts\":{\"idleTimeout\":180000000000}},\"forwardedHeaders\":{}}},\"providers\":{\"providersThrottleDuration\":2000000000,\"docker\":{\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"swarmModeRefreshSeconds\":15000000000}},\"api\":{\"dashboard\":true},\"log\":{\"level\":\"DEBUG\",\"format\":\"common\"}}" time="2019-09-22T20:49:46Z" level=info msg="\nStats collection is disabled.\nHelp us improve Traefik by turning this feature on :)\nMore details on: https://docs.traefik.io/v2.0/contributing/data-collection/\n" time="2019-09-22T20:49:46Z" level=debug msg="No default certificate, generating one" time="2019-09-22T20:49:47Z" level=info msg="Starting provider aggregator.ProviderAggregator {}" time="2019-09-22T20:49:47Z" level=debug msg="Start TCP Server" entryPointName=http time="2019-09-22T20:49:47Z" level=info msg="Starting provider *docker.Provider {\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"swarmModeRefreshSeconds\":15000000000}" time="2019-09-22T20:49:47Z" level=debug msg="Provider connection established with docker 19.03.1 (API 1.40)" providerName=docker time="2019-09-22T20:49:47Z" level=debug msg="Configuration received from provider docker: {\"http\":{\"routers\":{\"traefik\":{\"entryPoints\":[\"http\"],\"service\":\"api@internal\",\"rule\":\"Host(`traefik.localhost`)\"}},\"services\":{\"/traefik\":{\"loadBalancer\":{\"servers\":[{\"url\":\"http://10.255.0.2:80\"}],\"passHostHeader\":true}}}},\"tcp\":{}}" providerName=docker time="2019-09-22T20:49:47Z" level=debug msg="Added outgoing tracing middleware api@internal" entryPointName=http middlewareName=tracing middlewareType=TracingForwarder routerName=traefik@docker time="2019-09-22T20:49:47Z" level=debug msg="Creating middleware" entryPointName=http middlewareType=Recovery middlewareName=traefik-internal-recovery time="2019-09-22T20:49:47Z" level=debug msg="No default certificate, generating one" ``` ### Output of `traefik version`: (_What version of Traefik are you using?_) <!-- For the Traefik Docker image: docker run [IMAGE] version ex: docker run traefik version For the alpine Traefik Docker image: docker run [IMAGE] traefik version ex: docker run traefik traefik version --> ``` Version: 2.0.0 Codename: montdor Go version: go1.13 Built: 2019-09-16T17:35:17Z OS/Arch: linux/amd64 ``` ### What is your environment & configuration (arguments, toml, provider, platform, ...)? ``` N/A ``` <!-- Add more configuration information here. --> ### If applicable, please paste the log output in DEBUG level (`--log.level=DEBUG` switch) ``` N/A ```
non_priority
dashboard does not initialize with network host do not file issues for general support questions the issue tracker is for reporting bugs and feature requests only for end user related support questions please refer to one of the following the traefik community forum do you want to request a feature or report a bug bug did you try using a x configuration for the version yes no if you just checked the yes box be aware that this is probably not a bug the configurations between x and x are not compatible please have a look here what did you do docker run rm name traefik label traefik enable true label traefik http routers traefik entrypoints http label traefik http routers traefik rule host traefik localhost label traefik http routers traefik service api internal volume var run docker sock var run docker sock network host traefik api true api dashboard true log level debug providers docker true providers docker exposedbydefault false entrypoints http address how to write a good bug report respect the issue template as much as possible the title should be short and descriptive explain the conditions which led you to report this issue the context the context should lead to something an idea or a problem that you’re facing remain clear and concise format your messages to help the reader focus on what matters and understand the structure of your message use markdown syntax what did you expect to see dashboard available at what did you see instead time level info msg configuration loaded from flags time level info msg traefik version built on time level debug msg static configuration loaded global checknewversion true serverstransport maxidleconnsperhost entrypoints http address transport lifecycle gracetimeout respondingtimeouts idletimeout forwardedheaders providers providersthrottleduration docker watch true endpoint unix var run docker sock defaultrule host normalize name swarmmoderefreshseconds api dashboard true log level debug format common time level info msg nstats collection is disabled nhelp us improve traefik by turning this feature on nmore details on time level debug msg no default certificate generating one time level info msg starting provider aggregator provideraggregator time level debug msg start tcp server entrypointname http time level info msg starting provider docker provider watch true endpoint unix var run docker sock defaultrule host normalize name swarmmoderefreshseconds time level debug msg provider connection established with docker api providername docker time level error msg port is missing providername docker container traefik time level debug msg configuration received from provider docker http tcp providername docker time level debug msg no default certificate generating one remove network host and dashboard initializes as expected time level info msg configuration loaded from flags time level info msg traefik version built on time level debug msg static configuration loaded global checknewversion true serverstransport maxidleconnsperhost entrypoints http address transport lifecycle gracetimeout respondingtimeouts idletimeout forwardedheaders providers providersthrottleduration docker watch true endpoint unix var run docker sock defaultrule host normalize name swarmmoderefreshseconds api dashboard true log level debug format common time level info msg nstats collection is disabled nhelp us improve traefik by turning this feature on nmore details on time level debug msg no default certificate generating one time level info msg starting provider aggregator provideraggregator time level debug msg start tcp server entrypointname http time level info msg starting provider docker provider watch true endpoint unix var run docker sock defaultrule host normalize name swarmmoderefreshseconds time level debug msg provider connection established with docker api providername docker time level debug msg configuration received from provider docker http routers traefik entrypoints service api internal rule host traefik localhost services traefik loadbalancer servers passhostheader true tcp providername docker time level debug msg added outgoing tracing middleware api internal entrypointname http middlewarename tracing middlewaretype tracingforwarder routername traefik docker time level debug msg creating middleware entrypointname http middlewaretype recovery middlewarename traefik internal recovery time level debug msg no default certificate generating one output of traefik version what version of traefik are you using for the traefik docker image docker run version ex docker run traefik version for the alpine traefik docker image docker run traefik version ex docker run traefik traefik version version codename montdor go version built os arch linux what is your environment configuration arguments toml provider platform n a add more configuration information here if applicable please paste the log output in debug level log level debug switch n a
0
350,378
10,483,220,491
IssuesEvent
2019-09-24 13:38:03
jenkins-x/jx
https://api.github.com/repos/jenkins-x/jx
closed
JX Boot can duplicate env variables when running the master build
area/boot kind/bug priority/important-soon
### Summary When running jx boot to apply changes to the cluster, its possible to get duplicate environment variables added to the jenkins-x.yml (GIT_AUTHOR_NAME, GIT_AUTHOR_EMAIL), this can lead to a local commit existing that breaks the `verify env` step
1.0
JX Boot can duplicate env variables when running the master build - ### Summary When running jx boot to apply changes to the cluster, its possible to get duplicate environment variables added to the jenkins-x.yml (GIT_AUTHOR_NAME, GIT_AUTHOR_EMAIL), this can lead to a local commit existing that breaks the `verify env` step
priority
jx boot can duplicate env variables when running the master build summary when running jx boot to apply changes to the cluster its possible to get duplicate environment variables added to the jenkins x yml git author name git author email this can lead to a local commit existing that breaks the verify env step
1
120,605
12,092,133,624
IssuesEvent
2020-04-19 14:30:45
whisst/World-Time
https://api.github.com/repos/whisst/World-Time
closed
Documents(SRS, SDD)
documentation
- We have to improve our Software system requirement and system design definication. - This project is the first time of us in software engineering so we get some trouble such as: donnot understand some terminology of software engineering, lack of scopes, etc...
1.0
Documents(SRS, SDD) - - We have to improve our Software system requirement and system design definication. - This project is the first time of us in software engineering so we get some trouble such as: donnot understand some terminology of software engineering, lack of scopes, etc...
non_priority
documents srs sdd we have to improve our software system requirement and system design definication this project is the first time of us in software engineering so we get some trouble such as donnot understand some terminology of software engineering lack of scopes etc
0
58,262
3,088,377,177
IssuesEvent
2015-08-25 16:16:40
pavel-pimenov/flylinkdc-r5xx
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
opened
В $SR передается не IP а доменное имя
bug imported Priority-Medium
_From [Pavel.Pimenov@gmail.com](https://code.google.com/u/Pavel.Pimenov@gmail.com/) on March 13, 2014 23:26:02_ Пример $SR chen video\multfilm\Ну, погоди!\Ну, погоди! 2.avi33492992 5/5TTH:B4O5M74UPKZ7I23CH36NA3SZOUZTJLWNVEIJMTQ (dc.a-galaxy.com:411)| Во флае фунция ClientManager::findHub(hubIpPort) ломается TODO разобраться в чем баг Если верить этому http://mydc.ru/topic915s20.html?p=6845#entry6845 то там может быть только IP _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1443_
1.0
В $SR передается не IP а доменное имя - _From [Pavel.Pimenov@gmail.com](https://code.google.com/u/Pavel.Pimenov@gmail.com/) on March 13, 2014 23:26:02_ Пример $SR chen video\multfilm\Ну, погоди!\Ну, погоди! 2.avi33492992 5/5TTH:B4O5M74UPKZ7I23CH36NA3SZOUZTJLWNVEIJMTQ (dc.a-galaxy.com:411)| Во флае фунция ClientManager::findHub(hubIpPort) ломается TODO разобраться в чем баг Если верить этому http://mydc.ru/topic915s20.html?p=6845#entry6845 то там может быть только IP _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1443_
priority
в sr передается не ip а доменное имя from on march пример sr chen video multfilm ну погоди ну погоди dc a galaxy com во флае фунция clientmanager findhub hubipport ломается todo разобраться в чем баг если верить этому то там может быть только ip original issue
1
153,349
13,503,610,093
IssuesEvent
2020-09-13 14:20:12
vignette-project/vignette
https://api.github.com/repos/vignette-project/vignette
opened
Lag Compensation for Prediction Data to Live2D
documentation enhancement priority:high
As part of #28, we have discussed how raw data would result on jittery rough data, even if the neural network used is theoretically as precise as a human eye predicting the facial movements of the subject. To compensate for jittery input, we will implement a sort of lag-compensation algorithm. ## Background John Carmack's work with Latency Mitigation for Virtual Reality Devices ([source](https://danluu.com/latency-mitigation/)) explains that the physical movement from the user's head up to the eyes is critical to the experience. While the document is designed mainly for virtual reality, one can argue the methodologies used to provide a seamless experience for virtual reality can be applied for a face tracking application, as face tracking like HMDs, are also very demanding "human-in-the-loop" interfaces. [Byeong Doo-Choi, et al.](https://ieeexplore.ieee.org/abstract/document/4162523/?casa_token=P6kkAYIaHNEAAAAA:0-vUZ2KPz5CG9KnUW6Cm8_nppf9cPIgoleX8zvZVCPAd3sFCeHP4JHMDrWqaOMSwKfNLoO_hedayUg) work with frame interpolation using a novel algorithm for motion prediction would enhance a target video's temporal resolution, by using Adaptive OBMC. ## Strategy As stated on the background, there are many strategies we can perform lag compensation for such raw jittery input from prediction data from the neural network, it is limited to these two strategies: ### Frame Interpolation by Motion Prediction Byeong Doo-Choi, et al. achieves frame interpolation by the following: > First, we propose the bilateral motion estimation scheme to obtain the motion field of an interpolated frame without yielding the hole and overlapping problems. Then, we partition a frame into several object regions by clustering motion vectors. We apply the variable-size block MC (VS-BMC) algorithm to object boundaries in order to reconstruct edge information with a higher quality. Finally, we use the adaptive overlapped block MC (OBMC), which adjusts the coefficients of overlapped windows based on the reliabilities of neighboring motion vectors. The adaptive OBMC (AOBMC) can overcome the limitations of the conventional OBMC, such as over-smoothing and poor de-blocking According to their experiments, such method would produce better image quality for the interpolated frames, which is helpful for prediction in our neural network, however it comes with a cost of having to process the video at runtime, as the experiment is only done on pre-rendered video frames already. ### View Bypass/Time Warping John Carmack's work with reducing input latency for VR HMDs suggests a multitude of methods, one of them is View Bypass - a method achieved by getting a newer sampling of the input. To achieve this, the input should be sampled once but can be used by both the simulation and the rendering task, thus reducing the latency for such. However, the input and the game thread must run in parallel and the programmer must be careful not to reference the game state otherwise it would cause a race condition. Another method mentioned by Carmack is Time Warping, which he states that: > After drawing a frame with the best information at your disposal, possibly with bypassed view parameters, instead of displaying it directly, fetch the latest user input, generate updated view parameters, and calculate a transformation that warps the rendered image into a position that approximates where it would be with the updated parameters. Using that transform, warp the rendered image into an updated form on screen that reflects the new input. If there are two dimensional overlays present on the screen that need to remain fixed, they must be drawn or composited in after the warp operation, to prevent them from incorrectly moving as the view parameters change. There are different methods of warping which is forward warping and reverse warping, and such warping methods can be used along with View Bypassing. However, the increased complexity for lag compensation of doing input with the main loop concurrently is possible as the input loop will be independent of the game state entirely. ## Conclusion
1.0
Lag Compensation for Prediction Data to Live2D - As part of #28, we have discussed how raw data would result on jittery rough data, even if the neural network used is theoretically as precise as a human eye predicting the facial movements of the subject. To compensate for jittery input, we will implement a sort of lag-compensation algorithm. ## Background John Carmack's work with Latency Mitigation for Virtual Reality Devices ([source](https://danluu.com/latency-mitigation/)) explains that the physical movement from the user's head up to the eyes is critical to the experience. While the document is designed mainly for virtual reality, one can argue the methodologies used to provide a seamless experience for virtual reality can be applied for a face tracking application, as face tracking like HMDs, are also very demanding "human-in-the-loop" interfaces. [Byeong Doo-Choi, et al.](https://ieeexplore.ieee.org/abstract/document/4162523/?casa_token=P6kkAYIaHNEAAAAA:0-vUZ2KPz5CG9KnUW6Cm8_nppf9cPIgoleX8zvZVCPAd3sFCeHP4JHMDrWqaOMSwKfNLoO_hedayUg) work with frame interpolation using a novel algorithm for motion prediction would enhance a target video's temporal resolution, by using Adaptive OBMC. ## Strategy As stated on the background, there are many strategies we can perform lag compensation for such raw jittery input from prediction data from the neural network, it is limited to these two strategies: ### Frame Interpolation by Motion Prediction Byeong Doo-Choi, et al. achieves frame interpolation by the following: > First, we propose the bilateral motion estimation scheme to obtain the motion field of an interpolated frame without yielding the hole and overlapping problems. Then, we partition a frame into several object regions by clustering motion vectors. We apply the variable-size block MC (VS-BMC) algorithm to object boundaries in order to reconstruct edge information with a higher quality. Finally, we use the adaptive overlapped block MC (OBMC), which adjusts the coefficients of overlapped windows based on the reliabilities of neighboring motion vectors. The adaptive OBMC (AOBMC) can overcome the limitations of the conventional OBMC, such as over-smoothing and poor de-blocking According to their experiments, such method would produce better image quality for the interpolated frames, which is helpful for prediction in our neural network, however it comes with a cost of having to process the video at runtime, as the experiment is only done on pre-rendered video frames already. ### View Bypass/Time Warping John Carmack's work with reducing input latency for VR HMDs suggests a multitude of methods, one of them is View Bypass - a method achieved by getting a newer sampling of the input. To achieve this, the input should be sampled once but can be used by both the simulation and the rendering task, thus reducing the latency for such. However, the input and the game thread must run in parallel and the programmer must be careful not to reference the game state otherwise it would cause a race condition. Another method mentioned by Carmack is Time Warping, which he states that: > After drawing a frame with the best information at your disposal, possibly with bypassed view parameters, instead of displaying it directly, fetch the latest user input, generate updated view parameters, and calculate a transformation that warps the rendered image into a position that approximates where it would be with the updated parameters. Using that transform, warp the rendered image into an updated form on screen that reflects the new input. If there are two dimensional overlays present on the screen that need to remain fixed, they must be drawn or composited in after the warp operation, to prevent them from incorrectly moving as the view parameters change. There are different methods of warping which is forward warping and reverse warping, and such warping methods can be used along with View Bypassing. However, the increased complexity for lag compensation of doing input with the main loop concurrently is possible as the input loop will be independent of the game state entirely. ## Conclusion
non_priority
lag compensation for prediction data to as part of we have discussed how raw data would result on jittery rough data even if the neural network used is theoretically as precise as a human eye predicting the facial movements of the subject to compensate for jittery input we will implement a sort of lag compensation algorithm background john carmack s work with latency mitigation for virtual reality devices explains that the physical movement from the user s head up to the eyes is critical to the experience while the document is designed mainly for virtual reality one can argue the methodologies used to provide a seamless experience for virtual reality can be applied for a face tracking application as face tracking like hmds are also very demanding human in the loop interfaces work with frame interpolation using a novel algorithm for motion prediction would enhance a target video s temporal resolution by using adaptive obmc strategy as stated on the background there are many strategies we can perform lag compensation for such raw jittery input from prediction data from the neural network it is limited to these two strategies frame interpolation by motion prediction byeong doo choi et al achieves frame interpolation by the following first we propose the bilateral motion estimation scheme to obtain the motion field of an interpolated frame without yielding the hole and overlapping problems then we partition a frame into several object regions by clustering motion vectors we apply the variable size block mc vs bmc algorithm to object boundaries in order to reconstruct edge information with a higher quality finally we use the adaptive overlapped block mc obmc which adjusts the coefficients of overlapped windows based on the reliabilities of neighboring motion vectors the adaptive obmc aobmc can overcome the limitations of the conventional obmc such as over smoothing and poor de blocking according to their experiments such method would produce better image quality for the interpolated frames which is helpful for prediction in our neural network however it comes with a cost of having to process the video at runtime as the experiment is only done on pre rendered video frames already view bypass time warping john carmack s work with reducing input latency for vr hmds suggests a multitude of methods one of them is view bypass a method achieved by getting a newer sampling of the input to achieve this the input should be sampled once but can be used by both the simulation and the rendering task thus reducing the latency for such however the input and the game thread must run in parallel and the programmer must be careful not to reference the game state otherwise it would cause a race condition another method mentioned by carmack is time warping which he states that after drawing a frame with the best information at your disposal possibly with bypassed view parameters instead of displaying it directly fetch the latest user input generate updated view parameters and calculate a transformation that warps the rendered image into a position that approximates where it would be with the updated parameters using that transform warp the rendered image into an updated form on screen that reflects the new input if there are two dimensional overlays present on the screen that need to remain fixed they must be drawn or composited in after the warp operation to prevent them from incorrectly moving as the view parameters change there are different methods of warping which is forward warping and reverse warping and such warping methods can be used along with view bypassing however the increased complexity for lag compensation of doing input with the main loop concurrently is possible as the input loop will be independent of the game state entirely conclusion
0
292,830
8,969,120,348
IssuesEvent
2019-01-29 09:59:18
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
closed
Instantiate to persistent refactor
Category: API Category: vCenter Priority: High Type: Feature
# Enhancement Request ## Description Ideally covering: - Avoid copying the VM Template image(s) that will be discarded - Remove get_img_name from FileHelper and place it in a more sound location - Compact all VM Reconfigured ## Use case All involving Instantiate to persistent ## Interface Changes None # Progress Status - [x] Branch created - [x] Code committed to development branch - [x] Testing - QA - [x] Documentation - [x] Release notes - resolved issues, compatibility, known issues - [x] Code committed to upstream release/hotfix branches - [x] Documentation committed to upstream release/hotfix branches
1.0
Instantiate to persistent refactor - # Enhancement Request ## Description Ideally covering: - Avoid copying the VM Template image(s) that will be discarded - Remove get_img_name from FileHelper and place it in a more sound location - Compact all VM Reconfigured ## Use case All involving Instantiate to persistent ## Interface Changes None # Progress Status - [x] Branch created - [x] Code committed to development branch - [x] Testing - QA - [x] Documentation - [x] Release notes - resolved issues, compatibility, known issues - [x] Code committed to upstream release/hotfix branches - [x] Documentation committed to upstream release/hotfix branches
priority
instantiate to persistent refactor enhancement request description ideally covering avoid copying the vm template image s that will be discarded remove get img name from filehelper and place it in a more sound location compact all vm reconfigured use case all involving instantiate to persistent interface changes none progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
1
256,636
27,561,701,877
IssuesEvent
2023-03-07 22:40:59
samqws-marketing/amzn-ion-hive-serde
https://api.github.com/repos/samqws-marketing/amzn-ion-hive-serde
closed
CVE-2018-5968 (High) detected in jackson-databind-2.6.5.jar - autoclosed
Mend: dependency security vulnerability
## CVE-2018-5968 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /integration-test/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar</p> <p> Dependency Hierarchy: - hive-serde-2.3.9.jar (Root Library) - hive-common-2.3.9.jar - :x: **jackson-databind-2.6.5.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/amzn-ion-hive-serde/commit/ffb6641ebb10aac58bb7eec412635e91e79fac24">ffb6641ebb10aac58bb7eec412635e91e79fac24</a></p> <p>Found in base branch: <b>0.3.0</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind through 2.8.11 and 2.9.x through 2.9.3 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 and CVE-2017-17485 deserialization flaws. This is exploitable via two different gadgets that bypass a blacklist. <p>Publish Date: 2018-01-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-5968>CVE-2018-5968</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968">http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968</a></p> <p>Release Date: 2018-01-22</p> <p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.6.7.3</p> <p>Direct dependency fix Resolution (org.apache.hive:hive-serde): 3.0.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
True
CVE-2018-5968 (High) detected in jackson-databind-2.6.5.jar - autoclosed - ## CVE-2018-5968 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /integration-test/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.5/d50be1723a09befd903887099ff2014ea9020333/jackson-databind-2.6.5.jar</p> <p> Dependency Hierarchy: - hive-serde-2.3.9.jar (Root Library) - hive-common-2.3.9.jar - :x: **jackson-databind-2.6.5.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/amzn-ion-hive-serde/commit/ffb6641ebb10aac58bb7eec412635e91e79fac24">ffb6641ebb10aac58bb7eec412635e91e79fac24</a></p> <p>Found in base branch: <b>0.3.0</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind through 2.8.11 and 2.9.x through 2.9.3 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 and CVE-2017-17485 deserialization flaws. This is exploitable via two different gadgets that bypass a blacklist. <p>Publish Date: 2018-01-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-5968>CVE-2018-5968</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968">http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968</a></p> <p>Release Date: 2018-01-22</p> <p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.6.7.3</p> <p>Direct dependency fix Resolution (org.apache.hive:hive-serde): 3.0.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
non_priority
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file integration test build gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy hive serde jar root library hive common jar x jackson databind jar vulnerable library found in head commit a href found in base branch vulnerability details fasterxml jackson databind through and x through allows unauthenticated remote code execution because of an incomplete fix for the cve and cve deserialization flaws this is exploitable via two different gadgets that bypass a blacklist publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind direct dependency fix resolution org apache hive hive serde check this box to open an automated fix pr
0
103,328
4,167,177,742
IssuesEvent
2016-06-20 08:29:45
Kunstmaan/KunstmaanBundlesCMS
https://api.github.com/repos/Kunstmaan/KunstmaanBundlesCMS
closed
Upgrade to Behat 3.0.*
Priority: Low Profile: Backend Target audience: Developers Type: Feature
Last April [Behat 3.0](http://behat.org) was released, we are currently locked at v2. We should upgrade.
1.0
Upgrade to Behat 3.0.* - Last April [Behat 3.0](http://behat.org) was released, we are currently locked at v2. We should upgrade.
priority
upgrade to behat last april was released we are currently locked at we should upgrade
1
767,793
26,940,755,616
IssuesEvent
2023-02-08 01:52:07
wso2/api-manager
https://api.github.com/repos/wso2/api-manager
closed
User should not be able to deploy a retired API
Type/Bug Priority/Normal Component/APIM 4.2.0-beta
### Description In APIM after we retire an API the deploy button still active and by clicking user can deploy the API to the gateway. ### Steps to Reproduce 1. Create and publish an API in the publisher 2. Try to invoke the API and note down the access token. 3. Move the API to 'Deprecated' and then to 'Retired' 4. Now go to 'Overview' page, you will see that the deploy button is still active. Click on it and deploy it on the gateway. 5. Now invoke the api using curl or postman you will get the success response. Expected behavior After changing an API state to 'Retire', user should not be able to deploy the API. ### Affected Component Analytics ### Version 4.2.0 ### Environment Details (with versions) _No response_ ### Relevant Log Output _No response_ ### Related Issues _No response_ ### Suggested Labels bug
1.0
User should not be able to deploy a retired API - ### Description In APIM after we retire an API the deploy button still active and by clicking user can deploy the API to the gateway. ### Steps to Reproduce 1. Create and publish an API in the publisher 2. Try to invoke the API and note down the access token. 3. Move the API to 'Deprecated' and then to 'Retired' 4. Now go to 'Overview' page, you will see that the deploy button is still active. Click on it and deploy it on the gateway. 5. Now invoke the api using curl or postman you will get the success response. Expected behavior After changing an API state to 'Retire', user should not be able to deploy the API. ### Affected Component Analytics ### Version 4.2.0 ### Environment Details (with versions) _No response_ ### Relevant Log Output _No response_ ### Related Issues _No response_ ### Suggested Labels bug
priority
user should not be able to deploy a retired api description in apim after we retire an api the deploy button still active and by clicking user can deploy the api to the gateway steps to reproduce create and publish an api in the publisher try to invoke the api and note down the access token move the api to deprecated and then to retired now go to overview page you will see that the deploy button is still active click on it and deploy it on the gateway now invoke the api using curl or postman you will get the success response expected behavior after changing an api state to retire user should not be able to deploy the api affected component analytics version environment details with versions no response relevant log output no response related issues no response suggested labels bug
1
308,857
9,458,439,678
IssuesEvent
2019-04-17 05:17:21
GameFreak527/Comp308_W2019_AwesomeSurveys_Group04
https://api.github.com/repos/GameFreak527/Comp308_W2019_AwesomeSurveys_Group04
closed
Registration Page
Dependency High Priority good first issue
A form will allow the user to enter profile information (username, password, email address, etc.), which will be stored in a MongoDB database structure (2 Marks: GUI, 2 Marks: Functionality).
1.0
Registration Page - A form will allow the user to enter profile information (username, password, email address, etc.), which will be stored in a MongoDB database structure (2 Marks: GUI, 2 Marks: Functionality).
priority
registration page a form will allow the user to enter profile information username password email address etc which will be stored in a mongodb database structure marks gui marks functionality
1
182,552
21,672,917,809
IssuesEvent
2022-05-08 08:41:02
turkdevops/core
https://api.github.com/repos/turkdevops/core
closed
CVE-2021-28092 (High) detected in is-svg-2.1.0.tgz - autoclosed
security vulnerability
## CVE-2021-28092 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-2.1.0.tgz</b></p></summary> <p>Check if a string or buffer is SVG</p> <p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-2.1.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-2.1.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/is-svg</p> <p> Dependency Hierarchy: - sdk-1.5.1.tgz (Root Library) - microbundle-0.3.1.tgz - rollup-plugin-postcss-1.2.9.tgz - cssnano-3.10.0.tgz - postcss-svgo-2.1.6.tgz - :x: **is-svg-2.1.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The is-svg package 2.1.0 through 4.2.1 for Node.js uses a regular expression that is vulnerable to Regular Expression Denial of Service (ReDoS). If an attacker provides a malicious string, is-svg will get stuck processing the input for a very long time. <p>Publish Date: 2021-03-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28092>CVE-2021-28092</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092</a></p> <p>Release Date: 2021-03-12</p> <p>Fix Resolution (is-svg): 4.2.2</p> <p>Direct dependency fix Resolution (@stackblitz/sdk): 1.5.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-28092 (High) detected in is-svg-2.1.0.tgz - autoclosed - ## CVE-2021-28092 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-2.1.0.tgz</b></p></summary> <p>Check if a string or buffer is SVG</p> <p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-2.1.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-2.1.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/is-svg</p> <p> Dependency Hierarchy: - sdk-1.5.1.tgz (Root Library) - microbundle-0.3.1.tgz - rollup-plugin-postcss-1.2.9.tgz - cssnano-3.10.0.tgz - postcss-svgo-2.1.6.tgz - :x: **is-svg-2.1.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The is-svg package 2.1.0 through 4.2.1 for Node.js uses a regular expression that is vulnerable to Regular Expression Denial of Service (ReDoS). If an attacker provides a malicious string, is-svg will get stuck processing the input for a very long time. <p>Publish Date: 2021-03-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28092>CVE-2021-28092</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092</a></p> <p>Release Date: 2021-03-12</p> <p>Fix Resolution (is-svg): 4.2.2</p> <p>Direct dependency fix Resolution (@stackblitz/sdk): 1.5.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in is svg tgz autoclosed cve high severity vulnerability vulnerable library is svg tgz check if a string or buffer is svg library home page a href path to dependency file package json path to vulnerable library node modules is svg dependency hierarchy sdk tgz root library microbundle tgz rollup plugin postcss tgz cssnano tgz postcss svgo tgz x is svg tgz vulnerable library found in base branch master vulnerability details the is svg package through for node js uses a regular expression that is vulnerable to regular expression denial of service redos if an attacker provides a malicious string is svg will get stuck processing the input for a very long time publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution is svg direct dependency fix resolution stackblitz sdk step up your open source security game with whitesource
0
537,682
15,733,150,361
IssuesEvent
2021-03-29 19:11:48
bcgov/entity
https://api.github.com/repos/bcgov/entity
closed
NR 2661646 INDUSTRIAL PERSONNEL AND MANAGEMENT SERVICES INC.
ENTITY OPS Priority1
#### ServiceNow incident: INC0082901 #### Contact information Staff Name: Myra Noviello Staff Email: Myra.Noviello@gov.bc.ca #### Description She’s trying to file a Statement of Extraprovincial Registration in Corporate Online and unable to retrieve the NR information. Business team needs to provide her number in her home jurisdiction. It would be on the certificate from home jurisdiction she would have sent to you guys to remove the condition on her name. Work around: - [x] Need to create new NR in NRO. - [x] Approve New NR in Namex without conditions and add comment stating consent received under NR 2661646. - [x] Cancel old NR in NameX - [x] Test filing in Corporate Online - [x] Email Myra back with new NR# and name results letter #### Tasks - [x] When ticket has been created, post the ticket in RocketChat '#Operations Tasks' channel - [x] Add **entity** or **relationships** label to zenhub ticket - [x] Add 'Priority1' label to zenhub ticket - [x] Assign zenhub ticket to milestone: current, and place in pipeline: sprint backlog - [x] Reply All to IT Ops email and provide zenhub ticket number opened and which team it was assigned to - [x] BAs to complete work & close zenhub ticket (see above) - [x] Author of zenhub ticket to mark ServiceNow ticket as resolved or ask IT Ops to do so
1.0
NR 2661646 INDUSTRIAL PERSONNEL AND MANAGEMENT SERVICES INC. - #### ServiceNow incident: INC0082901 #### Contact information Staff Name: Myra Noviello Staff Email: Myra.Noviello@gov.bc.ca #### Description She’s trying to file a Statement of Extraprovincial Registration in Corporate Online and unable to retrieve the NR information. Business team needs to provide her number in her home jurisdiction. It would be on the certificate from home jurisdiction she would have sent to you guys to remove the condition on her name. Work around: - [x] Need to create new NR in NRO. - [x] Approve New NR in Namex without conditions and add comment stating consent received under NR 2661646. - [x] Cancel old NR in NameX - [x] Test filing in Corporate Online - [x] Email Myra back with new NR# and name results letter #### Tasks - [x] When ticket has been created, post the ticket in RocketChat '#Operations Tasks' channel - [x] Add **entity** or **relationships** label to zenhub ticket - [x] Add 'Priority1' label to zenhub ticket - [x] Assign zenhub ticket to milestone: current, and place in pipeline: sprint backlog - [x] Reply All to IT Ops email and provide zenhub ticket number opened and which team it was assigned to - [x] BAs to complete work & close zenhub ticket (see above) - [x] Author of zenhub ticket to mark ServiceNow ticket as resolved or ask IT Ops to do so
priority
nr industrial personnel and management services inc servicenow incident contact information staff name myra noviello staff email myra noviello gov bc ca description she’s trying to file a statement of extraprovincial registration in corporate online and unable to retrieve the nr information business team needs to provide her number in her home jurisdiction it would be on the certificate from home jurisdiction she would have sent to you guys to remove the condition on her name work around need to create new nr in nro approve new nr in namex without conditions and add comment stating consent received under nr cancel old nr in namex test filing in corporate online email myra back with new nr and name results letter tasks when ticket has been created post the ticket in rocketchat operations tasks channel add entity or relationships label to zenhub ticket add label to zenhub ticket assign zenhub ticket to milestone current and place in pipeline sprint backlog reply all to it ops email and provide zenhub ticket number opened and which team it was assigned to bas to complete work close zenhub ticket see above author of zenhub ticket to mark servicenow ticket as resolved or ask it ops to do so
1
324,842
24,019,581,817
IssuesEvent
2022-09-15 06:18:22
chkware/vscode-ext
https://api.github.com/repos/chkware/vscode-ext
closed
List of testcases - v0.1.0
documentation
What - We need a list of testcases to check if all the features for extensions are working properly Why - We haven't wrote any test code for this extension (soon to come) for the sake of make it on discovery phase
1.0
List of testcases - v0.1.0 - What - We need a list of testcases to check if all the features for extensions are working properly Why - We haven't wrote any test code for this extension (soon to come) for the sake of make it on discovery phase
non_priority
list of testcases what we need a list of testcases to check if all the features for extensions are working properly why we haven t wrote any test code for this extension soon to come for the sake of make it on discovery phase
0
443,032
12,758,325,511
IssuesEvent
2020-06-29 01:51:17
MyDataTaiwan/mylog14
https://api.github.com/repos/MyDataTaiwan/mylog14
closed
[Bug Report] Language to French can't record body temperature
QA fix verified priority-critical second-week
**Description** Language to French can't record body temperature **Steps to Reproduce** 1. change Language to French 2. open the app 3. Click on '+' 3. Input body temperature 4. Click on save ( no response ) ![Screenshot_20200605-110658](https://user-images.githubusercontent.com/62582260/83835038-58dfa580-a722-11ea-8605-535c9620b6f5.png) **Environment** * MyLog: V0.8.3 * OS:HTC EXODUS 1s,Android 8.1.0
1.0
[Bug Report] Language to French can't record body temperature - **Description** Language to French can't record body temperature **Steps to Reproduce** 1. change Language to French 2. open the app 3. Click on '+' 3. Input body temperature 4. Click on save ( no response ) ![Screenshot_20200605-110658](https://user-images.githubusercontent.com/62582260/83835038-58dfa580-a722-11ea-8605-535c9620b6f5.png) **Environment** * MyLog: V0.8.3 * OS:HTC EXODUS 1s,Android 8.1.0
priority
language to french can t record body temperature description language to french can t record body temperature steps to reproduce change language to french open the app click on input body temperature click on save no response environment mylog os htc exodus android
1
743,310
25,894,392,624
IssuesEvent
2022-12-14 20:55:58
GoogleCloudPlatform/alloydb-auth-proxy
https://api.github.com/repos/GoogleCloudPlatform/alloydb-auth-proxy
closed
Fail fast on invalid instance URI
type: feature request priority: p0
When a user provides an invalid instance URI, the Proxy does not immediately fail. Instead it fails when a user attempts to connect to the local listener. Instead, the Proxy should validate instance URIs as part of startup.
1.0
Fail fast on invalid instance URI - When a user provides an invalid instance URI, the Proxy does not immediately fail. Instead it fails when a user attempts to connect to the local listener. Instead, the Proxy should validate instance URIs as part of startup.
priority
fail fast on invalid instance uri when a user provides an invalid instance uri the proxy does not immediately fail instead it fails when a user attempts to connect to the local listener instead the proxy should validate instance uris as part of startup
1
162,165
6,148,526,487
IssuesEvent
2017-06-27 18:01:54
openshift/origin
https://api.github.com/repos/openshift/origin
closed
oc cluster up fails on Gentoo
component/cli priority/P2
When trying to run oc cluster up on Gentoo it fails with the following problems: ``` oc cluster up -- Checking OpenShift client ... OK -- Checking Docker client ... OK -- Checking Docker version ... OK -- Checking for existing OpenShift container ... OK -- Checking for openshift/origin:v1.3.0 image ... OK -- Checking Docker daemon configuration ... OK -- Checking for available ports ... WARNING: Binding DNS on port 8053 instead of 53, which may be not be resolvable from all clients. -- Checking type of volume mount ... Using nsenter mounter for OpenShift volumes -- Creating host directories ... OK -- Finding server IP ... Using 10.6.7.126 as the server IP -- Starting OpenShift container ... Creating initial OpenShift configuration FAIL Error: could not create OpenShift configuration Caused By: Error: cannot start container 5346a66289ab4c20f70c4c3829fe7a7399c678ab8ecc7d38abb0eb0e9f9c180c Caused By: Error: API error (500): {"message":"linux mounts: Path /var/lib/origin/openshift.local.volumes is mounted on / but it is not a shared or slave mount."} ``` When moving /bin/findmnt away like mentioned in #10860 it fails like this: ``` oc cluster up -- Checking OpenShift client ... OK -- Checking Docker client ... OK -- Checking Docker version ... OK -- Checking for existing OpenShift container ... Deleted existing OpenShift container -- Checking for openshift/origin:v1.3.0 image ... OK -- Checking Docker daemon configuration ... OK -- Checking for available ports ... WARNING: Binding DNS on port 8053 instead of 53, which may be not be resolvable from all clients. -- Checking type of volume mount ... Using Docker shared volumes for OpenShift volumes -- Creating host directories ... OK -- Finding server IP ... Using 10.6.7.126 as the server IP -- Starting OpenShift container ... Creating initial OpenShift configuration Starting OpenShift using container 'origin' Waiting for API server to start listening OpenShift server started -- Installing registry ... error: Post https://10.6.7.126:8443/oapi/v1/clusterrolebindings: read tcp 10.6.7.126:33148->10.6.7.126:8443: read: connection reset by peer error: Post https://10.6.7.126:8443/oapi/v1/namespaces/default/deploymentconfigs: dial tcp 10.6.7.126:8443: getsockopt: connection refused error: Post https://10.6.7.126:8443/api/v1/namespaces/default/services: dial tcp 10.6.7.126:8443: getsockopt: connection refused FAIL Error: cannot install registry Details: Last 10 lines of "origin" container log: I1011 10:09:53.110122 29642 manager.go:138] cAdvisor running in container: "/docker/3a7d694745de2f358641460cb9e6d37f04b10bc8e33de88e8e5dee1403d21118" I1011 10:09:53.118359 29642 node.go:380] Using iptables Proxier. I1011 10:09:53.118413 29642 proxier.go:209] missing br-netfilter module or unset br-nf-call-iptables; proxy may not work as intended I1011 10:09:53.118424 29642 node.go:396] Tearing down userspace rules. I1011 10:09:53.132295 29642 create_dockercfg_secrets.go:119] Dockercfg secret controller initialized, starting. W1011 10:09:53.150071 29642 manager.go:146] unable to connect to Rkt api service: rkt: cannot tcp Dial rkt api service: dial tcp [::1]:15441: getsockopt: connection refused I1011 10:09:53.175331 29642 fs.go:116] Filesystem partitions: map[/dev/mapper/sdb2_crypt:{mountpoint:/var/lib/docker major:252 minor:2 fsType:ext4 blockSize:0} /dev/mapper/luks-fed3e982-8233-4159-91b5-983294831e68:{mountpoint:/var/lib/origin/openshift.local.volumes major:252 minor:0 fsType:ext4 blockSize:0}] I1011 10:09:53.223591 29642 manager.go:192] Machine: {NumCores:8 CpuFrequency:2601000 MemoryCapacity:16684916736 MachineID: SystemUUID:4C4C4544-004D-5610-8052-C4C04F423632 BootID:34ed0ca6-7a35-492c-a6f3-e61c21e9226f Filesystems:[{Device:overlay Capacity:237250830336 Type:vfs Inodes:14721024} {Device:/dev/mapper/sdb2_crypt Capacity:237250830336 Type:vfs Inodes:14721024} {Device:/dev/mapper/luks-fed3e982-8233-4159-91b5-983294831e68 Capacity:125360812032 Type:vfs Inodes:7782400}] DiskMap:map[252:2:{Name:dm-2 Major:252 Minor:2 Size:241169334272 Scheduler:none} 8:0:{Name:sda Major:8 Minor:0 Size:128035676160 Scheduler:noop} 8:16:{Name:sdb Major:8 Minor:16 Size:1000204886016 Scheduler:cfq} 252:0:{Name:dm-0 Major:252 Minor:0 Size:127495659520 Scheduler:none} 252:1:{Name:dm-1 Major:252 Minor:1 Size:759029563392 Scheduler:none}] NetworkDevices:[{Name:bond0 MacAddress:f6:ad:87:03:0c:e6 Speed:0 Mtu:1500} {Name:eth0 MacAddress:20:47:47:72:4f:34 Speed:10 Mtu:1500} {Name:sit0 MacAddress:00:00:00:00 Speed:0 Mtu:1480} {Name:usb0 MacAddress:9c:eb:e8:27:fc:ed Speed:0 Mtu:1500} {Name:virbr0 MacAddress:52:54:00:19:67:28 Speed:0 Mtu:1500} {Name:virbr0-nic MacAddress:52:54:00:19:67:28 Speed:0 Mtu:1500} {Name:wlan0 MacAddress:4c:34:88:ec:da:51 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:16684916736 Cores:[{Id:0 Threads:[0 4] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:262144 Type:Unified Level:2}]} {Id:1 Threads:[1 5] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:262144 Type:Unified Level:2}]} {Id:2 Threads:[2 6] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:262144 Type:Unified Level:2}]} {Id:3 Threads:[3 7] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:262144 Type:Unified Level:2}]}] Caches:[{Size:6291456 Type:Unified Level:3}]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} F1011 10:09:53.240007 29642 node.go:443] error: Could not initialize Kubernetes Proxy. You must run this process as root to use the service proxy: failed to initialize iptables: error appending rule: exit status 1: iptables: No chain/target/match by that name. I1011 10:09:53.295361 29642 manager.go:198] Version: {KernelVersion:4.7.2-gentoo ContainerOsVersion:CentOS Linux 7 (Core) DockerVersion:1.12.1 CadvisorVersion: CadvisorRevision:} Caused By: Error: exit directly ``` The error with br-netfilter is discussed in kubernetes-23385, and is not a real error. ``` grep -i BRIDGE_NETFILTER .config CONFIG_BRIDGE_NETFILTER=y ``` ##### Version ``` ./oc version oc v1.3.0 kubernetes v1.3.0+52492b4 features: Basic-Auth GSSAPI Kerberos SPNEGO Server https://<someenterprise>:8443 kubernetes v1.2.0-36-g4a3f9c5 ``` util-linux-2.26.2 (contains findmnt)
1.0
oc cluster up fails on Gentoo - When trying to run oc cluster up on Gentoo it fails with the following problems: ``` oc cluster up -- Checking OpenShift client ... OK -- Checking Docker client ... OK -- Checking Docker version ... OK -- Checking for existing OpenShift container ... OK -- Checking for openshift/origin:v1.3.0 image ... OK -- Checking Docker daemon configuration ... OK -- Checking for available ports ... WARNING: Binding DNS on port 8053 instead of 53, which may be not be resolvable from all clients. -- Checking type of volume mount ... Using nsenter mounter for OpenShift volumes -- Creating host directories ... OK -- Finding server IP ... Using 10.6.7.126 as the server IP -- Starting OpenShift container ... Creating initial OpenShift configuration FAIL Error: could not create OpenShift configuration Caused By: Error: cannot start container 5346a66289ab4c20f70c4c3829fe7a7399c678ab8ecc7d38abb0eb0e9f9c180c Caused By: Error: API error (500): {"message":"linux mounts: Path /var/lib/origin/openshift.local.volumes is mounted on / but it is not a shared or slave mount."} ``` When moving /bin/findmnt away like mentioned in #10860 it fails like this: ``` oc cluster up -- Checking OpenShift client ... OK -- Checking Docker client ... OK -- Checking Docker version ... OK -- Checking for existing OpenShift container ... Deleted existing OpenShift container -- Checking for openshift/origin:v1.3.0 image ... OK -- Checking Docker daemon configuration ... OK -- Checking for available ports ... WARNING: Binding DNS on port 8053 instead of 53, which may be not be resolvable from all clients. -- Checking type of volume mount ... Using Docker shared volumes for OpenShift volumes -- Creating host directories ... OK -- Finding server IP ... Using 10.6.7.126 as the server IP -- Starting OpenShift container ... Creating initial OpenShift configuration Starting OpenShift using container 'origin' Waiting for API server to start listening OpenShift server started -- Installing registry ... error: Post https://10.6.7.126:8443/oapi/v1/clusterrolebindings: read tcp 10.6.7.126:33148->10.6.7.126:8443: read: connection reset by peer error: Post https://10.6.7.126:8443/oapi/v1/namespaces/default/deploymentconfigs: dial tcp 10.6.7.126:8443: getsockopt: connection refused error: Post https://10.6.7.126:8443/api/v1/namespaces/default/services: dial tcp 10.6.7.126:8443: getsockopt: connection refused FAIL Error: cannot install registry Details: Last 10 lines of "origin" container log: I1011 10:09:53.110122 29642 manager.go:138] cAdvisor running in container: "/docker/3a7d694745de2f358641460cb9e6d37f04b10bc8e33de88e8e5dee1403d21118" I1011 10:09:53.118359 29642 node.go:380] Using iptables Proxier. I1011 10:09:53.118413 29642 proxier.go:209] missing br-netfilter module or unset br-nf-call-iptables; proxy may not work as intended I1011 10:09:53.118424 29642 node.go:396] Tearing down userspace rules. I1011 10:09:53.132295 29642 create_dockercfg_secrets.go:119] Dockercfg secret controller initialized, starting. W1011 10:09:53.150071 29642 manager.go:146] unable to connect to Rkt api service: rkt: cannot tcp Dial rkt api service: dial tcp [::1]:15441: getsockopt: connection refused I1011 10:09:53.175331 29642 fs.go:116] Filesystem partitions: map[/dev/mapper/sdb2_crypt:{mountpoint:/var/lib/docker major:252 minor:2 fsType:ext4 blockSize:0} /dev/mapper/luks-fed3e982-8233-4159-91b5-983294831e68:{mountpoint:/var/lib/origin/openshift.local.volumes major:252 minor:0 fsType:ext4 blockSize:0}] I1011 10:09:53.223591 29642 manager.go:192] Machine: {NumCores:8 CpuFrequency:2601000 MemoryCapacity:16684916736 MachineID: SystemUUID:4C4C4544-004D-5610-8052-C4C04F423632 BootID:34ed0ca6-7a35-492c-a6f3-e61c21e9226f Filesystems:[{Device:overlay Capacity:237250830336 Type:vfs Inodes:14721024} {Device:/dev/mapper/sdb2_crypt Capacity:237250830336 Type:vfs Inodes:14721024} {Device:/dev/mapper/luks-fed3e982-8233-4159-91b5-983294831e68 Capacity:125360812032 Type:vfs Inodes:7782400}] DiskMap:map[252:2:{Name:dm-2 Major:252 Minor:2 Size:241169334272 Scheduler:none} 8:0:{Name:sda Major:8 Minor:0 Size:128035676160 Scheduler:noop} 8:16:{Name:sdb Major:8 Minor:16 Size:1000204886016 Scheduler:cfq} 252:0:{Name:dm-0 Major:252 Minor:0 Size:127495659520 Scheduler:none} 252:1:{Name:dm-1 Major:252 Minor:1 Size:759029563392 Scheduler:none}] NetworkDevices:[{Name:bond0 MacAddress:f6:ad:87:03:0c:e6 Speed:0 Mtu:1500} {Name:eth0 MacAddress:20:47:47:72:4f:34 Speed:10 Mtu:1500} {Name:sit0 MacAddress:00:00:00:00 Speed:0 Mtu:1480} {Name:usb0 MacAddress:9c:eb:e8:27:fc:ed Speed:0 Mtu:1500} {Name:virbr0 MacAddress:52:54:00:19:67:28 Speed:0 Mtu:1500} {Name:virbr0-nic MacAddress:52:54:00:19:67:28 Speed:0 Mtu:1500} {Name:wlan0 MacAddress:4c:34:88:ec:da:51 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:16684916736 Cores:[{Id:0 Threads:[0 4] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:262144 Type:Unified Level:2}]} {Id:1 Threads:[1 5] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:262144 Type:Unified Level:2}]} {Id:2 Threads:[2 6] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:262144 Type:Unified Level:2}]} {Id:3 Threads:[3 7] Caches:[{Size:32768 Type:Data Level:1} {Size:32768 Type:Instruction Level:1} {Size:262144 Type:Unified Level:2}]}] Caches:[{Size:6291456 Type:Unified Level:3}]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} F1011 10:09:53.240007 29642 node.go:443] error: Could not initialize Kubernetes Proxy. You must run this process as root to use the service proxy: failed to initialize iptables: error appending rule: exit status 1: iptables: No chain/target/match by that name. I1011 10:09:53.295361 29642 manager.go:198] Version: {KernelVersion:4.7.2-gentoo ContainerOsVersion:CentOS Linux 7 (Core) DockerVersion:1.12.1 CadvisorVersion: CadvisorRevision:} Caused By: Error: exit directly ``` The error with br-netfilter is discussed in kubernetes-23385, and is not a real error. ``` grep -i BRIDGE_NETFILTER .config CONFIG_BRIDGE_NETFILTER=y ``` ##### Version ``` ./oc version oc v1.3.0 kubernetes v1.3.0+52492b4 features: Basic-Auth GSSAPI Kerberos SPNEGO Server https://<someenterprise>:8443 kubernetes v1.2.0-36-g4a3f9c5 ``` util-linux-2.26.2 (contains findmnt)
priority
oc cluster up fails on gentoo when trying to run oc cluster up on gentoo it fails with the following problems oc cluster up checking openshift client ok checking docker client ok checking docker version ok checking for existing openshift container ok checking for openshift origin image ok checking docker daemon configuration ok checking for available ports warning binding dns on port instead of which may be not be resolvable from all clients checking type of volume mount using nsenter mounter for openshift volumes creating host directories ok finding server ip using as the server ip starting openshift container creating initial openshift configuration fail error could not create openshift configuration caused by error cannot start container caused by error api error message linux mounts path var lib origin openshift local volumes is mounted on but it is not a shared or slave mount when moving bin findmnt away like mentioned in it fails like this oc cluster up checking openshift client ok checking docker client ok checking docker version ok checking for existing openshift container deleted existing openshift container checking for openshift origin image ok checking docker daemon configuration ok checking for available ports warning binding dns on port instead of which may be not be resolvable from all clients checking type of volume mount using docker shared volumes for openshift volumes creating host directories ok finding server ip using as the server ip starting openshift container creating initial openshift configuration starting openshift using container origin waiting for api server to start listening openshift server started installing registry error post read tcp read connection reset by peer error post dial tcp getsockopt connection refused error post dial tcp getsockopt connection refused fail error cannot install registry details last lines of origin container log manager go cadvisor running in container docker node go using iptables proxier proxier go missing br netfilter module or unset br nf call iptables proxy may not work as intended node go tearing down userspace rules create dockercfg secrets go dockercfg secret controller initialized starting manager go unable to connect to rkt api service rkt cannot tcp dial rkt api service dial tcp getsockopt connection refused fs go filesystem partitions map manager go machine numcores cpufrequency memorycapacity machineid systemuuid bootid filesystems diskmap map networkdevices topology caches id threads caches id threads caches id threads caches caches cloudprovider unknown instancetype unknown instanceid none node go error could not initialize kubernetes proxy you must run this process as root to use the service proxy failed to initialize iptables error appending rule exit status iptables no chain target match by that name manager go version kernelversion gentoo containerosversion centos linux core dockerversion cadvisorversion cadvisorrevision caused by error exit directly the error with br netfilter is discussed in kubernetes and is not a real error grep i bridge netfilter config config bridge netfilter y version oc version oc kubernetes features basic auth gssapi kerberos spnego server kubernetes util linux contains findmnt
1
10,073
26,192,792,118
IssuesEvent
2023-01-03 10:37:42
gitpod-io/gitpod
https://api.github.com/repos/gitpod-io/gitpod
closed
Epic: Proof of Value Reference Architecture
meta: stale type: epic self-hosted: reference-architecture
### Summary Create a specific reference architecture for a PoC phase of the customer life cycle. Aimed at balancing ease-of-set-up with enough scale to prove Gitpod's value in a team setting. ### Context We are working on a reference architecture for prod settings. This reference archtiecture is aimed at creating a scalable, reliable way to operate Gitpod in a production setting. For customers only wanting to try out Gitpod in a PoC phase (i.e. they have already tried the #9075 locally and now want to test with their team), this prod Reference Architecture is too cumbersome to set up - and there is no need to create such a reliable set up in this phase yet. Important: This is not a self-hosted team project, but a Gitpod project. All teams need to give their input, the self-hosted team may just be coordinating. ### Value - Speed up the PoC phase of the sales pipeline - Give the community a easy-to-set-up way of creating a Gitpod installation that should work for a lot of use cases, but should not be used in prod ### Acceptance Criteria - Detailed documentation of the reference architecture, and the reasoning for the infrastructure decisions made. This also includes cost estimates. - A terraform script that creates the reference architecture documented. This script can be used by users at their own discretion (i.e. it is not officially supported), and will be used to run internal automated tests see #7316) ### Measurement - Qualitative feedback on reference artchitecture from users that have set it up - Feedback from Sales Engineering who have accompanied customers in PoC phases that used it - Community feedback ### Issues - [x] https://github.com/gitpod-io/gitpod/issues/11225 - [x] https://github.com/gitpod-io/gitpod/issues/11226 - [x] https://github.com/gitpod-io/gitpod/issues/12001 - [ ] https://github.com/gitpod-io/gitpod/issues/12517
1.0
Epic: Proof of Value Reference Architecture - ### Summary Create a specific reference architecture for a PoC phase of the customer life cycle. Aimed at balancing ease-of-set-up with enough scale to prove Gitpod's value in a team setting. ### Context We are working on a reference architecture for prod settings. This reference archtiecture is aimed at creating a scalable, reliable way to operate Gitpod in a production setting. For customers only wanting to try out Gitpod in a PoC phase (i.e. they have already tried the #9075 locally and now want to test with their team), this prod Reference Architecture is too cumbersome to set up - and there is no need to create such a reliable set up in this phase yet. Important: This is not a self-hosted team project, but a Gitpod project. All teams need to give their input, the self-hosted team may just be coordinating. ### Value - Speed up the PoC phase of the sales pipeline - Give the community a easy-to-set-up way of creating a Gitpod installation that should work for a lot of use cases, but should not be used in prod ### Acceptance Criteria - Detailed documentation of the reference architecture, and the reasoning for the infrastructure decisions made. This also includes cost estimates. - A terraform script that creates the reference architecture documented. This script can be used by users at their own discretion (i.e. it is not officially supported), and will be used to run internal automated tests see #7316) ### Measurement - Qualitative feedback on reference artchitecture from users that have set it up - Feedback from Sales Engineering who have accompanied customers in PoC phases that used it - Community feedback ### Issues - [x] https://github.com/gitpod-io/gitpod/issues/11225 - [x] https://github.com/gitpod-io/gitpod/issues/11226 - [x] https://github.com/gitpod-io/gitpod/issues/12001 - [ ] https://github.com/gitpod-io/gitpod/issues/12517
non_priority
epic proof of value reference architecture summary create a specific reference architecture for a poc phase of the customer life cycle aimed at balancing ease of set up with enough scale to prove gitpod s value in a team setting context we are working on a reference architecture for prod settings this reference archtiecture is aimed at creating a scalable reliable way to operate gitpod in a production setting for customers only wanting to try out gitpod in a poc phase i e they have already tried the locally and now want to test with their team this prod reference architecture is too cumbersome to set up and there is no need to create such a reliable set up in this phase yet important this is not a self hosted team project but a gitpod project all teams need to give their input the self hosted team may just be coordinating value speed up the poc phase of the sales pipeline give the community a easy to set up way of creating a gitpod installation that should work for a lot of use cases but should not be used in prod acceptance criteria detailed documentation of the reference architecture and the reasoning for the infrastructure decisions made this also includes cost estimates a terraform script that creates the reference architecture documented this script can be used by users at their own discretion i e it is not officially supported and will be used to run internal automated tests see measurement qualitative feedback on reference artchitecture from users that have set it up feedback from sales engineering who have accompanied customers in poc phases that used it community feedback issues
0
682,266
23,339,048,825
IssuesEvent
2022-08-09 12:37:42
googleapis/google-api-java-client-services
https://api.github.com/repos/googleapis/google-api-java-client-services
closed
Issue with google-api-client 2.0.0 (originally YouTube Client)
type: bug priority: p1
I am trying to create YouTube instance which does not work since there is a internal check if the major version is more then one or not. This check is keeps us from creating a YouTube instance, which would work without the check since the 2.0.0 release of the Google-Client-API did not have any breaking changes for the YouTube Library.
1.0
Issue with google-api-client 2.0.0 (originally YouTube Client) - I am trying to create YouTube instance which does not work since there is a internal check if the major version is more then one or not. This check is keeps us from creating a YouTube instance, which would work without the check since the 2.0.0 release of the Google-Client-API did not have any breaking changes for the YouTube Library.
priority
issue with google api client originally youtube client i am trying to create youtube instance which does not work since there is a internal check if the major version is more then one or not this check is keeps us from creating a youtube instance which would work without the check since the release of the google client api did not have any breaking changes for the youtube library
1
2,103
3,046,983,088
IssuesEvent
2015-08-11 00:02:03
lionheart/openradar-mirror
https://api.github.com/repos/lionheart/openradar-mirror
opened
18541225: Searching ctrl+cmd+space character menu for "lambda" does not return lambda.
classification:ui/usability reproducible:always
#### Description If I bring up the emoji/character selection menu in a text field by pressing ctrl-cmd-space and type "lambda", the only result is the ƛ (latin small letter lambda with stroke (U+019B)), rather than any variant of the greek letter lambda. This is because the unicode spec misspells "lambda" as "lamda", but I still should be able to search for the character by its proper spelling. - Product Version: OSX 10.9.4 build 13E28 Created: 2014-10-03T19:29:25.975657 Originated: 2014-10-03T00:00:00 Open Radar Link: http://www.openradar.me/18541225
True
18541225: Searching ctrl+cmd+space character menu for "lambda" does not return lambda. - #### Description If I bring up the emoji/character selection menu in a text field by pressing ctrl-cmd-space and type "lambda", the only result is the ƛ (latin small letter lambda with stroke (U+019B)), rather than any variant of the greek letter lambda. This is because the unicode spec misspells "lambda" as "lamda", but I still should be able to search for the character by its proper spelling. - Product Version: OSX 10.9.4 build 13E28 Created: 2014-10-03T19:29:25.975657 Originated: 2014-10-03T00:00:00 Open Radar Link: http://www.openradar.me/18541225
non_priority
searching ctrl cmd space character menu for lambda does not return lambda description if i bring up the emoji character selection menu in a text field by pressing ctrl cmd space and type lambda the only result is the ƛ latin small letter lambda with stroke u rather than any variant of the greek letter lambda this is because the unicode spec misspells lambda as lamda but i still should be able to search for the character by its proper spelling product version osx build created originated open radar link
0
724,437
24,930,759,644
IssuesEvent
2022-10-31 11:24:40
Chr1s70ph/ETIT-Master
https://api.github.com/repos/Chr1s70ph/ETIT-Master
closed
Improve exams and weekplan commands
enhancement Priority: Medium accepted In Progress
Both commands still require comments, reformatting, and in case of the exams command: sorting of the results
1.0
Improve exams and weekplan commands - Both commands still require comments, reformatting, and in case of the exams command: sorting of the results
priority
improve exams and weekplan commands both commands still require comments reformatting and in case of the exams command sorting of the results
1
55,908
3,075,126,316
IssuesEvent
2015-08-20 11:50:30
pavel-pimenov/flylinkdc-r5xx
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
opened
r500 не пересчитывает интервалы поиска
bug imported Performance Priority-High
_From [Tirael...@gmail.com](https://code.google.com/u/108935377450235604965/) on March 11, 2011 09:45:45_ Если открыть несколько поисковых вкладок, а потом закрывать их до того как они поищутся следующий поиск всё равно происходит с заранее установленному интервалу, в r400 если закрывать вкладку поиска у последующих пересчитывается интервал. Реализуйте пожалуйста это и в r500 . _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=389_
1.0
r500 не пересчитывает интервалы поиска - _From [Tirael...@gmail.com](https://code.google.com/u/108935377450235604965/) on March 11, 2011 09:45:45_ Если открыть несколько поисковых вкладок, а потом закрывать их до того как они поищутся следующий поиск всё равно происходит с заранее установленному интервалу, в r400 если закрывать вкладку поиска у последующих пересчитывается интервал. Реализуйте пожалуйста это и в r500 . _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=389_
priority
не пересчитывает интервалы поиска from on march если открыть несколько поисковых вкладок а потом закрывать их до того как они поищутся следующий поиск всё равно происходит с заранее установленному интервалу в если закрывать вкладку поиска у последующих пересчитывается интервал реализуйте пожалуйста это и в original issue
1
8,461
2,989,997,556
IssuesEvent
2015-07-21 05:47:55
dotnet/wcf
https://api.github.com/repos/dotnet/wcf
closed
CreateChannel_Using_Http_NoSecurity test fails in NET Native
test bug
The assert implies the exception message is wrong. When resolving this issue, please make a better Assert message to show what the wrong message was. Running Test: DuplexChannelFactoryTest.CreateChannel_Using_Http_NoSecurity Assert.True:InvalidCastException exception string should contain 'IRequestChannel' Xunit.AssertTestException: Assert.True: InvalidCastException exception string should contain 'IRequestChannel' at Xunit.Assert.HandleFail(String assertionName, String message) in e:\ProjectK\src\QA\ToF\tests\FX\Conformance\Common\Xunit\Assertion.cs:line 592 at Xunit.Assert.True(Boolean condition, String message) in e:\ProjectK\src\QA\ToF\tests\FX\Conformance\Common\Xunit\Assertion.cs:line 147 at DuplexChannelFactoryTest.CreateChannel_Using_Http_NoSecurity() in e:\OSS\wcf\src\System.ServiceModel.Primitives\tests\ServiceModel\DuplexChannelFactoryTest.cs:line 174 at EntryPointMain.b__1a() in e:\ProjectK\src\QA\ToF\PN\x86\dbg\IL\FX\Conformance\System.Private.ServiceModel\4.0.0.0\WcfOpen\unit\System.ServiceModel.Primitives\System.ServiceModel.Primitives.Tests.main.cs:line 41 at System.Action.InvokeOpenStaticThunk() at System.Action.Invoke() at CoreFXTestLibrary.Internal.Runner.RunTestMethod$catch$0() in e:\ProjectK\src\QA\ToF\tests\FX\Conformance\Common\Xunit\Internal\Runner.cs:line 160 ---- Test FAILED ---------------
1.0
CreateChannel_Using_Http_NoSecurity test fails in NET Native - The assert implies the exception message is wrong. When resolving this issue, please make a better Assert message to show what the wrong message was. Running Test: DuplexChannelFactoryTest.CreateChannel_Using_Http_NoSecurity Assert.True:InvalidCastException exception string should contain 'IRequestChannel' Xunit.AssertTestException: Assert.True: InvalidCastException exception string should contain 'IRequestChannel' at Xunit.Assert.HandleFail(String assertionName, String message) in e:\ProjectK\src\QA\ToF\tests\FX\Conformance\Common\Xunit\Assertion.cs:line 592 at Xunit.Assert.True(Boolean condition, String message) in e:\ProjectK\src\QA\ToF\tests\FX\Conformance\Common\Xunit\Assertion.cs:line 147 at DuplexChannelFactoryTest.CreateChannel_Using_Http_NoSecurity() in e:\OSS\wcf\src\System.ServiceModel.Primitives\tests\ServiceModel\DuplexChannelFactoryTest.cs:line 174 at EntryPointMain.b__1a() in e:\ProjectK\src\QA\ToF\PN\x86\dbg\IL\FX\Conformance\System.Private.ServiceModel\4.0.0.0\WcfOpen\unit\System.ServiceModel.Primitives\System.ServiceModel.Primitives.Tests.main.cs:line 41 at System.Action.InvokeOpenStaticThunk() at System.Action.Invoke() at CoreFXTestLibrary.Internal.Runner.RunTestMethod$catch$0() in e:\ProjectK\src\QA\ToF\tests\FX\Conformance\Common\Xunit\Internal\Runner.cs:line 160 ---- Test FAILED ---------------
non_priority
createchannel using http nosecurity test fails in net native the assert implies the exception message is wrong when resolving this issue please make a better assert message to show what the wrong message was running test duplexchannelfactorytest createchannel using http nosecurity assert true invalidcastexception exception string should contain irequestchannel xunit asserttestexception assert true invalidcastexception exception string should contain irequestchannel at xunit assert handlefail string assertionname string message in e projectk src qa tof tests fx conformance common xunit assertion cs line at xunit assert true boolean condition string message in e projectk src qa tof tests fx conformance common xunit assertion cs line at duplexchannelfactorytest createchannel using http nosecurity in e oss wcf src system servicemodel primitives tests servicemodel duplexchannelfactorytest cs line at entrypointmain b in e projectk src qa tof pn dbg il fx conformance system private servicemodel wcfopen unit system servicemodel primitives system servicemodel primitives tests main cs line at system action invokeopenstaticthunk at system action invoke at corefxtestlibrary internal runner runtestmethod catch in e projectk src qa tof tests fx conformance common xunit internal runner cs line test failed
0
413,101
12,060,395,703
IssuesEvent
2020-04-15 21:07:29
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.sakshamanand.com - Green sidebar is displayed misaligned
browser-chrome priority-normal severity-minor
<!-- @browser: Chrome 69.0.3497 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/69.0.3497.100 Safari/537.36 --> <!-- @reported_with: web --> **URL**: https://www.sakshamanand.com/tipu/ **Browser / Version**: Chrome 69.0.3497 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: In the screenshot, left is Firefox, right is chrome. The white div is vertically centred in firefox, but it is near the bottom in chrome. **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2018/10/5a32a43e-db4d-4a46-90b0-1b002b747fd9-thumb.jpg)](https://webcompat.com/uploads/2018/10/5a32a43e-db4d-4a46-90b0-1b002b747fd9.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.sakshamanand.com - Green sidebar is displayed misaligned - <!-- @browser: Chrome 69.0.3497 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/69.0.3497.100 Safari/537.36 --> <!-- @reported_with: web --> **URL**: https://www.sakshamanand.com/tipu/ **Browser / Version**: Chrome 69.0.3497 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: In the screenshot, left is Firefox, right is chrome. The white div is vertically centred in firefox, but it is near the bottom in chrome. **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2018/10/5a32a43e-db4d-4a46-90b0-1b002b747fd9-thumb.jpg)](https://webcompat.com/uploads/2018/10/5a32a43e-db4d-4a46-90b0-1b002b747fd9.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
green sidebar is displayed misaligned url browser version chrome operating system windows tested another browser yes problem type design is broken description in the screenshot left is firefox right is chrome the white div is vertically centred in firefox but it is near the bottom in chrome steps to reproduce from with ❤️
1
518,067
15,023,526,953
IssuesEvent
2021-02-01 18:21:04
fecgov/fec-cms
https://api.github.com/repos/fecgov/fec-cms
closed
Wagtail training with audit team
High priority Work: Content Work: Front-end
**What we're after:** We have written documentation that explains the process of building an audit report page and we need to know conduct a training session and go through that documentation with the audit team. **Draft audit page documentation:** - https://github.com/fecgov/fec-cms/issues/4255 Draft audit page documentation **Action items:** - [x] Identify everyone who needs to attend - [x] Schedule training **Completion criteria:** - [x] Training is conducted with all necessary audit staff
1.0
Wagtail training with audit team - **What we're after:** We have written documentation that explains the process of building an audit report page and we need to know conduct a training session and go through that documentation with the audit team. **Draft audit page documentation:** - https://github.com/fecgov/fec-cms/issues/4255 Draft audit page documentation **Action items:** - [x] Identify everyone who needs to attend - [x] Schedule training **Completion criteria:** - [x] Training is conducted with all necessary audit staff
priority
wagtail training with audit team what we re after we have written documentation that explains the process of building an audit report page and we need to know conduct a training session and go through that documentation with the audit team draft audit page documentation draft audit page documentation action items identify everyone who needs to attend schedule training completion criteria training is conducted with all necessary audit staff
1
700,864
24,075,737,619
IssuesEvent
2022-09-18 19:24:02
aquasecurity/trivy
https://api.github.com/repos/aquasecurity/trivy
opened
Integrate govulncheck
kind/feature priority/important-soon scan/vulnerability
To provide low-noise results in Go projects, it should integrate `govulncheck` into Trivy. https://pkg.go.dev/golang.org/x/vuln/cmd/govulncheck
1.0
Integrate govulncheck - To provide low-noise results in Go projects, it should integrate `govulncheck` into Trivy. https://pkg.go.dev/golang.org/x/vuln/cmd/govulncheck
priority
integrate govulncheck to provide low noise results in go projects it should integrate govulncheck into trivy
1
334,836
10,146,269,218
IssuesEvent
2019-08-05 07:41:40
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.paypal.com - see bug description
browser-fenix engine-gecko priority-critical
<!-- @browser: --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.paypal.com/us/webapps/mpp/security/report-identity-theft **Browser / Version**: **Operating System**: **Tested Another Browser**: No **Problem type**: Something else **Description**: i have no account with PayPal someone has stolen my identity. **Steps to Reproduce**: I keep getting emails saying i bought items and i have not <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.paypal.com - see bug description - <!-- @browser: --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.paypal.com/us/webapps/mpp/security/report-identity-theft **Browser / Version**: **Operating System**: **Tested Another Browser**: No **Problem type**: Something else **Description**: i have no account with PayPal someone has stolen my identity. **Steps to Reproduce**: I keep getting emails saying i bought items and i have not <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
see bug description url browser version operating system tested another browser no problem type something else description i have no account with paypal someone has stolen my identity steps to reproduce i keep getting emails saying i bought items and i have not browser configuration none from with ❤️
1
73,668
24,749,595,135
IssuesEvent
2022-10-21 12:42:29
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
Unreliable heat conduction truss test in MPI+Threads recipe
P: minor T: defect
## Bug Description Stochastic test failures on distrbuted + threads ``` Working Directory: /data/civet0/build/moose/modules/heat_conduction/test/tests/truss_heat_conduction Running command: mpiexec -n 2 /data/civet0/build/moose/modules/heat_conduction/heat_conduction-opt -i block_w_line.i --distributed-mesh --error --error-override --timing Outputs/perf_graph=true --no-gdb-backtrace --n-threads=2 Framework Information: MOOSE Version: git commit a737d941ef3 on 2022-10-20 LibMesh Version: 69dad44d8ef2b871ec1a6a5b5788b00385642c39 PETSc Version: 3.16.6 SLEPc Version: 3.16.2 Current Time: Fri Oct 21 00:18:08 2022 Executable Timestamp: Fri Oct 21 00:14:06 2022 Parallelism: Num Processors: 2 Num Threads: 2 Mesh: Parallel Type: replicated (forced) Mesh Dimension: 3 Spatial Dimension: 3 Nodes: Total: 419 Local: 214 Min/Max/Avg: 205/214/209 Elems: Total: 160 Local: 80 Min/Max/Avg: 80/80/80 Num Subdomains: 2 Num Partitions: 2 Partitioner: metis Nonlinear System: Num DOFs: 419 Num Local DOFs: 214 Variables: "temperature" Finite Element Types: "LAGRANGE" Approximation Orders: "FIRST" Auxiliary System: Num DOFs: 160 Num Local DOFs: 80 Variables: "area" Finite Element Types: "MONOMIAL" Approximation Orders: "CONSTANT" Execution Information: Executioner: Transient TimeStepper: ConstantDT Solver Mode: Preconditioned JFNK PETSc Preconditioner: hypre boomeramg strong_threshold: 0.7 (auto) Time Step 0, time = 0 Time Step 1, time = 1, dt = 1 0 Nonlinear |R| = 8.124038e+06 0 Linear |R| = 8.124038e+06 1 Linear |R| = 2.936180e+06 2 Linear |R| = 1.391816e+06 3 Linear |R| = 5.426702e+02 4 Linear |R| = 1.858057e-02 1 Nonlinear |R| = 4.987863e-01 0 Linear |R| = 4.987863e-01 1 Linear |R| = 4.984998e-01 2 Linear |R| = 4.898628e-01 3 Linear |R| = 4.763700e-01 4 Linear |R| = 1.116249e-01 5 Linear |R| = 8.691014e-04 6 Linear |R| = 1.822808e-04 7 Linear |R| = 2.942690e-06 2 Nonlinear |R| = 4.987863e-01 0 Linear |R| = 4.987863e-01 1 Linear |R| = 4.984998e-01 2 Linear |R| = 4.904613e-01 3 Linear |R| = 4.770534e-01 4 Linear |R| = 1.592058e-01 5 Linear |R| = 1.990848e-03 6 Linear |R| = 6.528066e-05 7 Linear |R| = 5.683374e-06 8 Linear |R| = 4.009811e-06 3 Nonlinear |R| = 2.567383e-01 0 Linear |R| = 2.567383e-01 1 Linear |R| = 2.527316e-01 2 Linear |R| = 1.957669e-01 3 Linear |R| = 1.947641e-01 4 Linear |R| = 1.236632e-01 5 Linear |R| = 2.129886e-03 6 Linear |R| = 8.013022e-05 7 Linear |R| = 8.195196e-06 8 Linear |R| = 5.685327e-06 9 Linear |R| = 4.612431e-06 10 Linear |R| = 3.981993e-06 11 Linear |R| = 3.554832e-06 12 Linear |R| = 3.241063e-06 13 Linear |R| = 2.997947e-06 14 Linear |R| = 2.802455e-06 15 Linear |R| = 2.640804e-06 16 Linear |R| = 2.504248e-06 4 Nonlinear |R| = 2.130816e-01 0 Linear |R| = 2.130816e-01 1 Linear |R| = 2.127822e-01 2 Linear |R| = 1.951439e-01 3 Linear |R| = 1.950112e-01 4 Linear |R| = 5.776265e-02 5 Linear |R| = 5.586006e-04 6 Linear |R| = 1.314725e-05 7 Linear |R| = 1.214406e-06 5 Nonlinear |R| = 2.112063e-01 0 Linear |R| = 2.112063e-01 1 Linear |R| = 2.104202e-01 2 Linear |R| = 1.910654e-01 3 Linear |R| = 1.906017e-01 4 Linear |R| = 9.825074e-02 5 Linear |R| = 1.446873e-03 6 Linear |R| = 2.038308e-05 7 Linear |R| = 3.983396e-07 6 Nonlinear |R| = 1.993684e-01 0 Linear |R| = 1.993684e-01 1 Linear |R| = 1.940511e-01 2 Linear |R| = 1.804851e-01 3 Linear |R| = 1.655430e-01 4 Linear |R| = 6.593206e-02 5 Linear |R| = 4.145496e-03 6 Linear |R| = 1.476153e-04 7 Linear |R| = 1.368834e-06 7 Nonlinear |R| = 1.993684e-01 0 Linear |R| = 1.993684e-01 1 Linear |R| = 1.940510e-01 2 Linear |R| = 1.821076e-01 3 Linear |R| = 1.676569e-01 4 Linear |R| = 1.486711e-01 5 Linear |R| = 2.073761e-03 6 Linear |R| = 7.024214e-05 7 Linear |R| = 2.797172e-06 8 Linear |R| = 1.987650e-06 8 Nonlinear |R| = 1.832074e-01 0 Linear |R| = 1.832074e-01 1 Linear |R| = 1.747611e-01 2 Linear |R| = 1.388639e-01 3 Linear |R| = 1.321860e-01 4 Linear |R| = 9.827800e-02 5 Linear |R| = 3.621897e-03 6 Linear |R| = 7.397675e-05 7 Linear |R| = 1.292977e-05 8 Linear |R| = 8.937240e-06 9 Linear |R| = 7.237751e-06 10 Linear |R| = 6.242950e-06 11 Linear |R| = 5.570493e-06 12 Linear |R| = 5.077053e-06 13 Linear |R| = 4.695115e-06 14 Linear |R| = 4.388160e-06 15 Linear |R| = 4.134479e-06 16 Linear |R| = 3.920247e-06 17 Linear |R| = 3.736199e-06 18 Linear |R| = 3.575857e-06 19 Linear |R| = 3.434531e-06 20 Linear |R| = 3.308739e-06 21 Linear |R| = 3.195827e-06 22 Linear |R| = 3.093738e-06 23 Linear |R| = 3.000846e-06 24 Linear |R| = 2.915848e-06 25 Linear |R| = 2.837687e-06 26 Linear |R| = 2.765492e-06 27 Linear |R| = 2.698540e-06 28 Linear |R| = 2.636228e-06 29 Linear |R| = 2.578041e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 9 Nonlinear |R| = 1.832074e-01 0 Linear |R| = 1.832074e-01 1 Linear |R| = 1.747611e-01 2 Linear |R| = 1.402027e-01 3 Linear |R| = 1.321711e-01 4 Linear |R| = 1.215777e-01 5 Linear |R| = 5.027232e-03 6 Linear |R| = 1.297047e-04 7 Linear |R| = 3.307898e-05 8 Linear |R| = 2.397256e-05 9 Linear |R| = 1.973076e-05 10 Linear |R| = 1.715673e-05 11 Linear |R| = 1.538315e-05 12 Linear |R| = 1.406595e-05 13 Linear |R| = 1.303788e-05 14 Linear |R| = 1.220676e-05 15 Linear |R| = 1.151661e-05 16 Linear |R| = 1.093176e-05 17 Linear |R| = 1.042776e-05 18 Linear |R| = 9.987631e-06 19 Linear |R| = 9.598842e-06 20 Linear |R| = 9.252230e-06 21 Linear |R| = 8.940646e-06 22 Linear |R| = 8.658564e-06 23 Linear |R| = 8.401597e-06 24 Linear |R| = 8.166232e-06 25 Linear |R| = 7.949599e-06 26 Linear |R| = 7.749337e-06 27 Linear |R| = 7.563487e-06 28 Linear |R| = 7.390402e-06 29 Linear |R| = 7.228673e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 Nonlinear solve did not converge due to DIVERGED_LINE_SEARCH iterations 9 Solve Did NOT Converge! Aborting as solve did not converge Time Step 1, time = 0.5, dt = 0.5 0 Nonlinear |R| = 4.062019e+06 0 Linear |R| = 4.062019e+06 1 Linear |R| = 1.408872e+06 2 Linear |R| = 5.767703e+05 3 Linear |R| = 1.337262e+03 4 Linear |R| = 3.477387e-03 1 Nonlinear |R| = 6.662115e-02 0 Linear |R| = 6.662115e-02 1 Linear |R| = 6.661958e-02 2 Linear |R| = 6.661213e-02 3 Linear |R| = 6.651483e-02 4 Linear |R| = 4.531902e-02 5 Linear |R| = 1.589527e-04 6 Linear |R| = 6.840782e-06 7 Linear |R| = 2.365846e-06 8 Linear |R| = 1.703323e-06 9 Linear |R| = 1.399136e-06 10 Linear |R| = 1.215352e-06 11 Linear |R| = 1.089037e-06 12 Linear |R| = 9.953671e-07 13 Linear |R| = 9.223387e-07 14 Linear |R| = 8.633339e-07 15 Linear |R| = 8.143765e-07 16 Linear |R| = 7.729024e-07 17 Linear |R| = 7.371806e-07 18 Linear |R| = 7.059931e-07 19 Linear |R| = 6.784570e-07 20 Linear |R| = 6.539097e-07 2 Nonlinear |R| = 4.354222e-02 0 Linear |R| = 4.354222e-02 1 Linear |R| = 4.261640e-02 2 Linear |R| = 4.259191e-02 3 Linear |R| = 4.191010e-02 4 Linear |R| = 4.166622e-02 5 Linear |R| = 4.131442e-04 6 Linear |R| = 4.163222e-05 7 Linear |R| = 2.060248e-05 8 Linear |R| = 1.596644e-05 9 Linear |R| = 1.346830e-05 10 Linear |R| = 1.186748e-05 11 Linear |R| = 1.072727e-05 12 Linear |R| = 9.864290e-06 13 Linear |R| = 9.180288e-06 14 Linear |R| = 8.621530e-06 15 Linear |R| = 8.153597e-06 16 Linear |R| = 7.754596e-06 17 Linear |R| = 7.408844e-06 18 Linear |R| = 7.105672e-06 19 Linear |R| = 6.836765e-06 20 Linear |R| = 6.596315e-06 21 Linear |R| = 6.379559e-06 22 Linear |R| = 6.182856e-06 23 Linear |R| = 6.003290e-06 24 Linear |R| = 5.838525e-06 25 Linear |R| = 5.686607e-06 26 Linear |R| = 5.546020e-06 27 Linear |R| = 5.415313e-06 28 Linear |R| = 5.293485e-06 29 Linear |R| = 5.179491e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 3 Nonlinear |R| = 4.261368e-02 0 Linear |R| = 4.261368e-02 1 Linear |R| = 4.210921e-02 2 Linear |R| = 4.162688e-02 3 Linear |R| = 3.903726e-02 4 Linear |R| = 2.703674e-02 5 Linear |R| = 1.452146e-03 6 Linear |R| = 1.293879e-04 7 Linear |R| = 4.296561e-05 8 Linear |R| = 3.191541e-05 9 Linear |R| = 2.650371e-05 10 Linear |R| = 2.315369e-05 11 Linear |R| = 2.081924e-05 12 Linear |R| = 1.907311e-05 13 Linear |R| = 1.770351e-05 14 Linear |R| = 1.659209e-05 15 Linear |R| = 1.566671e-05 16 Linear |R| = 1.488070e-05 17 Linear |R| = 1.420223e-05 18 Linear |R| = 1.360884e-05 19 Linear |R| = 1.308411e-05 20 Linear |R| = 1.261575e-05 21 Linear |R| = 1.219433e-05 22 Linear |R| = 1.181250e-05 23 Linear |R| = 1.146443e-05 24 Linear |R| = 1.114543e-05 25 Linear |R| = 1.085164e-05 26 Linear |R| = 1.057994e-05 27 Linear |R| = 1.032766e-05 28 Linear |R| = 1.009262e-05 29 Linear |R| = 9.872924e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 4 Nonlinear |R| = 3.713079e-02 Solve Converged! Time Step 2, time = 1, dt = 0.5 0 Nonlinear |R| = 4.062019e+06 0 Linear |R| = 4.062019e+06 1 Linear |R| = 1.519291e+06 2 Linear |R| = 7.680909e+05 3 Linear |R| = 5.089047e+02 4 Linear |R| = 2.470842e-01 1 Nonlinear |R| = 2.320993e-01 0 Linear |R| = 2.320993e-01 1 Linear |R| = 2.316175e-01 2 Linear |R| = 2.097271e-01 3 Linear |R| = 1.765354e-01 4 Linear |R| = 1.686036e-01 5 Linear |R| = 9.050986e-02 6 Linear |R| = 2.253444e-02 7 Linear |R| = 1.103272e-03 8 Linear |R| = 6.923692e-04 9 Linear |R| = 5.537875e-04 10 Linear |R| = 4.737331e-04 11 Linear |R| = 4.208348e-04 12 Linear |R| = 3.823479e-04 13 Linear |R| = 3.528327e-04 14 Linear |R| = 3.292417e-04 15 Linear |R| = 3.098151e-04 16 Linear |R| = 2.934741e-04 17 Linear |R| = 2.794668e-04 18 Linear |R| = 2.672972e-04 19 Linear |R| = 2.565894e-04 20 Linear |R| = 2.470732e-04 21 Linear |R| = 2.385429e-04 22 Linear |R| = 2.308397e-04 23 Linear |R| = 2.238356e-04 24 Linear |R| = 2.174344e-04 25 Linear |R| = 2.115512e-04 26 Linear |R| = 2.061221e-04 27 Linear |R| = 2.010907e-04 28 Linear |R| = 1.964105e-04 29 Linear |R| = 1.920427e-04 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 2 Nonlinear |R| = 2.320993e-01 0 Linear |R| = 2.320993e-01 1 Linear |R| = 2.316175e-01 2 Linear |R| = 2.096388e-01 3 Linear |R| = 1.856877e-01 4 Linear |R| = 1.841090e-01 5 Linear |R| = 1.931490e-02 6 Linear |R| = 1.679579e-02 7 Linear |R| = 2.409159e-03 8 Linear |R| = 1.728413e-03 9 Linear |R| = 1.419190e-03 10 Linear |R| = 1.232370e-03 11 Linear |R| = 1.104061e-03 12 Linear |R| = 1.008943e-03 13 Linear |R| = 9.348332e-04 14 Linear |R| = 8.749601e-04 15 Linear |R| = 8.252987e-04 16 Linear |R| = 7.832360e-04 17 Linear |R| = 7.470093e-04 18 Linear |R| = 7.153860e-04 19 Linear |R| = 6.874650e-04 20 Linear |R| = 6.625766e-04 21 Linear |R| = 6.402134e-04 22 Linear |R| = 6.199684e-04 23 Linear |R| = 6.015305e-04 24 Linear |R| = 5.846441e-04 25 Linear |R| = 5.691063e-04 26 Linear |R| = 5.547439e-04 27 Linear |R| = 5.414172e-04 28 Linear |R| = 5.290064e-04 29 Linear |R| = 5.174122e-04 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 3 Nonlinear |R| = 2.320870e-01 0 Linear |R| = 2.320870e-01 1 Linear |R| = 2.316584e-01 2 Linear |R| = 2.097528e-01 3 Linear |R| = 2.030920e-01 4 Linear |R| = 1.834407e-01 5 Linear |R| = 1.819411e-01 6 Linear |R| = 8.400089e-03 7 Linear |R| = 8.859178e-04 8 Linear |R| = 6.315971e-04 9 Linear |R| = 5.176682e-04 10 Linear |R| = 4.490556e-04 11 Linear |R| = 4.021027e-04 12 Linear |R| = 3.673259e-04 13 Linear |R| = 3.402699e-04 14 Linear |R| = 3.184134e-04 15 Linear |R| = 3.003021e-04 16 Linear |R| = 2.849593e-04 17 Linear |R| = 2.717568e-04 18 Linear |R| = 2.602292e-04 19 Linear |R| = 2.500587e-04 20 Linear |R| = 2.409907e-04 21 Linear |R| = 2.328458e-04 22 Linear |R| = 2.254719e-04 23 Linear |R| = 2.187591e-04 24 Linear |R| = 2.126104e-04 25 Linear |R| = 2.069542e-04 26 Linear |R| = 2.017253e-04 27 Linear |R| = 1.968749e-04 28 Linear |R| = 1.923573e-04 29 Linear |R| = 1.881380e-04 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 Nonlinear solve did not converge due to DIVERGED_LINE_SEARCH iterations 3 Solve Did NOT Converge! Aborting as solve did not converge Time Step 2, time = 0.75, dt = 0.25 0 Nonlinear |R| = 2.031010e+06 0 Linear |R| = 2.031010e+06 1 Linear |R| = 7.460102e+05 2 Linear |R| = 3.531009e+05 3 Linear |R| = 6.736684e+03 4 Linear |R| = 6.661378e-03 1 Nonlinear |R| = 3.293862e-01 0 Linear |R| = 3.293862e-01 1 Linear |R| = 2.214630e-01 2 Linear |R| = 2.130098e-01 3 Linear |R| = 2.130038e-01 4 Linear |R| = 4.081695e-03 5 Linear |R| = 1.601902e-05 6 Linear |R| = 2.080131e-06 2 Nonlinear |R| = 2.395315e-01 0 Linear |R| = 2.395315e-01 1 Linear |R| = 1.214649e-01 2 Linear |R| = 2.149745e-02 3 Linear |R| = 2.090277e-04 4 Linear |R| = 2.536993e-10 3 Nonlinear |R| = 5.566991e-09 Solve Converged! Time Step 3, time = 1, dt = 0.25 0 Nonlinear |R| = 2.031010e+06 0 Linear |R| = 2.031010e+06 1 Linear |R| = 7.682209e+05 2 Linear |R| = 3.924322e+05 3 Linear |R| = 6.833789e+03 4 Linear |R| = 7.331835e-03 1 Nonlinear |R| = 3.380724e-01 0 Linear |R| = 3.380724e-01 1 Linear |R| = 3.001736e-01 2 Linear |R| = 2.994000e-01 3 Linear |R| = 2.638543e-01 4 Linear |R| = 3.343074e-03 5 Linear |R| = 6.175358e-05 6 Linear |R| = 2.374424e-07 2 Nonlinear |R| = 1.834137e-01 0 Linear |R| = 1.834137e-01 1 Linear |R| = 1.576530e-01 2 Linear |R| = 4.220904e-02 3 Linear |R| = 4.207755e-04 4 Linear |R| = 2.969888e-10 3 Nonlinear |R| = 1.813251e-08 Solve Converged! Performance Graph: -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | Section | Calls | Self(s) | Avg(s) | % | Mem(MB) | Total(s) | Avg(s) | % | Mem(MB) | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | HeatConductionTestApp (main) | 1 | 0.003 | 0.003 | 0.36 | 2 | 0.904 | 0.904 | 100.00 | 40 | | Action::SetupMeshAction::Mesh::SetupMeshAction::act::setup_mesh | 1 | 0.001 | 0.001 | 0.07 | 0 | 0.001 | 0.001 | 0.07 | 0 | | Action::SetupMeshAction::Mesh::SetupMeshAction::act::set_mesh_base | 2 | 0.000 | 0.000 | 0.00 | 0 | 0.000 | 0.000 | 0.00 | 0 | | MooseApp::executeMeshGenerators | 1 | 0.000 | 0.000 | 0.00 | 0 | 0.000 | 0.000 | 0.00 | 0 | | FEProblem::outputStep | 4 | 0.000 | 0.000 | 0.03 | 0 | 0.016 | 0.004 | 1.82 | 7 | | Exodus::outputStep | 4 | 0.013 | 0.003 | 1.42 | 7 | 0.013 | 0.003 | 1.42 | 7 | | Transient::PicardSolve | 5 | 0.001 | 0.000 | 0.16 | 0 | 0.746 | 0.149 | 82.51 | 3 | | FEProblem::outputStep | 7 | 0.000 | 0.000 | 0.02 | 0 | 0.000 | 0.000 | 0.03 | 0 | | FEProblem::solve | 5 | 0.143 | 0.029 | 15.77 | 3 | 0.735 | 0.147 | 81.28 | 3 | | FEProblem::computeResidualInternal | 5 | 0.000 | 0.000 | 0.01 | 0 | 0.006 | 0.001 | 0.63 | 0 | | FEProblem::computeResidualInternal | 490 | 0.008 | 0.000 | 0.91 | 0 | 0.523 | 0.001 | 57.83 | 0 | | FEProblem::computeJacobianInternal | 24 | 0.000 | 0.000 | 0.04 | 0 | 0.051 | 0.002 | 5.63 | 0 | | FEProblem::computeUserObjects | 3 | 0.008 | 0.003 | 0.83 | 0 | 0.008 | 0.003 | 0.83 | 0 | | Transient::final | 1 | 0.000 | 0.000 | 0.01 | 0 | 0.000 | 0.000 | 0.02 | 0 | | FEProblem::outputStep | 1 | 0.000 | 0.000 | 0.00 | 0 | 0.000 | 0.000 | 0.00 | 0 | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Framework Information: MOOSE Version: git commit a737d941ef3 on 2022-10-20 LibMesh Version: 69dad44d8ef2b871ec1a6a5b5788b00385642c39 PETSc Version: 3.16.6 SLEPc Version: 3.16.2 Current Time: Fri Oct 21 00:18:08 2022 Executable Timestamp: Fri Oct 21 00:14:06 2022 Parallelism: Num Processors: 2 Num Threads: 2 Mesh: Parallel Type: replicated (forced) Mesh Dimension: 3 Spatial Dimension: 3 Nodes: Total: 419 Local: 214 Min/Max/Avg: 205/214/209 Elems: Total: 160 Local: 80 Min/Max/Avg: 80/80/80 Num Subdomains: 2 Num Partitions: 2 Partitioner: metis Nonlinear System: Num DOFs: 419 Num Local DOFs: 214 Variables: "temperature" Finite Element Types: "LAGRANGE" Approximation Orders: "FIRST" Auxiliary System: Num DOFs: 160 Num Local DOFs: 80 Variables: "area" Finite Element Types: "MONOMIAL" Approximation Orders: "CONSTANT" Execution Information: Executioner: Transient TimeStepper: ConstantDT Solver Mode: Preconditioned JFNK PETSc Preconditioner: hypre boomeramg strong_threshold: 0.7 (auto) Time Step 0, time = 0 Time Step 1, time = 1, dt = 1 0 Nonlinear |R| = 8.124038e+06 0 Linear |R| = 8.124038e+06 1 Linear |R| = 2.936180e+06 2 Linear |R| = 1.391816e+06 3 Linear |R| = 5.426702e+02 4 Linear |R| = 1.858057e-02 1 Nonlinear |R| = 4.987863e-01 0 Linear |R| = 4.987863e-01 1 Linear |R| = 4.984998e-01 2 Linear |R| = 4.898628e-01 3 Linear |R| = 4.763700e-01 4 Linear |R| = 1.116249e-01 5 Linear |R| = 8.691014e-04 6 Linear |R| = 1.822808e-04 7 Linear |R| = 2.942690e-06 2 Nonlinear |R| = 4.987863e-01 0 Linear |R| = 4.987863e-01 1 Linear |R| = 4.984998e-01 2 Linear |R| = 4.904613e-01 3 Linear |R| = 4.770534e-01 4 Linear |R| = 1.592058e-01 5 Linear |R| = 1.990848e-03 6 Linear |R| = 6.528066e-05 7 Linear |R| = 5.683374e-06 8 Linear |R| = 4.009811e-06 3 Nonlinear |R| = 2.567383e-01 0 Linear |R| = 2.567383e-01 1 Linear |R| = 2.527316e-01 2 Linear |R| = 1.957669e-01 3 Linear |R| = 1.947641e-01 4 Linear |R| = 1.236632e-01 5 Linear |R| = 2.129886e-03 6 Linear |R| = 8.013022e-05 7 Linear |R| = 8.195196e-06 8 Linear |R| = 5.685327e-06 9 Linear |R| = 4.612431e-06 10 Linear |R| = 3.981993e-06 11 Linear |R| = 3.554832e-06 12 Linear |R| = 3.241063e-06 13 Linear |R| = 2.997947e-06 14 Linear |R| = 2.802455e-06 15 Linear |R| = 2.640804e-06 16 Linear |R| = 2.504248e-06 4 Nonlinear |R| = 2.130816e-01 0 Linear |R| = 2.130816e-01 1 Linear |R| = 2.127822e-01 2 Linear |R| = 1.951439e-01 3 Linear |R| = 1.950112e-01 4 Linear |R| = 5.776265e-02 5 Linear |R| = 5.586006e-04 6 Linear |R| = 1.314725e-05 7 Linear |R| = 1.214406e-06 5 Nonlinear |R| = 2.112063e-01 0 Linear |R| = 2.112063e-01 1 Linear |R| = 2.104202e-01 2 Linear |R| = 1.910654e-01 3 Linear |R| = 1.906017e-01 4 Linear |R| = 9.825074e-02 5 Linear |R| = 1.446873e-03 6 Linear |R| = 2.038308e-05 7 Linear |R| = 3.983396e-07 6 Nonlinear |R| = 1.993684e-01 0 Linear |R| = 1.993684e-01 1 Linear |R| = 1.940511e-01 2 Linear |R| = 1.804851e-01 3 Linear |R| = 1.655430e-01 4 Linear |R| = 6.593206e-02 5 Linear |R| = 4.145496e-03 6 Linear |R| = 1.476153e-04 7 Linear |R| = 1.368834e-06 7 Nonlinear |R| = 1.993684e-01 0 Linear |R| = 1.993684e-01 1 Linear |R| = 1.940510e-01 2 Linear |R| = 1.821076e-01 3 Linear |R| = 1.676569e-01 4 Linear |R| = 1.486711e-01 5 Linear |R| = 2.073761e-03 6 Linear |R| = 7.024214e-05 7 Linear |R| = 2.797172e-06 8 Linear |R| = 1.987650e-06 8 Nonlinear |R| = 1.832074e-01 0 Linear |R| = 1.832074e-01 1 Linear |R| = 1.747611e-01 2 Linear |R| = 1.388639e-01 3 Linear |R| = 1.321860e-01 4 Linear |R| = 9.827800e-02 5 Linear |R| = 3.621897e-03 6 Linear |R| = 7.397675e-05 7 Linear |R| = 1.292977e-05 8 Linear |R| = 8.937240e-06 9 Linear |R| = 7.237751e-06 10 Linear |R| = 6.242950e-06 11 Linear |R| = 5.570493e-06 12 Linear |R| = 5.077053e-06 13 Linear |R| = 4.695115e-06 14 Linear |R| = 4.388160e-06 15 Linear |R| = 4.134479e-06 16 Linear |R| = 3.920247e-06 17 Linear |R| = 3.736199e-06 18 Linear |R| = 3.575857e-06 19 Linear |R| = 3.434531e-06 20 Linear |R| = 3.308739e-06 21 Linear |R| = 3.195827e-06 22 Linear |R| = 3.093738e-06 23 Linear |R| = 3.000846e-06 24 Linear |R| = 2.915848e-06 25 Linear |R| = 2.837687e-06 26 Linear |R| = 2.765492e-06 27 Linear |R| = 2.698540e-06 28 Linear |R| = 2.636228e-06 29 Linear |R| = 2.578041e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 9 Nonlinear |R| = 1.832074e-01 0 Linear |R| = 1.832074e-01 1 Linear |R| = 1.747611e-01 2 Linear |R| = 1.402027e-01 3 Linear |R| = 1.321711e-01 4 Linear |R| = 1.215777e-01 5 Linear |R| = 5.027232e-03 6 Linear |R| = 1.297047e-04 7 Linear |R| = 3.307898e-05 8 Linear |R| = 2.397256e-05 9 Linear |R| = 1.973076e-05 10 Linear |R| = 1.715673e-05 11 Linear |R| = 1.538315e-05 12 Linear |R| = 1.406595e-05 13 Linear |R| = 1.303788e-05 14 Linear |R| = 1.220676e-05 15 Linear |R| = 1.151661e-05 16 Linear |R| = 1.093176e-05 17 Linear |R| = 1.042776e-05 18 Linear |R| = 9.987631e-06 19 Linear |R| = 9.598842e-06 20 Linear |R| = 9.252230e-06 21 Linear |R| = 8.940646e-06 22 Linear |R| = 8.658564e-06 23 Linear |R| = 8.401597e-06 24 Linear |R| = 8.166232e-06 25 Linear |R| = 7.949599e-06 26 Linear |R| = 7.749337e-06 27 Linear |R| = 7.563487e-06 28 Linear |R| = 7.390402e-06 29 Linear |R| = 7.228673e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 Nonlinear solve did not converge due to DIVERGED_LINE_SEARCH iterations 9 Solve Did NOT Converge! Aborting as solve did not converge Time Step 1, time = 0.5, dt = 0.5 0 Nonlinear |R| = 4.062019e+06 0 Linear |R| = 4.062019e+06 1 Linear |R| = 1.408872e+06 2 Linear |R| = 5.767703e+05 3 Linear |R| = 1.337262e+03 4 Linear |R| = 3.477387e-03 1 Nonlinear |R| = 6.662115e-02 0 Linear |R| = 6.662115e-02 1 Linear |R| = 6.661958e-02 2 Linear |R| = 6.661213e-02 3 Linear |R| = 6.651483e-02 4 Linear |R| = 4.531902e-02 5 Linear |R| = 1.589527e-04 6 Linear |R| = 6.840782e-06 7 Linear |R| = 2.365846e-06 8 Linear |R| = 1.703323e-06 9 Linear |R| = 1.399136e-06 10 Linear |R| = 1.215352e-06 11 Linear |R| = 1.089037e-06 12 Linear |R| = 9.953671e-07 13 Linear |R| = 9.223387e-07 14 Linear |R| = 8.633339e-07 15 Linear |R| = 8.143765e-07 16 Linear |R| = 7.729024e-07 17 Linear |R| = 7.371806e-07 18 Linear |R| = 7.059931e-07 19 Linear |R| = 6.784570e-07 20 Linear |R| = 6.539097e-07 2 Nonlinear |R| = 4.354222e-02 0 Linear |R| = 4.354222e-02 1 Linear |R| = 4.261640e-02 2 Linear |R| = 4.259191e-02 3 Linear |R| = 4.191010e-02 4 Linear |R| = 4.166622e-02 5 Linear |R| = 4.131442e-04 6 Linear |R| = 4.163222e-05 7 Linear |R| = 2.060248e-05 8 Linear |R| = 1.596644e-05 9 Linear |R| = 1.346830e-05 10 Linear |R| = 1.186748e-05 11 Linear |R| = 1.072727e-05 12 Linear |R| = 9.864290e-06 13 Linear |R| = 9.180288e-06 14 Linear |R| = 8.621530e-06 15 Linear |R| = 8.153597e-06 16 Linear |R| = 7.754596e-06 17 Linear |R| = 7.408844e-06 18 Linear |R| = 7.105672e-06 19 Linear |R| = 6.836765e-06 20 Linear |R| = 6.596315e-06 21 Linear |R| = 6.379559e-06 22 Linear |R| = 6.182856e-06 23 Linear |R| = 6.003290e-06 24 Linear |R| = 5.838525e-06 25 Linear |R| = 5.686607e-06 26 Linear |R| = 5.546020e-06 27 Linear |R| = 5.415313e-06 28 Linear |R| = 5.293485e-06 29 Linear |R| = 5.179491e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 3 Nonlinear |R| = 4.261368e-02 0 Linear |R| = 4.261368e-02 1 Linear |R| = 4.210921e-02 2 Linear |R| = 4.162688e-02 3 Linear |R| = 3.903726e-02 4 Linear |R| = 2.703674e-02 5 Linear |R| = 1.452146e-03 6 Linear |R| = 1.293879e-04 7 Linear |R| = 4.296561e-05 8 Linear |R| = 3.191541e-05 9 Linear |R| = 2.650371e-05 10 Linear |R| = 2.315369e-05 11 Linear |R| = 2.081924e-05 12 Linear |R| = 1.907311e-05 13 Linear |R| = 1.770351e-05 14 Linear |R| = 1.659209e-05 15 Linear |R| = 1.566671e-05 16 Linear |R| = 1.488070e-05 17 Linear |R| = 1.420223e-05 18 Linear |R| = 1.360884e-05 19 Linear |R| = 1.308411e-05 20 Linear |R| = 1.261575e-05 21 Linear |R| = 1.219433e-05 22 Linear |R| = 1.181250e-05 23 Linear |R| = 1.146443e-05 24 Linear |R| = 1.114543e-05 25 Linear |R| = 1.085164e-05 26 Linear |R| = 1.057994e-05 27 Linear |R| = 1.032766e-05 28 Linear |R| = 1.009262e-05 29 Linear |R| = 9.872924e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 4 Nonlinear |R| = 3.713079e-02 Solve Converged! Time Step 2, time = 1, dt = 0.5 0 Nonlinear |R| = 4.062019e+06 0 Linear |R| = 4.062019e+06 1 Linear |R| = 1.519291e+06 2 Linear |R| = 7.680909e+05 3 Linear |R| = 5.089047e+02 4 Linear |R| = 2.470842e-01 1 Nonlinear |R| = 2.320993e-01 0 Linear |R| = 2.320993e-01 1 Linear |R| = 2.316175e-01 2 Linear |R| = 2.097271e-01 3 Linear |R| = 1.765354e-01 4 Linear |R| = 1.686036e-01 5 Linear |R| = 9.050986e-02 6 Linear |R| = 2.253444e-02 7 Linear |R| = 1.103272e-03 8 Linear |R| = 6.923692e-04 9 Linear |R| = 5.537875e-04 10 Linear |R| = 4.737331e-04 11 Linear |R| = 4.208348e-04 12 Linear |R| = 3.823479e-04 13 Linear |R| = 3.528327e-04 14 Linear |R| = 3.292417e-04 15 Linear |R| = 3.098151e-04 16 Linear |R| = 2.934741e-04 17 Linear |R| = 2.794668e-04 18 Linear |R| = 2.672972e-04 19 Linear |R| = 2.565894e-04 20 Linear |R| = 2.470732e-04 21 Linear |R| = 2.385429e-04 22 Linear |R| = 2.308397e-04 23 Linear |R| = 2.238356e-04 24 Linear |R| = 2.174344e-04 25 Linear |R| = 2.115512e-04 26 Linear |R| = 2.061221e-04 27 Linear |R| = 2.010907e-04 28 Linear |R| = 1.964105e-04 29 Linear |R| = 1.920427e-04 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 2 Nonlinear |R| = 2.320993e-01 0 Linear |R| = 2.320993e-01 1 Linear |R| = 2.316175e-01 2 Linear |R| = 2.096388e-01 3 Linear |R| = 1.856877e-01 4 Linear |R| = 1.841090e-01 5 Linear |R| = 1.931490e-02 6 Linear |R| = 1.679579e-02 7 Linear |R| = 2.409159e-03 8 Linear |R| = 1.728413e-03 9 Linear |R| = 1.419190e-03 10 Linear |R| = 1.232370e-03 11 Linear |R| = 1.104061e-03 12 Linear |R| = 1.008943e-03 13 Linear |R| = 9.348332e-04 14 Linear |R| = 8.749601e-04 15 Linear |R| = 8.252987e-04 16 Linear |R| = 7.832360e-04 17 Linear |R| = 7.470093e-04 18 Linear |R| = 7.153860e-04 19 Linear |R| = 6.874650e-04 20 Linear |R| = 6.625766e-04 21 Linear |R| = 6.402134e-04 22 Linear |R| = 6.199684e-04 23 Linear |R| = 6.015305e-04 24 Linear |R| = 5.846441e-04 25 Linear |R| = 5.691063e-04 26 Linear |R| = 5.547439e-04 27 Linear |R| = 5.414172e-04 28 Linear |R| = 5.290064e-04 29 Linear |R| = 5.174122e-04 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 3 Nonlinear |R| = 2.320870e-01 0 Linear |R| = 2.320870e-01 1 Linear |R| = 2.316584e-01 2 Linear |R| = 2.097528e-01 3 Linear |R| = 2.030920e-01 4 Linear |R| = 1.834407e-01 5 Linear |R| = 1.819411e-01 6 Linear |R| = 8.400089e-03 7 Linear |R| = 8.859178e-04 8 Linear |R| = 6.315971e-04 9 Linear |R| = 5.176682e-04 10 Linear |R| = 4.490556e-04 11 Linear |R| = 4.021027e-04 12 Linear |R| = 3.673259e-04 13 Linear |R| = 3.402699e-04 14 Linear |R| = 3.184134e-04 15 Linear |R| = 3.003021e-04 16 Linear |R| = 2.849593e-04 17 Linear |R| = 2.717568e-04 18 Linear |R| = 2.602292e-04 19 Linear |R| = 2.500587e-04 20 Linear |R| = 2.409907e-04 21 Linear |R| = 2.328458e-04 22 Linear |R| = 2.254719e-04 23 Linear |R| = 2.187591e-04 24 Linear |R| = 2.126104e-04 25 Linear |R| = 2.069542e-04 26 Linear |R| = 2.017253e-04 27 Linear |R| = 1.968749e-04 28 Linear |R| = 1.923573e-04 29 Linear |R| = 1.881380e-04 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 Nonlinear solve did not converge due to DIVERGED_LINE_SEARCH iterations 3 Solve Did NOT Converge! Aborting as solve did not converge Time Step 2, time = 0.75, dt = 0.25 0 Nonlinear |R| = 2.031010e+06 0 Linear |R| = 2.031010e+06 1 Linear |R| = 7.460102e+05 2 Linear |R| = 3.531009e+05 3 Linear |R| = 6.736684e+03 4 Linear |R| = 6.661378e-03 1 Nonlinear |R| = 3.293862e-01 0 Linear |R| = 3.293862e-01 1 Linear |R| = 2.214630e-01 2 Linear |R| = 2.130098e-01 3 Linear |R| = 2.130038e-01 4 Linear |R| = 4.081695e-03 5 Linear |R| = 1.601902e-05 6 Linear |R| = 2.080131e-06 2 Nonlinear |R| = 2.395315e-01 0 Linear |R| = 2.395315e-01 1 Linear |R| = 1.214649e-01 2 Linear |R| = 2.149745e-02 3 Linear |R| = 2.090277e-04 4 Linear |R| = 2.536993e-10 3 Nonlinear |R| = 5.566991e-09 Solve Converged! Time Step 3, time = 1, dt = 0.25 0 Nonlinear |R| = 2.031010e+06 0 Linear |R| = 2.031010e+06 1 Linear |R| = 7.682209e+05 2 Linear |R| = 3.924322e+05 3 Linear |R| = 6.833789e+03 4 Linear |R| = 7.331835e-03 1 Nonlinear |R| = 3.380724e-01 0 Linear |R| = 3.380724e-01 1 Linear |R| = 3.001736e-01 2 Linear |R| = 2.994000e-01 3 Linear |R| = 2.638543e-01 4 Linear |R| = 3.343074e-03 5 Linear |R| = 6.175358e-05 6 Linear |R| = 2.374424e-07 2 Nonlinear |R| = 1.834137e-01 0 Linear |R| = 1.834137e-01 1 Linear |R| = 1.576530e-01 2 Linear |R| = 4.220904e-02 3 Linear |R| = 4.207755e-04 4 Linear |R| = 2.969888e-10 3 Nonlinear |R| = 1.813251e-08 Solve Converged! Performance Graph: -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | Section | Calls | Self(s) | Avg(s) | % | Mem(MB) | Total(s) | Avg(s) | % | Mem(MB) | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | HeatConductionTestApp (main) | 1 | 0.003 | 0.003 | 0.36 | 2 | 0.904 | 0.904 | 100.00 | 40 | | Action::SetupMeshAction::Mesh::SetupMeshAction::act::setup_mesh | 1 | 0.001 | 0.001 | 0.07 | 0 | 0.001 | 0.001 | 0.07 | 0 | | Action::SetupMeshAction::Mesh::SetupMeshAction::act::set_mesh_base | 2 | 0.000 | 0.000 | 0.00 | 0 | 0.000 | 0.000 | 0.00 | 0 | | MooseApp::executeMeshGenerators | 1 | 0.000 | 0.000 | 0.00 | 0 | 0.000 | 0.000 | 0.00 | 0 | | FEProblem::outputStep | 4 | 0.000 | 0.000 | 0.03 | 0 | 0.016 | 0.004 | 1.82 | 7 | | Exodus::outputStep | 4 | 0.013 | 0.003 | 1.42 | 7 | 0.013 | 0.003 | 1.42 | 7 | | Transient::PicardSolve | 5 | 0.001 | 0.000 | 0.16 | 0 | 0.746 | 0.149 | 82.51 | 3 | | FEProblem::outputStep | 7 | 0.000 | 0.000 | 0.02 | 0 | 0.000 | 0.000 | 0.03 | 0 | | FEProblem::solve | 5 | 0.143 | 0.029 | 15.77 | 3 | 0.735 | 0.147 | 81.28 | 3 | | FEProblem::computeResidualInternal | 5 | 0.000 | 0.000 | 0.01 | 0 | 0.006 | 0.001 | 0.63 | 0 | | FEProblem::computeResidualInternal | 490 | 0.008 | 0.000 | 0.91 | 0 | 0.523 | 0.001 | 57.83 | 0 | | FEProblem::computeJacobianInternal | 24 | 0.000 | 0.000 | 0.04 | 0 | 0.051 | 0.002 | 5.63 | 0 | | FEProblem::computeUserObjects | 3 | 0.008 | 0.003 | 0.83 | 0 | 0.008 | 0.003 | 0.83 | 0 | | Transient::final | 1 | 0.000 | 0.000 | 0.01 | 0 | 0.000 | 0.000 | 0.02 | 0 | | FEProblem::outputStep | 1 | 0.000 | 0.000 | 0.00 | 0 | 0.000 | 0.000 | 0.00 | 0 | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Running exodiff: /data/civet0/build/moose/framework/contrib/exodiff/exodiff -m -F 1e-10 -t 1e-06 /data/civet0/build/moose/modules/heat_conduction/test/tests/truss_heat_conduction/gold/block_w_line_out.e /data/civet0/build/moose/modules/heat_conduction/test/tests/truss_heat_conduction/block_w_line_out.e ERROR: ***************************************************************** EXODIFF (Version: 2.90) Modified: 2018-02-15 Authors: Richard Drake, rrdrake@sandia.gov Greg Sjaardema, gdsjaar@sandia.gov Run on 2022/10/21 00:18:09 MDT ***************************************************************** Reading first file ... Reading second file ... FILE 1: /data/civet0/build/moose/modules/heat_conduction/test/tests/truss_heat_conduction/gold/block_w_line_out.e Title: block_w_line_out.e Dim = 3, Blocks = 2, Nodes = 419, Elements = 160, Nodesets = 8, Sidesets = 8 Vars: Global = 0, Nodal = 1, Element = 1, Nodeset = 0, Sideset = 0, Times = 2 FILE 2: /data/civet0/build/moose/modules/heat_conduction/test/tests/truss_heat_conduction/block_w_line_out.e Title: block_w_line_out.e Dim = 3, Blocks = 2, Nodes = 419, Elements = 160, Nodesets = 8, Sidesets = 8 Vars: Global = 0, Nodal = 1, Element = 1, Nodeset = 0, Sideset = 0, Times = 4 exodiff: ERROR: .. First file has 2 result times while the second file has 4. exodiff: INFO .. Map option is not needed. Nodal coordinates will be compared .. tol: 1e-06 (absolute), floor: 0 Time step values will be compared .. tol: 1e-06 (relative), floor: 1e-10 No Global variables on either file. Nodal variables to be compared: temperature tol: 1e-06 (relative), floor: 1e-10 Element variables to be compared: area tol: 1e-06 (relative), floor: 1e-10 No Element Attribute variables on either file. No Nodeset variables on either file. No Sideset variables on either file. Sideset Distribution Factors will be compared .. tol: 1e-06 (relative), floor: 0 ============================================================== NOTE: All node and element ids are reported as global ids. Sideset Distribution Factors: --------- Time step 1, 0.0000000e+00 ~ 0.0000000e+00, rel diff: 0.00000e+00 --------- Nodal variables: Element variables: --------- Time step 2, 1.0000000e+00 ~ 5.0000000e-01, rel diff: 5.00000e-01 (FAILED) Nodal variables: temperature rel diff: 6.7393323e+00 ~ 2.4136390e+00 = 6.41858e-01 (node 402) Element variables: exodiff: Files are different ################################################################################ Tester failed, reason: EXODIFF ``` ## Steps to Reproduce run that test many times ## Impact lost time
1.0
Unreliable heat conduction truss test in MPI+Threads recipe - ## Bug Description Stochastic test failures on distrbuted + threads ``` Working Directory: /data/civet0/build/moose/modules/heat_conduction/test/tests/truss_heat_conduction Running command: mpiexec -n 2 /data/civet0/build/moose/modules/heat_conduction/heat_conduction-opt -i block_w_line.i --distributed-mesh --error --error-override --timing Outputs/perf_graph=true --no-gdb-backtrace --n-threads=2 Framework Information: MOOSE Version: git commit a737d941ef3 on 2022-10-20 LibMesh Version: 69dad44d8ef2b871ec1a6a5b5788b00385642c39 PETSc Version: 3.16.6 SLEPc Version: 3.16.2 Current Time: Fri Oct 21 00:18:08 2022 Executable Timestamp: Fri Oct 21 00:14:06 2022 Parallelism: Num Processors: 2 Num Threads: 2 Mesh: Parallel Type: replicated (forced) Mesh Dimension: 3 Spatial Dimension: 3 Nodes: Total: 419 Local: 214 Min/Max/Avg: 205/214/209 Elems: Total: 160 Local: 80 Min/Max/Avg: 80/80/80 Num Subdomains: 2 Num Partitions: 2 Partitioner: metis Nonlinear System: Num DOFs: 419 Num Local DOFs: 214 Variables: "temperature" Finite Element Types: "LAGRANGE" Approximation Orders: "FIRST" Auxiliary System: Num DOFs: 160 Num Local DOFs: 80 Variables: "area" Finite Element Types: "MONOMIAL" Approximation Orders: "CONSTANT" Execution Information: Executioner: Transient TimeStepper: ConstantDT Solver Mode: Preconditioned JFNK PETSc Preconditioner: hypre boomeramg strong_threshold: 0.7 (auto) Time Step 0, time = 0 Time Step 1, time = 1, dt = 1 0 Nonlinear |R| = 8.124038e+06 0 Linear |R| = 8.124038e+06 1 Linear |R| = 2.936180e+06 2 Linear |R| = 1.391816e+06 3 Linear |R| = 5.426702e+02 4 Linear |R| = 1.858057e-02 1 Nonlinear |R| = 4.987863e-01 0 Linear |R| = 4.987863e-01 1 Linear |R| = 4.984998e-01 2 Linear |R| = 4.898628e-01 3 Linear |R| = 4.763700e-01 4 Linear |R| = 1.116249e-01 5 Linear |R| = 8.691014e-04 6 Linear |R| = 1.822808e-04 7 Linear |R| = 2.942690e-06 2 Nonlinear |R| = 4.987863e-01 0 Linear |R| = 4.987863e-01 1 Linear |R| = 4.984998e-01 2 Linear |R| = 4.904613e-01 3 Linear |R| = 4.770534e-01 4 Linear |R| = 1.592058e-01 5 Linear |R| = 1.990848e-03 6 Linear |R| = 6.528066e-05 7 Linear |R| = 5.683374e-06 8 Linear |R| = 4.009811e-06 3 Nonlinear |R| = 2.567383e-01 0 Linear |R| = 2.567383e-01 1 Linear |R| = 2.527316e-01 2 Linear |R| = 1.957669e-01 3 Linear |R| = 1.947641e-01 4 Linear |R| = 1.236632e-01 5 Linear |R| = 2.129886e-03 6 Linear |R| = 8.013022e-05 7 Linear |R| = 8.195196e-06 8 Linear |R| = 5.685327e-06 9 Linear |R| = 4.612431e-06 10 Linear |R| = 3.981993e-06 11 Linear |R| = 3.554832e-06 12 Linear |R| = 3.241063e-06 13 Linear |R| = 2.997947e-06 14 Linear |R| = 2.802455e-06 15 Linear |R| = 2.640804e-06 16 Linear |R| = 2.504248e-06 4 Nonlinear |R| = 2.130816e-01 0 Linear |R| = 2.130816e-01 1 Linear |R| = 2.127822e-01 2 Linear |R| = 1.951439e-01 3 Linear |R| = 1.950112e-01 4 Linear |R| = 5.776265e-02 5 Linear |R| = 5.586006e-04 6 Linear |R| = 1.314725e-05 7 Linear |R| = 1.214406e-06 5 Nonlinear |R| = 2.112063e-01 0 Linear |R| = 2.112063e-01 1 Linear |R| = 2.104202e-01 2 Linear |R| = 1.910654e-01 3 Linear |R| = 1.906017e-01 4 Linear |R| = 9.825074e-02 5 Linear |R| = 1.446873e-03 6 Linear |R| = 2.038308e-05 7 Linear |R| = 3.983396e-07 6 Nonlinear |R| = 1.993684e-01 0 Linear |R| = 1.993684e-01 1 Linear |R| = 1.940511e-01 2 Linear |R| = 1.804851e-01 3 Linear |R| = 1.655430e-01 4 Linear |R| = 6.593206e-02 5 Linear |R| = 4.145496e-03 6 Linear |R| = 1.476153e-04 7 Linear |R| = 1.368834e-06 7 Nonlinear |R| = 1.993684e-01 0 Linear |R| = 1.993684e-01 1 Linear |R| = 1.940510e-01 2 Linear |R| = 1.821076e-01 3 Linear |R| = 1.676569e-01 4 Linear |R| = 1.486711e-01 5 Linear |R| = 2.073761e-03 6 Linear |R| = 7.024214e-05 7 Linear |R| = 2.797172e-06 8 Linear |R| = 1.987650e-06 8 Nonlinear |R| = 1.832074e-01 0 Linear |R| = 1.832074e-01 1 Linear |R| = 1.747611e-01 2 Linear |R| = 1.388639e-01 3 Linear |R| = 1.321860e-01 4 Linear |R| = 9.827800e-02 5 Linear |R| = 3.621897e-03 6 Linear |R| = 7.397675e-05 7 Linear |R| = 1.292977e-05 8 Linear |R| = 8.937240e-06 9 Linear |R| = 7.237751e-06 10 Linear |R| = 6.242950e-06 11 Linear |R| = 5.570493e-06 12 Linear |R| = 5.077053e-06 13 Linear |R| = 4.695115e-06 14 Linear |R| = 4.388160e-06 15 Linear |R| = 4.134479e-06 16 Linear |R| = 3.920247e-06 17 Linear |R| = 3.736199e-06 18 Linear |R| = 3.575857e-06 19 Linear |R| = 3.434531e-06 20 Linear |R| = 3.308739e-06 21 Linear |R| = 3.195827e-06 22 Linear |R| = 3.093738e-06 23 Linear |R| = 3.000846e-06 24 Linear |R| = 2.915848e-06 25 Linear |R| = 2.837687e-06 26 Linear |R| = 2.765492e-06 27 Linear |R| = 2.698540e-06 28 Linear |R| = 2.636228e-06 29 Linear |R| = 2.578041e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 9 Nonlinear |R| = 1.832074e-01 0 Linear |R| = 1.832074e-01 1 Linear |R| = 1.747611e-01 2 Linear |R| = 1.402027e-01 3 Linear |R| = 1.321711e-01 4 Linear |R| = 1.215777e-01 5 Linear |R| = 5.027232e-03 6 Linear |R| = 1.297047e-04 7 Linear |R| = 3.307898e-05 8 Linear |R| = 2.397256e-05 9 Linear |R| = 1.973076e-05 10 Linear |R| = 1.715673e-05 11 Linear |R| = 1.538315e-05 12 Linear |R| = 1.406595e-05 13 Linear |R| = 1.303788e-05 14 Linear |R| = 1.220676e-05 15 Linear |R| = 1.151661e-05 16 Linear |R| = 1.093176e-05 17 Linear |R| = 1.042776e-05 18 Linear |R| = 9.987631e-06 19 Linear |R| = 9.598842e-06 20 Linear |R| = 9.252230e-06 21 Linear |R| = 8.940646e-06 22 Linear |R| = 8.658564e-06 23 Linear |R| = 8.401597e-06 24 Linear |R| = 8.166232e-06 25 Linear |R| = 7.949599e-06 26 Linear |R| = 7.749337e-06 27 Linear |R| = 7.563487e-06 28 Linear |R| = 7.390402e-06 29 Linear |R| = 7.228673e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 Nonlinear solve did not converge due to DIVERGED_LINE_SEARCH iterations 9 Solve Did NOT Converge! Aborting as solve did not converge Time Step 1, time = 0.5, dt = 0.5 0 Nonlinear |R| = 4.062019e+06 0 Linear |R| = 4.062019e+06 1 Linear |R| = 1.408872e+06 2 Linear |R| = 5.767703e+05 3 Linear |R| = 1.337262e+03 4 Linear |R| = 3.477387e-03 1 Nonlinear |R| = 6.662115e-02 0 Linear |R| = 6.662115e-02 1 Linear |R| = 6.661958e-02 2 Linear |R| = 6.661213e-02 3 Linear |R| = 6.651483e-02 4 Linear |R| = 4.531902e-02 5 Linear |R| = 1.589527e-04 6 Linear |R| = 6.840782e-06 7 Linear |R| = 2.365846e-06 8 Linear |R| = 1.703323e-06 9 Linear |R| = 1.399136e-06 10 Linear |R| = 1.215352e-06 11 Linear |R| = 1.089037e-06 12 Linear |R| = 9.953671e-07 13 Linear |R| = 9.223387e-07 14 Linear |R| = 8.633339e-07 15 Linear |R| = 8.143765e-07 16 Linear |R| = 7.729024e-07 17 Linear |R| = 7.371806e-07 18 Linear |R| = 7.059931e-07 19 Linear |R| = 6.784570e-07 20 Linear |R| = 6.539097e-07 2 Nonlinear |R| = 4.354222e-02 0 Linear |R| = 4.354222e-02 1 Linear |R| = 4.261640e-02 2 Linear |R| = 4.259191e-02 3 Linear |R| = 4.191010e-02 4 Linear |R| = 4.166622e-02 5 Linear |R| = 4.131442e-04 6 Linear |R| = 4.163222e-05 7 Linear |R| = 2.060248e-05 8 Linear |R| = 1.596644e-05 9 Linear |R| = 1.346830e-05 10 Linear |R| = 1.186748e-05 11 Linear |R| = 1.072727e-05 12 Linear |R| = 9.864290e-06 13 Linear |R| = 9.180288e-06 14 Linear |R| = 8.621530e-06 15 Linear |R| = 8.153597e-06 16 Linear |R| = 7.754596e-06 17 Linear |R| = 7.408844e-06 18 Linear |R| = 7.105672e-06 19 Linear |R| = 6.836765e-06 20 Linear |R| = 6.596315e-06 21 Linear |R| = 6.379559e-06 22 Linear |R| = 6.182856e-06 23 Linear |R| = 6.003290e-06 24 Linear |R| = 5.838525e-06 25 Linear |R| = 5.686607e-06 26 Linear |R| = 5.546020e-06 27 Linear |R| = 5.415313e-06 28 Linear |R| = 5.293485e-06 29 Linear |R| = 5.179491e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 3 Nonlinear |R| = 4.261368e-02 0 Linear |R| = 4.261368e-02 1 Linear |R| = 4.210921e-02 2 Linear |R| = 4.162688e-02 3 Linear |R| = 3.903726e-02 4 Linear |R| = 2.703674e-02 5 Linear |R| = 1.452146e-03 6 Linear |R| = 1.293879e-04 7 Linear |R| = 4.296561e-05 8 Linear |R| = 3.191541e-05 9 Linear |R| = 2.650371e-05 10 Linear |R| = 2.315369e-05 11 Linear |R| = 2.081924e-05 12 Linear |R| = 1.907311e-05 13 Linear |R| = 1.770351e-05 14 Linear |R| = 1.659209e-05 15 Linear |R| = 1.566671e-05 16 Linear |R| = 1.488070e-05 17 Linear |R| = 1.420223e-05 18 Linear |R| = 1.360884e-05 19 Linear |R| = 1.308411e-05 20 Linear |R| = 1.261575e-05 21 Linear |R| = 1.219433e-05 22 Linear |R| = 1.181250e-05 23 Linear |R| = 1.146443e-05 24 Linear |R| = 1.114543e-05 25 Linear |R| = 1.085164e-05 26 Linear |R| = 1.057994e-05 27 Linear |R| = 1.032766e-05 28 Linear |R| = 1.009262e-05 29 Linear |R| = 9.872924e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 4 Nonlinear |R| = 3.713079e-02 Solve Converged! Time Step 2, time = 1, dt = 0.5 0 Nonlinear |R| = 4.062019e+06 0 Linear |R| = 4.062019e+06 1 Linear |R| = 1.519291e+06 2 Linear |R| = 7.680909e+05 3 Linear |R| = 5.089047e+02 4 Linear |R| = 2.470842e-01 1 Nonlinear |R| = 2.320993e-01 0 Linear |R| = 2.320993e-01 1 Linear |R| = 2.316175e-01 2 Linear |R| = 2.097271e-01 3 Linear |R| = 1.765354e-01 4 Linear |R| = 1.686036e-01 5 Linear |R| = 9.050986e-02 6 Linear |R| = 2.253444e-02 7 Linear |R| = 1.103272e-03 8 Linear |R| = 6.923692e-04 9 Linear |R| = 5.537875e-04 10 Linear |R| = 4.737331e-04 11 Linear |R| = 4.208348e-04 12 Linear |R| = 3.823479e-04 13 Linear |R| = 3.528327e-04 14 Linear |R| = 3.292417e-04 15 Linear |R| = 3.098151e-04 16 Linear |R| = 2.934741e-04 17 Linear |R| = 2.794668e-04 18 Linear |R| = 2.672972e-04 19 Linear |R| = 2.565894e-04 20 Linear |R| = 2.470732e-04 21 Linear |R| = 2.385429e-04 22 Linear |R| = 2.308397e-04 23 Linear |R| = 2.238356e-04 24 Linear |R| = 2.174344e-04 25 Linear |R| = 2.115512e-04 26 Linear |R| = 2.061221e-04 27 Linear |R| = 2.010907e-04 28 Linear |R| = 1.964105e-04 29 Linear |R| = 1.920427e-04 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 2 Nonlinear |R| = 2.320993e-01 0 Linear |R| = 2.320993e-01 1 Linear |R| = 2.316175e-01 2 Linear |R| = 2.096388e-01 3 Linear |R| = 1.856877e-01 4 Linear |R| = 1.841090e-01 5 Linear |R| = 1.931490e-02 6 Linear |R| = 1.679579e-02 7 Linear |R| = 2.409159e-03 8 Linear |R| = 1.728413e-03 9 Linear |R| = 1.419190e-03 10 Linear |R| = 1.232370e-03 11 Linear |R| = 1.104061e-03 12 Linear |R| = 1.008943e-03 13 Linear |R| = 9.348332e-04 14 Linear |R| = 8.749601e-04 15 Linear |R| = 8.252987e-04 16 Linear |R| = 7.832360e-04 17 Linear |R| = 7.470093e-04 18 Linear |R| = 7.153860e-04 19 Linear |R| = 6.874650e-04 20 Linear |R| = 6.625766e-04 21 Linear |R| = 6.402134e-04 22 Linear |R| = 6.199684e-04 23 Linear |R| = 6.015305e-04 24 Linear |R| = 5.846441e-04 25 Linear |R| = 5.691063e-04 26 Linear |R| = 5.547439e-04 27 Linear |R| = 5.414172e-04 28 Linear |R| = 5.290064e-04 29 Linear |R| = 5.174122e-04 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 3 Nonlinear |R| = 2.320870e-01 0 Linear |R| = 2.320870e-01 1 Linear |R| = 2.316584e-01 2 Linear |R| = 2.097528e-01 3 Linear |R| = 2.030920e-01 4 Linear |R| = 1.834407e-01 5 Linear |R| = 1.819411e-01 6 Linear |R| = 8.400089e-03 7 Linear |R| = 8.859178e-04 8 Linear |R| = 6.315971e-04 9 Linear |R| = 5.176682e-04 10 Linear |R| = 4.490556e-04 11 Linear |R| = 4.021027e-04 12 Linear |R| = 3.673259e-04 13 Linear |R| = 3.402699e-04 14 Linear |R| = 3.184134e-04 15 Linear |R| = 3.003021e-04 16 Linear |R| = 2.849593e-04 17 Linear |R| = 2.717568e-04 18 Linear |R| = 2.602292e-04 19 Linear |R| = 2.500587e-04 20 Linear |R| = 2.409907e-04 21 Linear |R| = 2.328458e-04 22 Linear |R| = 2.254719e-04 23 Linear |R| = 2.187591e-04 24 Linear |R| = 2.126104e-04 25 Linear |R| = 2.069542e-04 26 Linear |R| = 2.017253e-04 27 Linear |R| = 1.968749e-04 28 Linear |R| = 1.923573e-04 29 Linear |R| = 1.881380e-04 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 Nonlinear solve did not converge due to DIVERGED_LINE_SEARCH iterations 3 Solve Did NOT Converge! Aborting as solve did not converge Time Step 2, time = 0.75, dt = 0.25 0 Nonlinear |R| = 2.031010e+06 0 Linear |R| = 2.031010e+06 1 Linear |R| = 7.460102e+05 2 Linear |R| = 3.531009e+05 3 Linear |R| = 6.736684e+03 4 Linear |R| = 6.661378e-03 1 Nonlinear |R| = 3.293862e-01 0 Linear |R| = 3.293862e-01 1 Linear |R| = 2.214630e-01 2 Linear |R| = 2.130098e-01 3 Linear |R| = 2.130038e-01 4 Linear |R| = 4.081695e-03 5 Linear |R| = 1.601902e-05 6 Linear |R| = 2.080131e-06 2 Nonlinear |R| = 2.395315e-01 0 Linear |R| = 2.395315e-01 1 Linear |R| = 1.214649e-01 2 Linear |R| = 2.149745e-02 3 Linear |R| = 2.090277e-04 4 Linear |R| = 2.536993e-10 3 Nonlinear |R| = 5.566991e-09 Solve Converged! Time Step 3, time = 1, dt = 0.25 0 Nonlinear |R| = 2.031010e+06 0 Linear |R| = 2.031010e+06 1 Linear |R| = 7.682209e+05 2 Linear |R| = 3.924322e+05 3 Linear |R| = 6.833789e+03 4 Linear |R| = 7.331835e-03 1 Nonlinear |R| = 3.380724e-01 0 Linear |R| = 3.380724e-01 1 Linear |R| = 3.001736e-01 2 Linear |R| = 2.994000e-01 3 Linear |R| = 2.638543e-01 4 Linear |R| = 3.343074e-03 5 Linear |R| = 6.175358e-05 6 Linear |R| = 2.374424e-07 2 Nonlinear |R| = 1.834137e-01 0 Linear |R| = 1.834137e-01 1 Linear |R| = 1.576530e-01 2 Linear |R| = 4.220904e-02 3 Linear |R| = 4.207755e-04 4 Linear |R| = 2.969888e-10 3 Nonlinear |R| = 1.813251e-08 Solve Converged! Performance Graph: -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | Section | Calls | Self(s) | Avg(s) | % | Mem(MB) | Total(s) | Avg(s) | % | Mem(MB) | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | HeatConductionTestApp (main) | 1 | 0.003 | 0.003 | 0.36 | 2 | 0.904 | 0.904 | 100.00 | 40 | | Action::SetupMeshAction::Mesh::SetupMeshAction::act::setup_mesh | 1 | 0.001 | 0.001 | 0.07 | 0 | 0.001 | 0.001 | 0.07 | 0 | | Action::SetupMeshAction::Mesh::SetupMeshAction::act::set_mesh_base | 2 | 0.000 | 0.000 | 0.00 | 0 | 0.000 | 0.000 | 0.00 | 0 | | MooseApp::executeMeshGenerators | 1 | 0.000 | 0.000 | 0.00 | 0 | 0.000 | 0.000 | 0.00 | 0 | | FEProblem::outputStep | 4 | 0.000 | 0.000 | 0.03 | 0 | 0.016 | 0.004 | 1.82 | 7 | | Exodus::outputStep | 4 | 0.013 | 0.003 | 1.42 | 7 | 0.013 | 0.003 | 1.42 | 7 | | Transient::PicardSolve | 5 | 0.001 | 0.000 | 0.16 | 0 | 0.746 | 0.149 | 82.51 | 3 | | FEProblem::outputStep | 7 | 0.000 | 0.000 | 0.02 | 0 | 0.000 | 0.000 | 0.03 | 0 | | FEProblem::solve | 5 | 0.143 | 0.029 | 15.77 | 3 | 0.735 | 0.147 | 81.28 | 3 | | FEProblem::computeResidualInternal | 5 | 0.000 | 0.000 | 0.01 | 0 | 0.006 | 0.001 | 0.63 | 0 | | FEProblem::computeResidualInternal | 490 | 0.008 | 0.000 | 0.91 | 0 | 0.523 | 0.001 | 57.83 | 0 | | FEProblem::computeJacobianInternal | 24 | 0.000 | 0.000 | 0.04 | 0 | 0.051 | 0.002 | 5.63 | 0 | | FEProblem::computeUserObjects | 3 | 0.008 | 0.003 | 0.83 | 0 | 0.008 | 0.003 | 0.83 | 0 | | Transient::final | 1 | 0.000 | 0.000 | 0.01 | 0 | 0.000 | 0.000 | 0.02 | 0 | | FEProblem::outputStep | 1 | 0.000 | 0.000 | 0.00 | 0 | 0.000 | 0.000 | 0.00 | 0 | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Framework Information: MOOSE Version: git commit a737d941ef3 on 2022-10-20 LibMesh Version: 69dad44d8ef2b871ec1a6a5b5788b00385642c39 PETSc Version: 3.16.6 SLEPc Version: 3.16.2 Current Time: Fri Oct 21 00:18:08 2022 Executable Timestamp: Fri Oct 21 00:14:06 2022 Parallelism: Num Processors: 2 Num Threads: 2 Mesh: Parallel Type: replicated (forced) Mesh Dimension: 3 Spatial Dimension: 3 Nodes: Total: 419 Local: 214 Min/Max/Avg: 205/214/209 Elems: Total: 160 Local: 80 Min/Max/Avg: 80/80/80 Num Subdomains: 2 Num Partitions: 2 Partitioner: metis Nonlinear System: Num DOFs: 419 Num Local DOFs: 214 Variables: "temperature" Finite Element Types: "LAGRANGE" Approximation Orders: "FIRST" Auxiliary System: Num DOFs: 160 Num Local DOFs: 80 Variables: "area" Finite Element Types: "MONOMIAL" Approximation Orders: "CONSTANT" Execution Information: Executioner: Transient TimeStepper: ConstantDT Solver Mode: Preconditioned JFNK PETSc Preconditioner: hypre boomeramg strong_threshold: 0.7 (auto) Time Step 0, time = 0 Time Step 1, time = 1, dt = 1 0 Nonlinear |R| = 8.124038e+06 0 Linear |R| = 8.124038e+06 1 Linear |R| = 2.936180e+06 2 Linear |R| = 1.391816e+06 3 Linear |R| = 5.426702e+02 4 Linear |R| = 1.858057e-02 1 Nonlinear |R| = 4.987863e-01 0 Linear |R| = 4.987863e-01 1 Linear |R| = 4.984998e-01 2 Linear |R| = 4.898628e-01 3 Linear |R| = 4.763700e-01 4 Linear |R| = 1.116249e-01 5 Linear |R| = 8.691014e-04 6 Linear |R| = 1.822808e-04 7 Linear |R| = 2.942690e-06 2 Nonlinear |R| = 4.987863e-01 0 Linear |R| = 4.987863e-01 1 Linear |R| = 4.984998e-01 2 Linear |R| = 4.904613e-01 3 Linear |R| = 4.770534e-01 4 Linear |R| = 1.592058e-01 5 Linear |R| = 1.990848e-03 6 Linear |R| = 6.528066e-05 7 Linear |R| = 5.683374e-06 8 Linear |R| = 4.009811e-06 3 Nonlinear |R| = 2.567383e-01 0 Linear |R| = 2.567383e-01 1 Linear |R| = 2.527316e-01 2 Linear |R| = 1.957669e-01 3 Linear |R| = 1.947641e-01 4 Linear |R| = 1.236632e-01 5 Linear |R| = 2.129886e-03 6 Linear |R| = 8.013022e-05 7 Linear |R| = 8.195196e-06 8 Linear |R| = 5.685327e-06 9 Linear |R| = 4.612431e-06 10 Linear |R| = 3.981993e-06 11 Linear |R| = 3.554832e-06 12 Linear |R| = 3.241063e-06 13 Linear |R| = 2.997947e-06 14 Linear |R| = 2.802455e-06 15 Linear |R| = 2.640804e-06 16 Linear |R| = 2.504248e-06 4 Nonlinear |R| = 2.130816e-01 0 Linear |R| = 2.130816e-01 1 Linear |R| = 2.127822e-01 2 Linear |R| = 1.951439e-01 3 Linear |R| = 1.950112e-01 4 Linear |R| = 5.776265e-02 5 Linear |R| = 5.586006e-04 6 Linear |R| = 1.314725e-05 7 Linear |R| = 1.214406e-06 5 Nonlinear |R| = 2.112063e-01 0 Linear |R| = 2.112063e-01 1 Linear |R| = 2.104202e-01 2 Linear |R| = 1.910654e-01 3 Linear |R| = 1.906017e-01 4 Linear |R| = 9.825074e-02 5 Linear |R| = 1.446873e-03 6 Linear |R| = 2.038308e-05 7 Linear |R| = 3.983396e-07 6 Nonlinear |R| = 1.993684e-01 0 Linear |R| = 1.993684e-01 1 Linear |R| = 1.940511e-01 2 Linear |R| = 1.804851e-01 3 Linear |R| = 1.655430e-01 4 Linear |R| = 6.593206e-02 5 Linear |R| = 4.145496e-03 6 Linear |R| = 1.476153e-04 7 Linear |R| = 1.368834e-06 7 Nonlinear |R| = 1.993684e-01 0 Linear |R| = 1.993684e-01 1 Linear |R| = 1.940510e-01 2 Linear |R| = 1.821076e-01 3 Linear |R| = 1.676569e-01 4 Linear |R| = 1.486711e-01 5 Linear |R| = 2.073761e-03 6 Linear |R| = 7.024214e-05 7 Linear |R| = 2.797172e-06 8 Linear |R| = 1.987650e-06 8 Nonlinear |R| = 1.832074e-01 0 Linear |R| = 1.832074e-01 1 Linear |R| = 1.747611e-01 2 Linear |R| = 1.388639e-01 3 Linear |R| = 1.321860e-01 4 Linear |R| = 9.827800e-02 5 Linear |R| = 3.621897e-03 6 Linear |R| = 7.397675e-05 7 Linear |R| = 1.292977e-05 8 Linear |R| = 8.937240e-06 9 Linear |R| = 7.237751e-06 10 Linear |R| = 6.242950e-06 11 Linear |R| = 5.570493e-06 12 Linear |R| = 5.077053e-06 13 Linear |R| = 4.695115e-06 14 Linear |R| = 4.388160e-06 15 Linear |R| = 4.134479e-06 16 Linear |R| = 3.920247e-06 17 Linear |R| = 3.736199e-06 18 Linear |R| = 3.575857e-06 19 Linear |R| = 3.434531e-06 20 Linear |R| = 3.308739e-06 21 Linear |R| = 3.195827e-06 22 Linear |R| = 3.093738e-06 23 Linear |R| = 3.000846e-06 24 Linear |R| = 2.915848e-06 25 Linear |R| = 2.837687e-06 26 Linear |R| = 2.765492e-06 27 Linear |R| = 2.698540e-06 28 Linear |R| = 2.636228e-06 29 Linear |R| = 2.578041e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 9 Nonlinear |R| = 1.832074e-01 0 Linear |R| = 1.832074e-01 1 Linear |R| = 1.747611e-01 2 Linear |R| = 1.402027e-01 3 Linear |R| = 1.321711e-01 4 Linear |R| = 1.215777e-01 5 Linear |R| = 5.027232e-03 6 Linear |R| = 1.297047e-04 7 Linear |R| = 3.307898e-05 8 Linear |R| = 2.397256e-05 9 Linear |R| = 1.973076e-05 10 Linear |R| = 1.715673e-05 11 Linear |R| = 1.538315e-05 12 Linear |R| = 1.406595e-05 13 Linear |R| = 1.303788e-05 14 Linear |R| = 1.220676e-05 15 Linear |R| = 1.151661e-05 16 Linear |R| = 1.093176e-05 17 Linear |R| = 1.042776e-05 18 Linear |R| = 9.987631e-06 19 Linear |R| = 9.598842e-06 20 Linear |R| = 9.252230e-06 21 Linear |R| = 8.940646e-06 22 Linear |R| = 8.658564e-06 23 Linear |R| = 8.401597e-06 24 Linear |R| = 8.166232e-06 25 Linear |R| = 7.949599e-06 26 Linear |R| = 7.749337e-06 27 Linear |R| = 7.563487e-06 28 Linear |R| = 7.390402e-06 29 Linear |R| = 7.228673e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 Nonlinear solve did not converge due to DIVERGED_LINE_SEARCH iterations 9 Solve Did NOT Converge! Aborting as solve did not converge Time Step 1, time = 0.5, dt = 0.5 0 Nonlinear |R| = 4.062019e+06 0 Linear |R| = 4.062019e+06 1 Linear |R| = 1.408872e+06 2 Linear |R| = 5.767703e+05 3 Linear |R| = 1.337262e+03 4 Linear |R| = 3.477387e-03 1 Nonlinear |R| = 6.662115e-02 0 Linear |R| = 6.662115e-02 1 Linear |R| = 6.661958e-02 2 Linear |R| = 6.661213e-02 3 Linear |R| = 6.651483e-02 4 Linear |R| = 4.531902e-02 5 Linear |R| = 1.589527e-04 6 Linear |R| = 6.840782e-06 7 Linear |R| = 2.365846e-06 8 Linear |R| = 1.703323e-06 9 Linear |R| = 1.399136e-06 10 Linear |R| = 1.215352e-06 11 Linear |R| = 1.089037e-06 12 Linear |R| = 9.953671e-07 13 Linear |R| = 9.223387e-07 14 Linear |R| = 8.633339e-07 15 Linear |R| = 8.143765e-07 16 Linear |R| = 7.729024e-07 17 Linear |R| = 7.371806e-07 18 Linear |R| = 7.059931e-07 19 Linear |R| = 6.784570e-07 20 Linear |R| = 6.539097e-07 2 Nonlinear |R| = 4.354222e-02 0 Linear |R| = 4.354222e-02 1 Linear |R| = 4.261640e-02 2 Linear |R| = 4.259191e-02 3 Linear |R| = 4.191010e-02 4 Linear |R| = 4.166622e-02 5 Linear |R| = 4.131442e-04 6 Linear |R| = 4.163222e-05 7 Linear |R| = 2.060248e-05 8 Linear |R| = 1.596644e-05 9 Linear |R| = 1.346830e-05 10 Linear |R| = 1.186748e-05 11 Linear |R| = 1.072727e-05 12 Linear |R| = 9.864290e-06 13 Linear |R| = 9.180288e-06 14 Linear |R| = 8.621530e-06 15 Linear |R| = 8.153597e-06 16 Linear |R| = 7.754596e-06 17 Linear |R| = 7.408844e-06 18 Linear |R| = 7.105672e-06 19 Linear |R| = 6.836765e-06 20 Linear |R| = 6.596315e-06 21 Linear |R| = 6.379559e-06 22 Linear |R| = 6.182856e-06 23 Linear |R| = 6.003290e-06 24 Linear |R| = 5.838525e-06 25 Linear |R| = 5.686607e-06 26 Linear |R| = 5.546020e-06 27 Linear |R| = 5.415313e-06 28 Linear |R| = 5.293485e-06 29 Linear |R| = 5.179491e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 3 Nonlinear |R| = 4.261368e-02 0 Linear |R| = 4.261368e-02 1 Linear |R| = 4.210921e-02 2 Linear |R| = 4.162688e-02 3 Linear |R| = 3.903726e-02 4 Linear |R| = 2.703674e-02 5 Linear |R| = 1.452146e-03 6 Linear |R| = 1.293879e-04 7 Linear |R| = 4.296561e-05 8 Linear |R| = 3.191541e-05 9 Linear |R| = 2.650371e-05 10 Linear |R| = 2.315369e-05 11 Linear |R| = 2.081924e-05 12 Linear |R| = 1.907311e-05 13 Linear |R| = 1.770351e-05 14 Linear |R| = 1.659209e-05 15 Linear |R| = 1.566671e-05 16 Linear |R| = 1.488070e-05 17 Linear |R| = 1.420223e-05 18 Linear |R| = 1.360884e-05 19 Linear |R| = 1.308411e-05 20 Linear |R| = 1.261575e-05 21 Linear |R| = 1.219433e-05 22 Linear |R| = 1.181250e-05 23 Linear |R| = 1.146443e-05 24 Linear |R| = 1.114543e-05 25 Linear |R| = 1.085164e-05 26 Linear |R| = 1.057994e-05 27 Linear |R| = 1.032766e-05 28 Linear |R| = 1.009262e-05 29 Linear |R| = 9.872924e-06 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 4 Nonlinear |R| = 3.713079e-02 Solve Converged! Time Step 2, time = 1, dt = 0.5 0 Nonlinear |R| = 4.062019e+06 0 Linear |R| = 4.062019e+06 1 Linear |R| = 1.519291e+06 2 Linear |R| = 7.680909e+05 3 Linear |R| = 5.089047e+02 4 Linear |R| = 2.470842e-01 1 Nonlinear |R| = 2.320993e-01 0 Linear |R| = 2.320993e-01 1 Linear |R| = 2.316175e-01 2 Linear |R| = 2.097271e-01 3 Linear |R| = 1.765354e-01 4 Linear |R| = 1.686036e-01 5 Linear |R| = 9.050986e-02 6 Linear |R| = 2.253444e-02 7 Linear |R| = 1.103272e-03 8 Linear |R| = 6.923692e-04 9 Linear |R| = 5.537875e-04 10 Linear |R| = 4.737331e-04 11 Linear |R| = 4.208348e-04 12 Linear |R| = 3.823479e-04 13 Linear |R| = 3.528327e-04 14 Linear |R| = 3.292417e-04 15 Linear |R| = 3.098151e-04 16 Linear |R| = 2.934741e-04 17 Linear |R| = 2.794668e-04 18 Linear |R| = 2.672972e-04 19 Linear |R| = 2.565894e-04 20 Linear |R| = 2.470732e-04 21 Linear |R| = 2.385429e-04 22 Linear |R| = 2.308397e-04 23 Linear |R| = 2.238356e-04 24 Linear |R| = 2.174344e-04 25 Linear |R| = 2.115512e-04 26 Linear |R| = 2.061221e-04 27 Linear |R| = 2.010907e-04 28 Linear |R| = 1.964105e-04 29 Linear |R| = 1.920427e-04 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 2 Nonlinear |R| = 2.320993e-01 0 Linear |R| = 2.320993e-01 1 Linear |R| = 2.316175e-01 2 Linear |R| = 2.096388e-01 3 Linear |R| = 1.856877e-01 4 Linear |R| = 1.841090e-01 5 Linear |R| = 1.931490e-02 6 Linear |R| = 1.679579e-02 7 Linear |R| = 2.409159e-03 8 Linear |R| = 1.728413e-03 9 Linear |R| = 1.419190e-03 10 Linear |R| = 1.232370e-03 11 Linear |R| = 1.104061e-03 12 Linear |R| = 1.008943e-03 13 Linear |R| = 9.348332e-04 14 Linear |R| = 8.749601e-04 15 Linear |R| = 8.252987e-04 16 Linear |R| = 7.832360e-04 17 Linear |R| = 7.470093e-04 18 Linear |R| = 7.153860e-04 19 Linear |R| = 6.874650e-04 20 Linear |R| = 6.625766e-04 21 Linear |R| = 6.402134e-04 22 Linear |R| = 6.199684e-04 23 Linear |R| = 6.015305e-04 24 Linear |R| = 5.846441e-04 25 Linear |R| = 5.691063e-04 26 Linear |R| = 5.547439e-04 27 Linear |R| = 5.414172e-04 28 Linear |R| = 5.290064e-04 29 Linear |R| = 5.174122e-04 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 3 Nonlinear |R| = 2.320870e-01 0 Linear |R| = 2.320870e-01 1 Linear |R| = 2.316584e-01 2 Linear |R| = 2.097528e-01 3 Linear |R| = 2.030920e-01 4 Linear |R| = 1.834407e-01 5 Linear |R| = 1.819411e-01 6 Linear |R| = 8.400089e-03 7 Linear |R| = 8.859178e-04 8 Linear |R| = 6.315971e-04 9 Linear |R| = 5.176682e-04 10 Linear |R| = 4.490556e-04 11 Linear |R| = 4.021027e-04 12 Linear |R| = 3.673259e-04 13 Linear |R| = 3.402699e-04 14 Linear |R| = 3.184134e-04 15 Linear |R| = 3.003021e-04 16 Linear |R| = 2.849593e-04 17 Linear |R| = 2.717568e-04 18 Linear |R| = 2.602292e-04 19 Linear |R| = 2.500587e-04 20 Linear |R| = 2.409907e-04 21 Linear |R| = 2.328458e-04 22 Linear |R| = 2.254719e-04 23 Linear |R| = 2.187591e-04 24 Linear |R| = 2.126104e-04 25 Linear |R| = 2.069542e-04 26 Linear |R| = 2.017253e-04 27 Linear |R| = 1.968749e-04 28 Linear |R| = 1.923573e-04 29 Linear |R| = 1.881380e-04 Linear solve did not converge due to DIVERGED_BREAKDOWN iterations 30 Nonlinear solve did not converge due to DIVERGED_LINE_SEARCH iterations 3 Solve Did NOT Converge! Aborting as solve did not converge Time Step 2, time = 0.75, dt = 0.25 0 Nonlinear |R| = 2.031010e+06 0 Linear |R| = 2.031010e+06 1 Linear |R| = 7.460102e+05 2 Linear |R| = 3.531009e+05 3 Linear |R| = 6.736684e+03 4 Linear |R| = 6.661378e-03 1 Nonlinear |R| = 3.293862e-01 0 Linear |R| = 3.293862e-01 1 Linear |R| = 2.214630e-01 2 Linear |R| = 2.130098e-01 3 Linear |R| = 2.130038e-01 4 Linear |R| = 4.081695e-03 5 Linear |R| = 1.601902e-05 6 Linear |R| = 2.080131e-06 2 Nonlinear |R| = 2.395315e-01 0 Linear |R| = 2.395315e-01 1 Linear |R| = 1.214649e-01 2 Linear |R| = 2.149745e-02 3 Linear |R| = 2.090277e-04 4 Linear |R| = 2.536993e-10 3 Nonlinear |R| = 5.566991e-09 Solve Converged! Time Step 3, time = 1, dt = 0.25 0 Nonlinear |R| = 2.031010e+06 0 Linear |R| = 2.031010e+06 1 Linear |R| = 7.682209e+05 2 Linear |R| = 3.924322e+05 3 Linear |R| = 6.833789e+03 4 Linear |R| = 7.331835e-03 1 Nonlinear |R| = 3.380724e-01 0 Linear |R| = 3.380724e-01 1 Linear |R| = 3.001736e-01 2 Linear |R| = 2.994000e-01 3 Linear |R| = 2.638543e-01 4 Linear |R| = 3.343074e-03 5 Linear |R| = 6.175358e-05 6 Linear |R| = 2.374424e-07 2 Nonlinear |R| = 1.834137e-01 0 Linear |R| = 1.834137e-01 1 Linear |R| = 1.576530e-01 2 Linear |R| = 4.220904e-02 3 Linear |R| = 4.207755e-04 4 Linear |R| = 2.969888e-10 3 Nonlinear |R| = 1.813251e-08 Solve Converged! Performance Graph: -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | Section | Calls | Self(s) | Avg(s) | % | Mem(MB) | Total(s) | Avg(s) | % | Mem(MB) | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | HeatConductionTestApp (main) | 1 | 0.003 | 0.003 | 0.36 | 2 | 0.904 | 0.904 | 100.00 | 40 | | Action::SetupMeshAction::Mesh::SetupMeshAction::act::setup_mesh | 1 | 0.001 | 0.001 | 0.07 | 0 | 0.001 | 0.001 | 0.07 | 0 | | Action::SetupMeshAction::Mesh::SetupMeshAction::act::set_mesh_base | 2 | 0.000 | 0.000 | 0.00 | 0 | 0.000 | 0.000 | 0.00 | 0 | | MooseApp::executeMeshGenerators | 1 | 0.000 | 0.000 | 0.00 | 0 | 0.000 | 0.000 | 0.00 | 0 | | FEProblem::outputStep | 4 | 0.000 | 0.000 | 0.03 | 0 | 0.016 | 0.004 | 1.82 | 7 | | Exodus::outputStep | 4 | 0.013 | 0.003 | 1.42 | 7 | 0.013 | 0.003 | 1.42 | 7 | | Transient::PicardSolve | 5 | 0.001 | 0.000 | 0.16 | 0 | 0.746 | 0.149 | 82.51 | 3 | | FEProblem::outputStep | 7 | 0.000 | 0.000 | 0.02 | 0 | 0.000 | 0.000 | 0.03 | 0 | | FEProblem::solve | 5 | 0.143 | 0.029 | 15.77 | 3 | 0.735 | 0.147 | 81.28 | 3 | | FEProblem::computeResidualInternal | 5 | 0.000 | 0.000 | 0.01 | 0 | 0.006 | 0.001 | 0.63 | 0 | | FEProblem::computeResidualInternal | 490 | 0.008 | 0.000 | 0.91 | 0 | 0.523 | 0.001 | 57.83 | 0 | | FEProblem::computeJacobianInternal | 24 | 0.000 | 0.000 | 0.04 | 0 | 0.051 | 0.002 | 5.63 | 0 | | FEProblem::computeUserObjects | 3 | 0.008 | 0.003 | 0.83 | 0 | 0.008 | 0.003 | 0.83 | 0 | | Transient::final | 1 | 0.000 | 0.000 | 0.01 | 0 | 0.000 | 0.000 | 0.02 | 0 | | FEProblem::outputStep | 1 | 0.000 | 0.000 | 0.00 | 0 | 0.000 | 0.000 | 0.00 | 0 | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Running exodiff: /data/civet0/build/moose/framework/contrib/exodiff/exodiff -m -F 1e-10 -t 1e-06 /data/civet0/build/moose/modules/heat_conduction/test/tests/truss_heat_conduction/gold/block_w_line_out.e /data/civet0/build/moose/modules/heat_conduction/test/tests/truss_heat_conduction/block_w_line_out.e ERROR: ***************************************************************** EXODIFF (Version: 2.90) Modified: 2018-02-15 Authors: Richard Drake, rrdrake@sandia.gov Greg Sjaardema, gdsjaar@sandia.gov Run on 2022/10/21 00:18:09 MDT ***************************************************************** Reading first file ... Reading second file ... FILE 1: /data/civet0/build/moose/modules/heat_conduction/test/tests/truss_heat_conduction/gold/block_w_line_out.e Title: block_w_line_out.e Dim = 3, Blocks = 2, Nodes = 419, Elements = 160, Nodesets = 8, Sidesets = 8 Vars: Global = 0, Nodal = 1, Element = 1, Nodeset = 0, Sideset = 0, Times = 2 FILE 2: /data/civet0/build/moose/modules/heat_conduction/test/tests/truss_heat_conduction/block_w_line_out.e Title: block_w_line_out.e Dim = 3, Blocks = 2, Nodes = 419, Elements = 160, Nodesets = 8, Sidesets = 8 Vars: Global = 0, Nodal = 1, Element = 1, Nodeset = 0, Sideset = 0, Times = 4 exodiff: ERROR: .. First file has 2 result times while the second file has 4. exodiff: INFO .. Map option is not needed. Nodal coordinates will be compared .. tol: 1e-06 (absolute), floor: 0 Time step values will be compared .. tol: 1e-06 (relative), floor: 1e-10 No Global variables on either file. Nodal variables to be compared: temperature tol: 1e-06 (relative), floor: 1e-10 Element variables to be compared: area tol: 1e-06 (relative), floor: 1e-10 No Element Attribute variables on either file. No Nodeset variables on either file. No Sideset variables on either file. Sideset Distribution Factors will be compared .. tol: 1e-06 (relative), floor: 0 ============================================================== NOTE: All node and element ids are reported as global ids. Sideset Distribution Factors: --------- Time step 1, 0.0000000e+00 ~ 0.0000000e+00, rel diff: 0.00000e+00 --------- Nodal variables: Element variables: --------- Time step 2, 1.0000000e+00 ~ 5.0000000e-01, rel diff: 5.00000e-01 (FAILED) Nodal variables: temperature rel diff: 6.7393323e+00 ~ 2.4136390e+00 = 6.41858e-01 (node 402) Element variables: exodiff: Files are different ################################################################################ Tester failed, reason: EXODIFF ``` ## Steps to Reproduce run that test many times ## Impact lost time
non_priority
unreliable heat conduction truss test in mpi threads recipe bug description stochastic test failures on distrbuted threads working directory data build moose modules heat conduction test tests truss heat conduction running command mpiexec n data build moose modules heat conduction heat conduction opt i block w line i distributed mesh error error override timing outputs perf graph true no gdb backtrace n threads framework information moose version git commit on libmesh version petsc version slepc version current time fri oct executable timestamp fri oct parallelism num processors num threads mesh parallel type replicated forced mesh dimension spatial dimension nodes total local min max avg elems total local min max avg num subdomains num partitions partitioner metis nonlinear system num dofs num local dofs variables temperature finite element types lagrange approximation orders first auxiliary system num dofs num local dofs variables area finite element types monomial approximation orders constant execution information executioner transient timestepper constantdt solver mode preconditioned jfnk petsc preconditioner hypre boomeramg strong threshold auto time step time time step time dt nonlinear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear solve did not converge due to diverged line search iterations solve did not converge aborting as solve did not converge time step time dt nonlinear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear r solve converged time step time dt nonlinear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear solve did not converge due to diverged line search iterations solve did not converge aborting as solve did not converge time step time dt nonlinear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r nonlinear r solve converged time step time dt nonlinear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r nonlinear r solve converged performance graph section calls self s avg s mem mb total s avg s mem mb heatconductiontestapp main action setupmeshaction mesh setupmeshaction act setup mesh action setupmeshaction mesh setupmeshaction act set mesh base mooseapp executemeshgenerators feproblem outputstep exodus outputstep transient picardsolve feproblem outputstep feproblem solve feproblem computeresidualinternal feproblem computeresidualinternal feproblem computejacobianinternal feproblem computeuserobjects transient final feproblem outputstep framework information moose version git commit on libmesh version petsc version slepc version current time fri oct executable timestamp fri oct parallelism num processors num threads mesh parallel type replicated forced mesh dimension spatial dimension nodes total local min max avg elems total local min max avg num subdomains num partitions partitioner metis nonlinear system num dofs num local dofs variables temperature finite element types lagrange approximation orders first auxiliary system num dofs num local dofs variables area finite element types monomial approximation orders constant execution information executioner transient timestepper constantdt solver mode preconditioned jfnk petsc preconditioner hypre boomeramg strong threshold auto time step time time step time dt nonlinear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear solve did not converge due to diverged line search iterations solve did not converge aborting as solve did not converge time step time dt nonlinear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear r solve converged time step time dt nonlinear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear r linear solve did not converge due to diverged breakdown iterations nonlinear solve did not converge due to diverged line search iterations solve did not converge aborting as solve did not converge time step time dt nonlinear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r nonlinear r solve converged time step time dt nonlinear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r linear r linear r nonlinear r linear r linear r linear r linear r linear r nonlinear r solve converged performance graph section calls self s avg s mem mb total s avg s mem mb heatconductiontestapp main action setupmeshaction mesh setupmeshaction act setup mesh action setupmeshaction mesh setupmeshaction act set mesh base mooseapp executemeshgenerators feproblem outputstep exodus outputstep transient picardsolve feproblem outputstep feproblem solve feproblem computeresidualinternal feproblem computeresidualinternal feproblem computejacobianinternal feproblem computeuserobjects transient final feproblem outputstep running exodiff data build moose framework contrib exodiff exodiff m f t data build moose modules heat conduction test tests truss heat conduction gold block w line out e data build moose modules heat conduction test tests truss heat conduction block w line out e error exodiff version modified authors richard drake rrdrake sandia gov greg sjaardema gdsjaar sandia gov run on mdt reading first file reading second file file data build moose modules heat conduction test tests truss heat conduction gold block w line out e title block w line out e dim blocks nodes elements nodesets sidesets vars global nodal element nodeset sideset times file data build moose modules heat conduction test tests truss heat conduction block w line out e title block w line out e dim blocks nodes elements nodesets sidesets vars global nodal element nodeset sideset times exodiff error first file has result times while the second file has exodiff info map option is not needed nodal coordinates will be compared tol absolute floor time step values will be compared tol relative floor no global variables on either file nodal variables to be compared temperature tol relative floor element variables to be compared area tol relative floor no element attribute variables on either file no nodeset variables on either file no sideset variables on either file sideset distribution factors will be compared tol relative floor note all node and element ids are reported as global ids sideset distribution factors time step rel diff nodal variables element variables time step rel diff failed nodal variables temperature rel diff node element variables exodiff files are different tester failed reason exodiff steps to reproduce run that test many times impact lost time
0
2,390
5,135,021,097
IssuesEvent
2017-01-11 10:54:12
Infocatcher/Private_Tab
https://api.github.com/repos/Infocatcher/Private_Tab
opened
[e10s] Unable to restore closed private tabs
bug compatibility
Disabled by default, for _extensions.privateTab.rememberClosedPrivateTabs_ = true resource:///modules/sessionstore/SessionStore.jsm SessionStoreInternal.undoCloseTab() -> SessionStoreInternal.restoreTab() ```js if ("attributes" in tabData) { // Ensure that we persist tab attributes restored from previous sessions. Object.keys(tabData.attributes).forEach(a => TabAttributes.persist(a)); } ... browser.messageManager.sendAsyncMessage("SessionStore:restoreHistory", {tabData: tabData, epoch: epoch, loadArguments}); ``` chrome://browser/content/content-sessionStore.js --> MessageListener.receiveMessage() -> MessageListener.restoreHistory() -> sendSyncMessage("SessionStore:restoreHistoryComplete", …) resource:///modules/sessionstore/SessionStore.jsm --> SessionStoreInternal.receiveMessage() ```js let event = win.document.createEvent("Events"); event.initEvent("SSTabRestoring", true, false); tab.dispatchEvent(event); ``` So, loading will be started before "SSTabRestoring" and in non-private mode.
True
[e10s] Unable to restore closed private tabs - Disabled by default, for _extensions.privateTab.rememberClosedPrivateTabs_ = true resource:///modules/sessionstore/SessionStore.jsm SessionStoreInternal.undoCloseTab() -> SessionStoreInternal.restoreTab() ```js if ("attributes" in tabData) { // Ensure that we persist tab attributes restored from previous sessions. Object.keys(tabData.attributes).forEach(a => TabAttributes.persist(a)); } ... browser.messageManager.sendAsyncMessage("SessionStore:restoreHistory", {tabData: tabData, epoch: epoch, loadArguments}); ``` chrome://browser/content/content-sessionStore.js --> MessageListener.receiveMessage() -> MessageListener.restoreHistory() -> sendSyncMessage("SessionStore:restoreHistoryComplete", …) resource:///modules/sessionstore/SessionStore.jsm --> SessionStoreInternal.receiveMessage() ```js let event = win.document.createEvent("Events"); event.initEvent("SSTabRestoring", true, false); tab.dispatchEvent(event); ``` So, loading will be started before "SSTabRestoring" and in non-private mode.
non_priority
unable to restore closed private tabs disabled by default for extensions privatetab rememberclosedprivatetabs true resource modules sessionstore sessionstore jsm sessionstoreinternal undoclosetab sessionstoreinternal restoretab js if attributes in tabdata ensure that we persist tab attributes restored from previous sessions object keys tabdata attributes foreach a tabattributes persist a browser messagemanager sendasyncmessage sessionstore restorehistory tabdata tabdata epoch epoch loadarguments chrome browser content content sessionstore js messagelistener receivemessage messagelistener restorehistory sendsyncmessage sessionstore restorehistorycomplete … resource modules sessionstore sessionstore jsm sessionstoreinternal receivemessage js let event win document createevent events event initevent sstabrestoring true false tab dispatchevent event so loading will be started before sstabrestoring and in non private mode
0
436,342
12,550,353,514
IssuesEvent
2020-06-06 10:48:26
googleapis/google-api-java-client-services
https://api.github.com/repos/googleapis/google-api-java-client-services
opened
Synthesis failed for firebasedynamiclinks
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate firebasedynamiclinks. :broken_heart: Here's the output from running `synth.py`: ``` 2020-06-06 03:48:19,869 autosynth [INFO] > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/google-api-java-client-services 2020-06-06 03:48:20,900 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2020-06-06 03:48:20,904 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2020-06-06 03:48:20,907 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2020-06-06 03:48:20,925 autosynth [DEBUG] > Running: git config push.default simple 2020-06-06 03:48:20,929 autosynth [DEBUG] > Running: git branch -f autosynth-firebasedynamiclinks 2020-06-06 03:48:20,933 autosynth [DEBUG] > Running: git checkout autosynth-firebasedynamiclinks Switched to branch 'autosynth-firebasedynamiclinks' 2020-06-06 03:48:21,372 autosynth [INFO] > Running synthtool 2020-06-06 03:48:21,372 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/google-api-services-firebasedynamiclinks/synth.metadata', 'synth.py', '--'] 2020-06-06 03:48:21,374 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata clients/google-api-services-firebasedynamiclinks/synth.metadata synth.py -- firebasedynamiclinks tee: /tmpfs/src/github/synthtool/logs/googleapis/google-api-java-client-services: Is a directory 2020-06-06 03:48:21,589 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-java-client-services/synth.py. On branch autosynth-firebasedynamiclinks nothing to commit, working tree clean 2020-06-06 03:48:21,974 synthtool [DEBUG] > Cloning https://github.com/googleapis/discovery-artifact-manager.git. DEBUG:synthtool:Cloning https://github.com/googleapis/discovery-artifact-manager.git. 2020-06-06 03:48:22,610 synthtool [DEBUG] > Cleaning output directory. DEBUG:synthtool:Cleaning output directory. 2020-06-06 03:48:22,613 synthtool [DEBUG] > Installing dependencies. DEBUG:synthtool:Installing dependencies. 2020-06-06 03:48:24,397 synthtool [INFO] > Generating firebasedynamiclinks v1. INFO:synthtool:Generating firebasedynamiclinks v1. 2020-06-06 03:48:24,398 synthtool [INFO] > 1.30.1 INFO:synthtool: 1.30.1 Writing json metadata to clients/google-api-services-firebasedynamiclinks/v1.metadata.json 2020-06-06 03:48:25,110 synthtool [DEBUG] > Wrote metadata to clients/google-api-services-firebasedynamiclinks/synth.metadata. DEBUG:synthtool:Wrote metadata to clients/google-api-services-firebasedynamiclinks/synth.metadata. 2020-06-06 03:48:25,144 autosynth [DEBUG] > Running: git clean -fdx Removing .cache/ Removing __pycache__/ Removing clients/google-api-services-firebasedynamiclinks/synth.metadata Removing generator/.cache/ Removing generator/ez_setup.pyc Removing generator/src/google_apis_client_generator.egg-info/ Removing generator/src/googleapis/__init__.pyc Removing generator/src/googleapis/codegen/__init__.pyc Removing generator/src/googleapis/codegen/api.pyc Removing generator/src/googleapis/codegen/api_exception.pyc Removing generator/src/googleapis/codegen/api_library_generator.pyc Removing generator/src/googleapis/codegen/data_types.pyc Removing generator/src/googleapis/codegen/django_helpers.pyc Removing generator/src/googleapis/codegen/filesys/__init__.pyc Removing generator/src/googleapis/codegen/filesys/files.pyc Removing generator/src/googleapis/codegen/filesys/filesystem_library_package.pyc Removing generator/src/googleapis/codegen/filesys/library_package.pyc Removing generator/src/googleapis/codegen/filesys/package_writer_foundry.pyc Removing generator/src/googleapis/codegen/filesys/tar_library_package.pyc Removing generator/src/googleapis/codegen/filesys/zip_library_package.pyc Removing generator/src/googleapis/codegen/generate_library.pyc Removing generator/src/googleapis/codegen/generator.pyc Removing generator/src/googleapis/codegen/generator_lookup.pyc Removing generator/src/googleapis/codegen/import_definition.pyc Removing generator/src/googleapis/codegen/java_generator.pyc Removing generator/src/googleapis/codegen/java_import_manager.pyc Removing generator/src/googleapis/codegen/language_model.pyc Removing generator/src/googleapis/codegen/schema.pyc Removing generator/src/googleapis/codegen/targets.pyc Removing generator/src/googleapis/codegen/template_helpers.pyc Removing generator/src/googleapis/codegen/template_objects.pyc Removing generator/src/googleapis/codegen/utilities/__init__.pyc Removing generator/src/googleapis/codegen/utilities/convert_size.pyc Removing generator/src/googleapis/codegen/utilities/html_stripper.pyc Removing generator/src/googleapis/codegen/utilities/json_expander.pyc Removing generator/src/googleapis/codegen/utilities/json_with_comments.pyc Removing generator/src/googleapis/codegen/utilities/maven_utils.pyc Removing generator/src/googleapis/codegen/utilities/name_validator.pyc Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 615, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 476, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 555, in _inner_main ).synthesize(base_synth_log_path) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 121, in synthesize with open(log_file_path, "rt") as fp: IsADirectoryError: [Errno 21] Is a directory: '/tmpfs/src/github/synthtool/logs/googleapis/google-api-java-client-services' ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/a12fce2f-fa64-48b1-abd7-5a9cfc4df713/targets/github%2Fsynthtool;config=default/tests;query=google-api-java-client-services;failed=false).
1.0
Synthesis failed for firebasedynamiclinks - Hello! Autosynth couldn't regenerate firebasedynamiclinks. :broken_heart: Here's the output from running `synth.py`: ``` 2020-06-06 03:48:19,869 autosynth [INFO] > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/google-api-java-client-services 2020-06-06 03:48:20,900 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2020-06-06 03:48:20,904 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2020-06-06 03:48:20,907 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2020-06-06 03:48:20,925 autosynth [DEBUG] > Running: git config push.default simple 2020-06-06 03:48:20,929 autosynth [DEBUG] > Running: git branch -f autosynth-firebasedynamiclinks 2020-06-06 03:48:20,933 autosynth [DEBUG] > Running: git checkout autosynth-firebasedynamiclinks Switched to branch 'autosynth-firebasedynamiclinks' 2020-06-06 03:48:21,372 autosynth [INFO] > Running synthtool 2020-06-06 03:48:21,372 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/google-api-services-firebasedynamiclinks/synth.metadata', 'synth.py', '--'] 2020-06-06 03:48:21,374 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata clients/google-api-services-firebasedynamiclinks/synth.metadata synth.py -- firebasedynamiclinks tee: /tmpfs/src/github/synthtool/logs/googleapis/google-api-java-client-services: Is a directory 2020-06-06 03:48:21,589 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-java-client-services/synth.py. On branch autosynth-firebasedynamiclinks nothing to commit, working tree clean 2020-06-06 03:48:21,974 synthtool [DEBUG] > Cloning https://github.com/googleapis/discovery-artifact-manager.git. DEBUG:synthtool:Cloning https://github.com/googleapis/discovery-artifact-manager.git. 2020-06-06 03:48:22,610 synthtool [DEBUG] > Cleaning output directory. DEBUG:synthtool:Cleaning output directory. 2020-06-06 03:48:22,613 synthtool [DEBUG] > Installing dependencies. DEBUG:synthtool:Installing dependencies. 2020-06-06 03:48:24,397 synthtool [INFO] > Generating firebasedynamiclinks v1. INFO:synthtool:Generating firebasedynamiclinks v1. 2020-06-06 03:48:24,398 synthtool [INFO] > 1.30.1 INFO:synthtool: 1.30.1 Writing json metadata to clients/google-api-services-firebasedynamiclinks/v1.metadata.json 2020-06-06 03:48:25,110 synthtool [DEBUG] > Wrote metadata to clients/google-api-services-firebasedynamiclinks/synth.metadata. DEBUG:synthtool:Wrote metadata to clients/google-api-services-firebasedynamiclinks/synth.metadata. 2020-06-06 03:48:25,144 autosynth [DEBUG] > Running: git clean -fdx Removing .cache/ Removing __pycache__/ Removing clients/google-api-services-firebasedynamiclinks/synth.metadata Removing generator/.cache/ Removing generator/ez_setup.pyc Removing generator/src/google_apis_client_generator.egg-info/ Removing generator/src/googleapis/__init__.pyc Removing generator/src/googleapis/codegen/__init__.pyc Removing generator/src/googleapis/codegen/api.pyc Removing generator/src/googleapis/codegen/api_exception.pyc Removing generator/src/googleapis/codegen/api_library_generator.pyc Removing generator/src/googleapis/codegen/data_types.pyc Removing generator/src/googleapis/codegen/django_helpers.pyc Removing generator/src/googleapis/codegen/filesys/__init__.pyc Removing generator/src/googleapis/codegen/filesys/files.pyc Removing generator/src/googleapis/codegen/filesys/filesystem_library_package.pyc Removing generator/src/googleapis/codegen/filesys/library_package.pyc Removing generator/src/googleapis/codegen/filesys/package_writer_foundry.pyc Removing generator/src/googleapis/codegen/filesys/tar_library_package.pyc Removing generator/src/googleapis/codegen/filesys/zip_library_package.pyc Removing generator/src/googleapis/codegen/generate_library.pyc Removing generator/src/googleapis/codegen/generator.pyc Removing generator/src/googleapis/codegen/generator_lookup.pyc Removing generator/src/googleapis/codegen/import_definition.pyc Removing generator/src/googleapis/codegen/java_generator.pyc Removing generator/src/googleapis/codegen/java_import_manager.pyc Removing generator/src/googleapis/codegen/language_model.pyc Removing generator/src/googleapis/codegen/schema.pyc Removing generator/src/googleapis/codegen/targets.pyc Removing generator/src/googleapis/codegen/template_helpers.pyc Removing generator/src/googleapis/codegen/template_objects.pyc Removing generator/src/googleapis/codegen/utilities/__init__.pyc Removing generator/src/googleapis/codegen/utilities/convert_size.pyc Removing generator/src/googleapis/codegen/utilities/html_stripper.pyc Removing generator/src/googleapis/codegen/utilities/json_expander.pyc Removing generator/src/googleapis/codegen/utilities/json_with_comments.pyc Removing generator/src/googleapis/codegen/utilities/maven_utils.pyc Removing generator/src/googleapis/codegen/utilities/name_validator.pyc Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 615, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 476, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 555, in _inner_main ).synthesize(base_synth_log_path) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 121, in synthesize with open(log_file_path, "rt") as fp: IsADirectoryError: [Errno 21] Is a directory: '/tmpfs/src/github/synthtool/logs/googleapis/google-api-java-client-services' ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/a12fce2f-fa64-48b1-abd7-5a9cfc4df713/targets/github%2Fsynthtool;config=default/tests;query=google-api-java-client-services;failed=false).
priority
synthesis failed for firebasedynamiclinks hello autosynth couldn t regenerate firebasedynamiclinks broken heart here s the output from running synth py autosynth logs will be written to tmpfs src github synthtool logs googleapis google api java client services autosynth running git config global core excludesfile home kbuilder autosynth gitignore autosynth running git config user name yoshi automation autosynth running git config user email yoshi automation google com autosynth running git config push default simple autosynth running git branch f autosynth firebasedynamiclinks autosynth running git checkout autosynth firebasedynamiclinks switched to branch autosynth firebasedynamiclinks autosynth running synthtool autosynth autosynth running tmpfs src github synthtool env bin m synthtool metadata clients google api services firebasedynamiclinks synth metadata synth py firebasedynamiclinks tee tmpfs src github synthtool logs googleapis google api java client services is a directory synthtool executing home kbuilder cache synthtool google api java client services synth py on branch autosynth firebasedynamiclinks nothing to commit working tree clean synthtool cloning debug synthtool cloning synthtool cleaning output directory debug synthtool cleaning output directory synthtool installing dependencies debug synthtool installing dependencies synthtool generating firebasedynamiclinks info synthtool generating firebasedynamiclinks synthtool info synthtool writing json metadata to clients google api services firebasedynamiclinks metadata json synthtool wrote metadata to clients google api services firebasedynamiclinks synth metadata debug synthtool wrote metadata to clients google api services firebasedynamiclinks synth metadata autosynth running git clean fdx removing cache removing pycache removing clients google api services firebasedynamiclinks synth metadata removing generator cache removing generator ez setup pyc removing generator src google apis client generator egg info removing generator src googleapis init pyc removing generator src googleapis codegen init pyc removing generator src googleapis codegen api pyc removing generator src googleapis codegen api exception pyc removing generator src googleapis codegen api library generator pyc removing generator src googleapis codegen data types pyc removing generator src googleapis codegen django helpers pyc removing generator src googleapis codegen filesys init pyc removing generator src googleapis codegen filesys files pyc removing generator src googleapis codegen filesys filesystem library package pyc removing generator src googleapis codegen filesys library package pyc removing generator src googleapis codegen filesys package writer foundry pyc removing generator src googleapis codegen filesys tar library package pyc removing generator src googleapis codegen filesys zip library package pyc removing generator src googleapis codegen generate library pyc removing generator src googleapis codegen generator pyc removing generator src googleapis codegen generator lookup pyc removing generator src googleapis codegen import definition pyc removing generator src googleapis codegen java generator pyc removing generator src googleapis codegen java import manager pyc removing generator src googleapis codegen language model pyc removing generator src googleapis codegen schema pyc removing generator src googleapis codegen targets pyc removing generator src googleapis codegen template helpers pyc removing generator src googleapis codegen template objects pyc removing generator src googleapis codegen utilities init pyc removing generator src googleapis codegen utilities convert size pyc removing generator src googleapis codegen utilities html stripper pyc removing generator src googleapis codegen utilities json expander pyc removing generator src googleapis codegen utilities json with comments pyc removing generator src googleapis codegen utilities maven utils pyc removing generator src googleapis codegen utilities name validator pyc traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize base synth log path file tmpfs src github synthtool autosynth synthesizer py line in synthesize with open log file path rt as fp isadirectoryerror is a directory tmpfs src github synthtool logs googleapis google api java client services google internal developers can see the full log
1
165,818
26,229,914,392
IssuesEvent
2023-01-04 22:41:59
MozillaFoundation/Design
https://api.github.com/repos/MozillaFoundation/Design
opened
[PNI] Audit and replace all lowercase instances with Title Case *PNI
design Buyer's Guide
Decision was made to have PNI be in title case last year - [ ] Find all instances where PNI is lowercase - [ ] Update to Title case
1.0
[PNI] Audit and replace all lowercase instances with Title Case *PNI - Decision was made to have PNI be in title case last year - [ ] Find all instances where PNI is lowercase - [ ] Update to Title case
non_priority
audit and replace all lowercase instances with title case pni decision was made to have pni be in title case last year find all instances where pni is lowercase update to title case
0
205,229
15,597,812,892
IssuesEvent
2021-03-18 17:20:49
NationalSecurityAgency/skills-service
https://api.github.com/repos/NationalSecurityAgency/skills-service
closed
Update `AuthorizationAspect` to check if proxied user is also a dashboard user
bug test
the scenario being that the token is actually for a dashboard user attempting to use an admin endpoint that uses the `AuthorizationAspect`
1.0
Update `AuthorizationAspect` to check if proxied user is also a dashboard user - the scenario being that the token is actually for a dashboard user attempting to use an admin endpoint that uses the `AuthorizationAspect`
non_priority
update authorizationaspect to check if proxied user is also a dashboard user the scenario being that the token is actually for a dashboard user attempting to use an admin endpoint that uses the authorizationaspect
0
27,320
4,300,684,824
IssuesEvent
2016-07-20 02:46:14
coreos/etcd
https://api.github.com/repos/coreos/etcd
opened
test: performance regression tests
area/performance area/testing kind/enhancement
It's often unclear whether a change to etcd will significantly impact performance. To catch performance issues early, etcd should have performance regression testing including historical data. The performance data could be t-tested for some statistical confidence. I don't expect this to need significant amounts of new code, a lot of resources, or much time to run-- it would launch a cluster, run tools/benchmark, store the results somewhere, and calculate the change in performance. It'd be neat if this could be integrated with CI (but, obviously, it wouldn't be able to run directly on travis/semaphore) so that performance regressions are gated before commit.
1.0
test: performance regression tests - It's often unclear whether a change to etcd will significantly impact performance. To catch performance issues early, etcd should have performance regression testing including historical data. The performance data could be t-tested for some statistical confidence. I don't expect this to need significant amounts of new code, a lot of resources, or much time to run-- it would launch a cluster, run tools/benchmark, store the results somewhere, and calculate the change in performance. It'd be neat if this could be integrated with CI (but, obviously, it wouldn't be able to run directly on travis/semaphore) so that performance regressions are gated before commit.
non_priority
test performance regression tests it s often unclear whether a change to etcd will significantly impact performance to catch performance issues early etcd should have performance regression testing including historical data the performance data could be t tested for some statistical confidence i don t expect this to need significant amounts of new code a lot of resources or much time to run it would launch a cluster run tools benchmark store the results somewhere and calculate the change in performance it d be neat if this could be integrated with ci but obviously it wouldn t be able to run directly on travis semaphore so that performance regressions are gated before commit
0
301,057
9,215,394,407
IssuesEvent
2019-03-11 02:54:20
cuappdev/ithaca-transit-backend
https://api.github.com/repos/cuappdev/ithaca-transit-backend
reopened
v2: Refactor Models
Priority: Medium Type: Maintenance
# Models ## Route - departureTime: Date - arrivalTime: Date - startCoords: Coordinates // see below - endCoords: Coordinates // see below - travelDistance: Double // distance between start and end coords in miles - startName: String - endName: String - boundingBox: Bounds // see below - totalDuration: Int // minutes between departureTime and arrivalTime - routeSummary: [RouteSummaryElement] // see @myo3's comment below - detailDirections: [Direction] // see comment below ## Direction - type: DirectionType - name: String // General Description (bus stop name, Google Place name) - startLocation: Location - endLocation: Location - startTime: Date - endTime: Date - path: [Coordinates] - travelDistance: Double - routeNumber: Int - stops: [Location] - stayOnBusTransfer: Bool - tripIdentifiers: [String]? - delay: Int? ### detailDirections Logic *see Route.swift Line 164 on for front-end implantation* 1. You will need to create `arrive` and `.transfer` directions. Use `DirectionType` accordingly and *correctly*. `stayOnBusTransfer` will help with the latter. 2. Remove the first and last stop in each direction with stops if stops isn't empty. 3. Make sure the last direction's name is the ending location. A typical route should be ``` [.walk, .depart, .arrive, .walk] ``` A route with two buses should be ``` [.walk, .depart, .arrive, .depart, .arrive, .walk] ``` There is no walking at `index 3` because the user waits at the bus stop. A route with a stay on bus transfer should be ``` [.walk, .depart, .transfer, .arrive, .walk] ``` # Other Relevant Classes **Bounds** ``` "bounds" : { "minLat" : Double, "minLong" : Double, "maxLat" : Double, "maxLong" : Double } ``` **Coordinates** ``` "coordinates" : { "lat" : Double, "long" : Double } ``` **DirectionType** ``` enum DirectionType: String { case walk // Directions that involving walking case depart // Directions where the user gets on the bus case arrive // Direction where the user gets off the bus case transfer // IMPORTANT: Direction where transfer is involved, but user stays on bus } ``` **Location** ``` "location" : { "name" : String, "id" : String, // Identifier associated with location, used mainly for stopID for bus stop locations "lat" : Double, "long" : Double } ```
1.0
v2: Refactor Models - # Models ## Route - departureTime: Date - arrivalTime: Date - startCoords: Coordinates // see below - endCoords: Coordinates // see below - travelDistance: Double // distance between start and end coords in miles - startName: String - endName: String - boundingBox: Bounds // see below - totalDuration: Int // minutes between departureTime and arrivalTime - routeSummary: [RouteSummaryElement] // see @myo3's comment below - detailDirections: [Direction] // see comment below ## Direction - type: DirectionType - name: String // General Description (bus stop name, Google Place name) - startLocation: Location - endLocation: Location - startTime: Date - endTime: Date - path: [Coordinates] - travelDistance: Double - routeNumber: Int - stops: [Location] - stayOnBusTransfer: Bool - tripIdentifiers: [String]? - delay: Int? ### detailDirections Logic *see Route.swift Line 164 on for front-end implantation* 1. You will need to create `arrive` and `.transfer` directions. Use `DirectionType` accordingly and *correctly*. `stayOnBusTransfer` will help with the latter. 2. Remove the first and last stop in each direction with stops if stops isn't empty. 3. Make sure the last direction's name is the ending location. A typical route should be ``` [.walk, .depart, .arrive, .walk] ``` A route with two buses should be ``` [.walk, .depart, .arrive, .depart, .arrive, .walk] ``` There is no walking at `index 3` because the user waits at the bus stop. A route with a stay on bus transfer should be ``` [.walk, .depart, .transfer, .arrive, .walk] ``` # Other Relevant Classes **Bounds** ``` "bounds" : { "minLat" : Double, "minLong" : Double, "maxLat" : Double, "maxLong" : Double } ``` **Coordinates** ``` "coordinates" : { "lat" : Double, "long" : Double } ``` **DirectionType** ``` enum DirectionType: String { case walk // Directions that involving walking case depart // Directions where the user gets on the bus case arrive // Direction where the user gets off the bus case transfer // IMPORTANT: Direction where transfer is involved, but user stays on bus } ``` **Location** ``` "location" : { "name" : String, "id" : String, // Identifier associated with location, used mainly for stopID for bus stop locations "lat" : Double, "long" : Double } ```
priority
refactor models models route departuretime date arrivaltime date startcoords coordinates see below endcoords coordinates see below traveldistance double distance between start and end coords in miles startname string endname string boundingbox bounds see below totalduration int minutes between departuretime and arrivaltime routesummary see s comment below detaildirections see comment below direction type directiontype name string general description bus stop name google place name startlocation location endlocation location starttime date endtime date path traveldistance double routenumber int stops stayonbustransfer bool tripidentifiers delay int detaildirections logic see route swift line on for front end implantation you will need to create arrive and transfer directions use directiontype accordingly and correctly stayonbustransfer will help with the latter remove the first and last stop in each direction with stops if stops isn t empty make sure the last direction s name is the ending location a typical route should be a route with two buses should be there is no walking at index because the user waits at the bus stop a route with a stay on bus transfer should be other relevant classes bounds bounds minlat double minlong double maxlat double maxlong double coordinates coordinates lat double long double directiontype enum directiontype string case walk directions that involving walking case depart directions where the user gets on the bus case arrive direction where the user gets off the bus case transfer important direction where transfer is involved but user stays on bus location location name string id string identifier associated with location used mainly for stopid for bus stop locations lat double long double
1
9,822
4,647,704,816
IssuesEvent
2016-10-01 17:04:33
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
OpenAPI generated files break incremental builds
area/build-release component/kubectl team/CSI-API Machinery SIG
Incremental builds seem to be broken by #31468. To repo: 1. Run `make WHAT=cmd/kubectl` 2. Run `make WHAT=cmd/kubectl` again Notice that it is rebuilding kubectl. We can debug by running `make WHAT=cmd/kubectl --debug=b` and see lines like: ``` File 'cmd/libs/go2idl/client-gen/test_apis/testgroup.k8s.io/v1/zz_generated.openapi.go' does not exist. Must remake target 'cmd/libs/go2idl/client-gen/test_apis/testgroup.k8s.io/v1/zz_generated.openapi.go'. Successfully remade target file 'cmd/libs/go2idl/client-gen/test_apis/testgroup.k8s.io/v1/zz_generated.openapi.go'. ``` It looks like the makefiles are creating a dependency on a file that never gets built. @thockin -- you managed to make this work with the other generated files -- any advice here?
1.0
OpenAPI generated files break incremental builds - Incremental builds seem to be broken by #31468. To repo: 1. Run `make WHAT=cmd/kubectl` 2. Run `make WHAT=cmd/kubectl` again Notice that it is rebuilding kubectl. We can debug by running `make WHAT=cmd/kubectl --debug=b` and see lines like: ``` File 'cmd/libs/go2idl/client-gen/test_apis/testgroup.k8s.io/v1/zz_generated.openapi.go' does not exist. Must remake target 'cmd/libs/go2idl/client-gen/test_apis/testgroup.k8s.io/v1/zz_generated.openapi.go'. Successfully remade target file 'cmd/libs/go2idl/client-gen/test_apis/testgroup.k8s.io/v1/zz_generated.openapi.go'. ``` It looks like the makefiles are creating a dependency on a file that never gets built. @thockin -- you managed to make this work with the other generated files -- any advice here?
non_priority
openapi generated files break incremental builds incremental builds seem to be broken by to repo run make what cmd kubectl run make what cmd kubectl again notice that it is rebuilding kubectl we can debug by running make what cmd kubectl debug b and see lines like file cmd libs client gen test apis testgroup io zz generated openapi go does not exist must remake target cmd libs client gen test apis testgroup io zz generated openapi go successfully remade target file cmd libs client gen test apis testgroup io zz generated openapi go it looks like the makefiles are creating a dependency on a file that never gets built thockin you managed to make this work with the other generated files any advice here
0
97,442
3,993,176,546
IssuesEvent
2016-05-10 07:07:24
MinetestForFun/server-minetestforfun
https://api.github.com/repos/MinetestForFun/server-minetestforfun
closed
text on signs is randomly turned of 90°
Modding ➤ BugFix Priority: Low
![capture du 2016-05-10 08 47 24](https://cloud.githubusercontent.com/assets/12512762/15137847/66e61162-168c-11e6-8eee-c39b30611a9d.png) Here is how I found today signs I have put in front of the public factory entrance yesterday evening. ![capture du 2016-05-10 08 47 54](https://cloud.githubusercontent.com/assets/12512762/15137848/66e62b7a-168c-11e6-904d-f962c5352da6.png) and how they should be. The text seems to turn 90° at some occasion, I already noticed it before. Note: the left of the signs on the example is oriented south, the right to the north.
1.0
text on signs is randomly turned of 90° - ![capture du 2016-05-10 08 47 24](https://cloud.githubusercontent.com/assets/12512762/15137847/66e61162-168c-11e6-8eee-c39b30611a9d.png) Here is how I found today signs I have put in front of the public factory entrance yesterday evening. ![capture du 2016-05-10 08 47 54](https://cloud.githubusercontent.com/assets/12512762/15137848/66e62b7a-168c-11e6-904d-f962c5352da6.png) and how they should be. The text seems to turn 90° at some occasion, I already noticed it before. Note: the left of the signs on the example is oriented south, the right to the north.
priority
text on signs is randomly turned of ° here is how i found today signs i have put in front of the public factory entrance yesterday evening and how they should be the text seems to turn ° at some occasion i already noticed it before note the left of the signs on the example is oriented south the right to the north
1
167,144
13,001,824,831
IssuesEvent
2020-07-24 01:02:14
elastic/uptime
https://api.github.com/repos/elastic/uptime
closed
[Uptime] Switch away from Auto-date histogram
Team:uptime bug test-plan test-plan-fail v7.9.0
We started using auto data histogram again and have noticed that it sometimes uses too few buckets resulting in awkward looking histograms. Let's switch back to using: https://github.com/elastic/kibana/blob/master/x-pack/plugins/uptime/server/lib/helper/get_histogram_interval.ts for our histograms on the overview page since it does what we expect, evenly dividing time.
2.0
[Uptime] Switch away from Auto-date histogram - We started using auto data histogram again and have noticed that it sometimes uses too few buckets resulting in awkward looking histograms. Let's switch back to using: https://github.com/elastic/kibana/blob/master/x-pack/plugins/uptime/server/lib/helper/get_histogram_interval.ts for our histograms on the overview page since it does what we expect, evenly dividing time.
non_priority
switch away from auto date histogram we started using auto data histogram again and have noticed that it sometimes uses too few buckets resulting in awkward looking histograms let s switch back to using for our histograms on the overview page since it does what we expect evenly dividing time
0
32,844
7,609,440,092
IssuesEvent
2018-05-01 00:29:48
cretonne/cretonne
https://api.github.com/repos/cretonne/cretonne
opened
Jump Tables
generated code quality
Codegen currently legalizes `br_table` instructions by [lowering them to sequences of branches](https://github.com/cretonne/cretonne/blob/master/lib/codegen/src/legalizer/mod.rs#L187). What we want is to use actual jump tables. This will require some new `binemit` capabilities, for emitting the actual jump table data, outside of the function body. In theory, the `RelocSink` trait has support for relocating jump tables (`reloc_jt`), however as this hasn't been tested, it may need to be modified or extended. A jump table is an array, and the elements can be either absolute addresses or offsets relative to some base, often the base of the jump table itself. PIC prefers relative offsets, so we should start with that, and do absolute offsets later if extra speed for non-PIC is needed. Using a jump table typically involves a bounds-check to see if the `br_table` index is in bounds, and then an indexed load into the table using the `br_table` index. In the case of a relative offset, the base is then added to the result. Then, the final result can be used as the operand to an indirect jump.
1.0
Jump Tables - Codegen currently legalizes `br_table` instructions by [lowering them to sequences of branches](https://github.com/cretonne/cretonne/blob/master/lib/codegen/src/legalizer/mod.rs#L187). What we want is to use actual jump tables. This will require some new `binemit` capabilities, for emitting the actual jump table data, outside of the function body. In theory, the `RelocSink` trait has support for relocating jump tables (`reloc_jt`), however as this hasn't been tested, it may need to be modified or extended. A jump table is an array, and the elements can be either absolute addresses or offsets relative to some base, often the base of the jump table itself. PIC prefers relative offsets, so we should start with that, and do absolute offsets later if extra speed for non-PIC is needed. Using a jump table typically involves a bounds-check to see if the `br_table` index is in bounds, and then an indexed load into the table using the `br_table` index. In the case of a relative offset, the base is then added to the result. Then, the final result can be used as the operand to an indirect jump.
non_priority
jump tables codegen currently legalizes br table instructions by what we want is to use actual jump tables this will require some new binemit capabilities for emitting the actual jump table data outside of the function body in theory the relocsink trait has support for relocating jump tables reloc jt however as this hasn t been tested it may need to be modified or extended a jump table is an array and the elements can be either absolute addresses or offsets relative to some base often the base of the jump table itself pic prefers relative offsets so we should start with that and do absolute offsets later if extra speed for non pic is needed using a jump table typically involves a bounds check to see if the br table index is in bounds and then an indexed load into the table using the br table index in the case of a relative offset the base is then added to the result then the final result can be used as the operand to an indirect jump
0
673,217
22,953,048,631
IssuesEvent
2022-07-19 09:07:03
GEWIS/gewisdb
https://api.github.com/repos/GEWIS/gewisdb
closed
`SubDecision` hash is not collision proof
Type: Bug Module: Checker Priority: Medium Status: Confirmed
### Current behaviour The current generation of hashes is not collision proof, as misuse of the decision data can result in the same hash. This is because it is a simple concatenation of the different attributes. ### Desired behaviour _No response_ ### Steps to reproduce Example: XX 12.34.56.7 and XX 123.4.56.7 both result in `XX1234567`. ### Database version e26026c ### What operating are you seeing the problem on? _No response_ ### What browsers are you seeing the problem on? _No response_ ### Other information _No response_
1.0
`SubDecision` hash is not collision proof - ### Current behaviour The current generation of hashes is not collision proof, as misuse of the decision data can result in the same hash. This is because it is a simple concatenation of the different attributes. ### Desired behaviour _No response_ ### Steps to reproduce Example: XX 12.34.56.7 and XX 123.4.56.7 both result in `XX1234567`. ### Database version e26026c ### What operating are you seeing the problem on? _No response_ ### What browsers are you seeing the problem on? _No response_ ### Other information _No response_
priority
subdecision hash is not collision proof current behaviour the current generation of hashes is not collision proof as misuse of the decision data can result in the same hash this is because it is a simple concatenation of the different attributes desired behaviour no response steps to reproduce example xx and xx both result in database version what operating are you seeing the problem on no response what browsers are you seeing the problem on no response other information no response
1
739,499
25,599,695,155
IssuesEvent
2022-12-01 19:02:47
NCAR/wrfcloud
https://api.github.com/repos/NCAR/wrfcloud
closed
Finalize output viewer
priority: blocker type: enhancement component: graphics component: web ui
*Replace italics below with details for this issue.* ## Describe the Enhancement ## *Provide a description of the enhancement request here.* ### Time Estimate ### *Estimate the amount of work required here.* *Issues should represent approximately 1 to 3 days of work.* ### Sub-Issues ### Consider breaking the enhancement down into sub-issues. - [ ] *Add a checkbox for each sub-issue here.* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** ### Projects and Milestone ### - [ ] Select **Project** - [ ] Select **Milestone** as the next official version or **Backlog of Development Ideas** ## Enhancement Checklist ## - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>/<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)**, **Project**, and **Development** issue Select: **Milestone** as the next official version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
1.0
Finalize output viewer - *Replace italics below with details for this issue.* ## Describe the Enhancement ## *Provide a description of the enhancement request here.* ### Time Estimate ### *Estimate the amount of work required here.* *Issues should represent approximately 1 to 3 days of work.* ### Sub-Issues ### Consider breaking the enhancement down into sub-issues. - [ ] *Add a checkbox for each sub-issue here.* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** ### Projects and Milestone ### - [ ] Select **Project** - [ ] Select **Milestone** as the next official version or **Backlog of Development Ideas** ## Enhancement Checklist ## - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>/<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)**, **Project**, and **Development** issue Select: **Milestone** as the next official version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
priority
finalize output viewer replace italics below with details for this issue describe the enhancement provide a description of the enhancement request here time estimate estimate the amount of work required here issues should represent approximately to days of work sub issues consider breaking the enhancement down into sub issues add a checkbox for each sub issue here relevant deadlines list relevant project deadlines here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority projects and milestone select project select milestone as the next official version or backlog of development ideas enhancement checklist complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s project and development issue select milestone as the next official version iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue
1
170,263
14,246,584,215
IssuesEvent
2020-11-19 10:15:49
postmanlabs/postman-app-support
https://api.github.com/repos/postmanlabs/postman-app-support
closed
Optional query parameters no longer show in documenter
feature product/documentation
Optional query parameters no longer show in documenter. We can see the names and descriptions of parameters that are checked (and showing in the URL), but others aren't showing. I believe they did in a previous version of the generated documentation. This is a problem because I would like my API users to know what options they can call my routes with without showing them all in the URL (it is messy and sometimes params can't combine). **Screenshots** This is an example of a request definition I'm doing in the postman app. ![image](https://user-images.githubusercontent.com/37486794/81792511-71d0ae80-9508-11ea-9fe0-efe0b66d30b9.png) I'd like to see all query params (even unchecked) in the generated documentation. But I get this incomplete documentation ![image](https://user-images.githubusercontent.com/37486794/81793836-2f0fd600-950a-11ea-868b-459d35b1552f.png) **App information (please complete the following information):** - App Type : Native App - Postman Version 7.24.0 - OS: Linux If I'm mistaken in the usage of postman documentation, please guide me and close this issue ! Thanks for the hard work
1.0
Optional query parameters no longer show in documenter - Optional query parameters no longer show in documenter. We can see the names and descriptions of parameters that are checked (and showing in the URL), but others aren't showing. I believe they did in a previous version of the generated documentation. This is a problem because I would like my API users to know what options they can call my routes with without showing them all in the URL (it is messy and sometimes params can't combine). **Screenshots** This is an example of a request definition I'm doing in the postman app. ![image](https://user-images.githubusercontent.com/37486794/81792511-71d0ae80-9508-11ea-9fe0-efe0b66d30b9.png) I'd like to see all query params (even unchecked) in the generated documentation. But I get this incomplete documentation ![image](https://user-images.githubusercontent.com/37486794/81793836-2f0fd600-950a-11ea-868b-459d35b1552f.png) **App information (please complete the following information):** - App Type : Native App - Postman Version 7.24.0 - OS: Linux If I'm mistaken in the usage of postman documentation, please guide me and close this issue ! Thanks for the hard work
non_priority
optional query parameters no longer show in documenter optional query parameters no longer show in documenter we can see the names and descriptions of parameters that are checked and showing in the url but others aren t showing i believe they did in a previous version of the generated documentation this is a problem because i would like my api users to know what options they can call my routes with without showing them all in the url it is messy and sometimes params can t combine screenshots this is an example of a request definition i m doing in the postman app i d like to see all query params even unchecked in the generated documentation but i get this incomplete documentation app information please complete the following information app type native app postman version os linux if i m mistaken in the usage of postman documentation please guide me and close this issue thanks for the hard work
0
101,051
16,490,754,415
IssuesEvent
2021-05-25 03:12:57
valdisiljuconoks/episerverless
https://api.github.com/repos/valdisiljuconoks/episerverless
opened
CVE-2020-8203 (High) detected in lodash-2.4.2.tgz
security vulnerability
## CVE-2020-8203 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/package.json</p> <p>Path to vulnerable library: episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/lodash/package.json,episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - intern-geezer-2.2.3.tgz (Root Library) - digdug-1.4.0.tgz - decompress-0.2.3.tgz - map-key-0.1.5.tgz - :x: **lodash-2.4.2.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-07-23</p> <p>Fix Resolution: lodash - 4.17.19</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-8203 (High) detected in lodash-2.4.2.tgz - ## CVE-2020-8203 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/package.json</p> <p>Path to vulnerable library: episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/lodash/package.json,episerverless/packages/EPiServer.CMS.UI.10.10.6/content/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - intern-geezer-2.2.3.tgz (Root Library) - digdug-1.4.0.tgz - decompress-0.2.3.tgz - map-key-0.1.5.tgz - :x: **lodash-2.4.2.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-07-23</p> <p>Fix Resolution: lodash - 4.17.19</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in lodash tgz cve high severity vulnerability vulnerable library lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file episerverless packages episerver cms ui content modules protected shell shell clientresources lib xstyle package json path to vulnerable library episerverless packages episerver cms ui content modules protected shell shell clientresources lib xstyle node modules lodash package json episerverless packages episerver cms ui content modules protected shell shell clientresources lib xstyle node modules lodash package json dependency hierarchy intern geezer tgz root library digdug tgz decompress tgz map key tgz x lodash tgz vulnerable library vulnerability details prototype pollution attack when using zipobjectdeep in lodash before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource
0
236,305
7,748,378,605
IssuesEvent
2018-05-30 08:08:39
Gloirin/m2gTest
https://api.github.com/repos/Gloirin/m2gTest
closed
0002504: Possibility to rearrange application tabs
Tinebase bug low priority
**Reported by rokka on 7 Apr 2010 08:29** **Version:** Mialena (2010-03) Milestone 2 See summary.
1.0
0002504: Possibility to rearrange application tabs - **Reported by rokka on 7 Apr 2010 08:29** **Version:** Mialena (2010-03) Milestone 2 See summary.
priority
possibility to rearrange application tabs reported by rokka on apr version mialena milestone see summary
1
5,190
2,610,182,828
IssuesEvent
2015-02-26 18:58:14
chrsmith/quchuseban
https://api.github.com/repos/chrsmith/quchuseban
opened
解密怎么样去脸上的色斑
auto-migrated Priority-Medium Type-Defect
``` 《摘要》 一直不敢碰触荒凉这个词语,我怕在这寒冷的冬天,被荒凉�� �个词紧紧包裹,阻挡我找寻春天的温暖。可是,你看,真的� ��满眼的灰黄和萧瑟。我不得不站在季节的末梢,看枯叶飘零 ,残阳无力,寒风肆虐。此时,一种繁华落尽的惆怅沧桑着�� �言。原来,世间无论多么盛大的甜美都会有失去的那一天,� ��经的繁华已然远去,无法找寻一丝痕迹。铅华褪尽的世界是 如此的苍凉和晦暗。总是回忆起你,你的笑你的温柔,还有�� �临别时的眼神,那是在嘲笑我把,是啊脸上的色斑已经提醒� ��你已经远去!怎么样去脸上的色斑, 《客户案例》   人为什么长黄褐斑, 该怎么祛斑啊,这个问题也是伴随了我很长时间。记得上大�� �之前,我的肤质本来还是不错的,白白的,可是自从经过开� ��后一段时间的军训,皮肤就变得越来越不好了,不光是晒得 很黑,脸色还有些发黄,鼻子两翼竟然还长出许多的斑斑点�� �,一下子好像老了许多一样。我很是郁闷,军训一结束,我� ��开始修护保养自己的皮肤。过不久,我的皮肤有回复了往日 的白皙,可是脸上的斑却还在那里,因为皮肤的白反而更明�� �了。于是我就开始了作战!与斑点战斗到底!在此期间,我尝�� �了各种祛斑方法,都说效果有多好多好,可结果就是我脸上� ��斑点依旧在那,有的是消失了但是又反弹了。那时真的是很 失望,都快失去信心了。</br>   我把这个烦恼告诉了我一个同事,我同事当时听完,就�� �我说有一款叫黛芙薇尔精华液的产品,美白祛斑功效特别好� ��她家亲戚就一直在使用,听了她的话,我立刻上网搜索黛芙 薇尔精华液,进入到官网,发现很多女性朋友,用了他们的�� �品后,肌肤长斑情况不但得到了改善,而且肌肤比以前还白� ��。这正是我想要的怎样去除脸上的黄褐斑的产品。通过和客 服的沟通,我订购了1个周期的黛芙薇尔精华液。当产品到手� ��,我就立刻使用。</br>   刚开始的两个周期效果不明显,我就咨询了黛芙薇尔精�� �液的客服,她们说:“黛芙薇尔精华液只从根本上治疗黄褐� ��,前两个周期主要是DNA修复阶段,后两个周期就会从皮肤表 面表现出来”。我想这和中国的医理是一样的。后两个周期�� �效果你们就可想而知了。我的黄褐斑真的“下岗”了,皮肤� ��净的像出生的婴儿肤一样。哇,没想到天然精华祛斑产品黛 芙薇尔精华液如此的神奇,使我打败了“顽固的黄褐斑”。�� �道寻找真正能祛斑的产品不易,想到自己也是看了别人的文� ��才能去除黄褐斑,因此我也将自己的祛斑经历写下来发表, 希望和我一样有祛斑痛苦经历的人能赶快摆脱黄褐斑的困扰�� �找到自己白皙靓丽的一面。 阅读了怎么样去脸上的色斑,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 怎么样去脸上的色斑,同时为您分享祛斑小方法 去斑方法,胡萝卜汁:2汤匙汁加入10滴柠檬汁,每天早、晚各� ��一次,20到30分钟后洗掉, 涂抹美白护肤霜。 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 2:56
1.0
解密怎么样去脸上的色斑 - ``` 《摘要》 一直不敢碰触荒凉这个词语,我怕在这寒冷的冬天,被荒凉�� �个词紧紧包裹,阻挡我找寻春天的温暖。可是,你看,真的� ��满眼的灰黄和萧瑟。我不得不站在季节的末梢,看枯叶飘零 ,残阳无力,寒风肆虐。此时,一种繁华落尽的惆怅沧桑着�� �言。原来,世间无论多么盛大的甜美都会有失去的那一天,� ��经的繁华已然远去,无法找寻一丝痕迹。铅华褪尽的世界是 如此的苍凉和晦暗。总是回忆起你,你的笑你的温柔,还有�� �临别时的眼神,那是在嘲笑我把,是啊脸上的色斑已经提醒� ��你已经远去!怎么样去脸上的色斑, 《客户案例》   人为什么长黄褐斑, 该怎么祛斑啊,这个问题也是伴随了我很长时间。记得上大�� �之前,我的肤质本来还是不错的,白白的,可是自从经过开� ��后一段时间的军训,皮肤就变得越来越不好了,不光是晒得 很黑,脸色还有些发黄,鼻子两翼竟然还长出许多的斑斑点�� �,一下子好像老了许多一样。我很是郁闷,军训一结束,我� ��开始修护保养自己的皮肤。过不久,我的皮肤有回复了往日 的白皙,可是脸上的斑却还在那里,因为皮肤的白反而更明�� �了。于是我就开始了作战!与斑点战斗到底!在此期间,我尝�� �了各种祛斑方法,都说效果有多好多好,可结果就是我脸上� ��斑点依旧在那,有的是消失了但是又反弹了。那时真的是很 失望,都快失去信心了。</br>   我把这个烦恼告诉了我一个同事,我同事当时听完,就�� �我说有一款叫黛芙薇尔精华液的产品,美白祛斑功效特别好� ��她家亲戚就一直在使用,听了她的话,我立刻上网搜索黛芙 薇尔精华液,进入到官网,发现很多女性朋友,用了他们的�� �品后,肌肤长斑情况不但得到了改善,而且肌肤比以前还白� ��。这正是我想要的怎样去除脸上的黄褐斑的产品。通过和客 服的沟通,我订购了1个周期的黛芙薇尔精华液。当产品到手� ��,我就立刻使用。</br>   刚开始的两个周期效果不明显,我就咨询了黛芙薇尔精�� �液的客服,她们说:“黛芙薇尔精华液只从根本上治疗黄褐� ��,前两个周期主要是DNA修复阶段,后两个周期就会从皮肤表 面表现出来”。我想这和中国的医理是一样的。后两个周期�� �效果你们就可想而知了。我的黄褐斑真的“下岗”了,皮肤� ��净的像出生的婴儿肤一样。哇,没想到天然精华祛斑产品黛 芙薇尔精华液如此的神奇,使我打败了“顽固的黄褐斑”。�� �道寻找真正能祛斑的产品不易,想到自己也是看了别人的文� ��才能去除黄褐斑,因此我也将自己的祛斑经历写下来发表, 希望和我一样有祛斑痛苦经历的人能赶快摆脱黄褐斑的困扰�� �找到自己白皙靓丽的一面。 阅读了怎么样去脸上的色斑,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 怎么样去脸上的色斑,同时为您分享祛斑小方法 去斑方法,胡萝卜汁:2汤匙汁加入10滴柠檬汁,每天早、晚各� ��一次,20到30分钟后洗掉, 涂抹美白护肤霜。 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 2:56
non_priority
解密怎么样去脸上的色斑 《摘要》 一直不敢碰触荒凉这个词语,我怕在这寒冷的冬天,被荒凉�� �个词紧紧包裹,阻挡我找寻春天的温暖。可是,你看,真的� ��满眼的灰黄和萧瑟。我不得不站在季节的末梢,看枯叶飘零 ,残阳无力,寒风肆虐。此时,一种繁华落尽的惆怅沧桑着�� �言。原来,世间无论多么盛大的甜美都会有失去的那一天,� ��经的繁华已然远去,无法找寻一丝痕迹。铅华褪尽的世界是 如此的苍凉和晦暗。总是回忆起你,你的笑你的温柔,还有�� �临别时的眼神,那是在嘲笑我把,是啊脸上的色斑已经提醒� ��你已经远去!怎么样去脸上的色斑, 《客户案例》   人为什么长黄褐斑 该怎么祛斑啊,这个问题也是伴随了我很长时间。记得上大�� �之前,我的肤质本来还是不错的,白白的,可是自从经过开� ��后一段时间的军训,皮肤就变得越来越不好了,不光是晒得 很黑,脸色还有些发黄,鼻子两翼竟然还长出许多的斑斑点�� �,一下子好像老了许多一样。我很是郁闷,军训一结束,我� ��开始修护保养自己的皮肤。过不久,我的皮肤有回复了往日 的白皙,可是脸上的斑却还在那里,因为皮肤的白反而更明�� �了。于是我就开始了作战 与斑点战斗到底 在此期间,我尝�� �了各种祛斑方法,都说效果有多好多好,可结果就是我脸上� ��斑点依旧在那,有的是消失了但是又反弹了。那时真的是很 失望,都快失去信心了。   我把这个烦恼告诉了我一个同事,我同事当时听完,就�� �我说有一款叫黛芙薇尔精华液的产品,美白祛斑功效特别好� ��她家亲戚就一直在使用,听了她的话,我立刻上网搜索黛芙 薇尔精华液,进入到官网,发现很多女性朋友,用了他们的�� �品后,肌肤长斑情况不但得到了改善,而且肌肤比以前还白� ��。这正是我想要的怎样去除脸上的黄褐斑的产品。通过和客 服的沟通, 。当产品到手� ��,我就立刻使用。   刚开始的两个周期效果不明显,我就咨询了黛芙薇尔精�� �液的客服,她们说:“黛芙薇尔精华液只从根本上治疗黄褐� ��,前两个周期主要是dna修复阶段,后两个周期就会从皮肤表 面表现出来”。我想这和中国的医理是一样的。后两个周期�� �效果你们就可想而知了。我的黄褐斑真的“下岗”了,皮肤� ��净的像出生的婴儿肤一样。哇,没想到天然精华祛斑产品黛 芙薇尔精华液如此的神奇,使我打败了“顽固的黄褐斑”。�� �道寻找真正能祛斑的产品不易,想到自己也是看了别人的文� ��才能去除黄褐斑,因此我也将自己的祛斑经历写下来发表, 希望和我一样有祛斑痛苦经历的人能赶快摆脱黄褐斑的困扰�� �找到自己白皙靓丽的一面。 阅读了怎么样去脸上的色斑,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》    黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗   答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来    ,服用黛芙薇尔美白,会伤身体吗 有副作用吗   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖    ,去除黄褐斑之后,会反弹吗   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗    ,你们的价格有点贵,能不能便宜一点   答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗    ,我适合用黛芙薇尔精华液吗   答:黛芙薇尔适用人群:    、生理紊乱引起的黄褐斑人群    、生育引起的妊娠斑人群    、年纪增长引起的老年斑人群    、化妆品色素沉积、辐射斑人群    、长期日照引起的日晒斑人群    、肌肤暗淡急需美白的人群 《祛斑小方法》 怎么样去脸上的色斑,同时为您分享祛斑小方法 去斑方法 胡萝卜汁: ,每天早、晚各� ��一次, , 涂抹美白护肤霜。 original issue reported on code google com by additive gmail com on jul at
0
237,146
7,756,624,852
IssuesEvent
2018-05-31 14:08:34
zephyrproject-rtos/
https://api.github.com/repos/zephyrproject-rtos/
closed
[Coverity CID :186035] Incorrect expression in /drivers/ipm/ipm_mcux.c
Coverity area: Drivers bug priority: medium
Static code scan issues seen in File: /drivers/ipm/ipm_mcux.c Category: Incorrect expression Function: mcux_mailbox_isr Component: Drivers CID: 186035 Please fix or provide comments to square it off in coverity in the link: https://scan9.coverity.com/reports.htm#v32951/p12996
1.0
[Coverity CID :186035] Incorrect expression in /drivers/ipm/ipm_mcux.c - Static code scan issues seen in File: /drivers/ipm/ipm_mcux.c Category: Incorrect expression Function: mcux_mailbox_isr Component: Drivers CID: 186035 Please fix or provide comments to square it off in coverity in the link: https://scan9.coverity.com/reports.htm#v32951/p12996
priority
incorrect expression in drivers ipm ipm mcux c static code scan issues seen in file drivers ipm ipm mcux c category incorrect expression function mcux mailbox isr component drivers cid please fix or provide comments to square it off in coverity in the link
1
16,881
10,576,282,600
IssuesEvent
2019-10-07 17:31:28
cityofaustin/atd-vz-data
https://api.github.com/repos/cityofaustin/atd-vz-data
opened
Bug: Search breaks if you give a type mismatch
MVP Project: Vision Zero Crash Data System Service: Dev Type: Bug Report Workgroup: VZ
Reproduce by putting a string entry into the search field with the field type dropdown having "Crash ID" selected ![Screen Shot 2019-10-07 at 12 27 15 PM](https://user-images.githubusercontent.com/5697474/66334147-16996680-e8fe-11e9-86ef-d9c209b31691.png) ![Screen Shot 2019-10-07 at 12 27 20 PM](https://user-images.githubusercontent.com/5697474/66334158-1a2ced80-e8fe-11e9-8e88-2c85a2fe8451.png)
1.0
Bug: Search breaks if you give a type mismatch - Reproduce by putting a string entry into the search field with the field type dropdown having "Crash ID" selected ![Screen Shot 2019-10-07 at 12 27 15 PM](https://user-images.githubusercontent.com/5697474/66334147-16996680-e8fe-11e9-86ef-d9c209b31691.png) ![Screen Shot 2019-10-07 at 12 27 20 PM](https://user-images.githubusercontent.com/5697474/66334158-1a2ced80-e8fe-11e9-8e88-2c85a2fe8451.png)
non_priority
bug search breaks if you give a type mismatch reproduce by putting a string entry into the search field with the field type dropdown having crash id selected
0
131,573
12,486,821,101
IssuesEvent
2020-05-31 05:03:50
zacaz90/Tutorial10_Group-1
https://api.github.com/repos/zacaz90/Tutorial10_Group-1
closed
Possible effects if project fails
5 H documentation
E.g. Potential benefits to the call center (in terms of reducing cost and being able to carry out more effective calls). 500 words (including competitive advantages)
1.0
Possible effects if project fails - E.g. Potential benefits to the call center (in terms of reducing cost and being able to carry out more effective calls). 500 words (including competitive advantages)
non_priority
possible effects if project fails e g potential benefits to the call center in terms of reducing cost and being able to carry out more effective calls words including competitive advantages
0
247,591
7,920,365,309
IssuesEvent
2018-07-04 23:43:27
Sinapse-Energia/AP-Sinapse
https://api.github.com/repos/Sinapse-Energia/AP-Sinapse
opened
[BUG][RF TX - RX] [BROADCAST / GROUPS] Acting over groups produces reconnections
Device: EN Priority: high Size: 5 Status: new Type: bug
It has been observed a lot of AP re-connections just after acting over groups, when there are a lot of EPDs with this group For example: Topic: ID_INST/IMEI/ACT/FFFFFA Msg: 3;60; Why? Any idea? - The answer bit is 0 when a message like this is sent, right? - Maybe we need to set the transit bit also to 0?
1.0
[BUG][RF TX - RX] [BROADCAST / GROUPS] Acting over groups produces reconnections - It has been observed a lot of AP re-connections just after acting over groups, when there are a lot of EPDs with this group For example: Topic: ID_INST/IMEI/ACT/FFFFFA Msg: 3;60; Why? Any idea? - The answer bit is 0 when a message like this is sent, right? - Maybe we need to set the transit bit also to 0?
priority
acting over groups produces reconnections it has been observed a lot of ap re connections just after acting over groups when there are a lot of epds with this group for example topic id inst imei act fffffa msg why any idea the answer bit is when a message like this is sent right maybe we need to set the transit bit also to
1
217,584
7,324,968,741
IssuesEvent
2018-03-03 02:39:54
NCEAS/metacat
https://api.github.com/repos/NCEAS/metacat
closed
Extend Metacat Interface and Client
Category: metacat Component: Bugzilla-Id Priority: Normal Status: Closed Tracker: Bug
--- Author Name: **Duane Costa** (Duane Costa) Original Redmine Issue: 1217, https://projects.ecoinformatics.org/ecoinfo/issues/1217 Original Date: 2003-11-21 Original Assignee: Duane Costa --- The Metacat interface, implemented by the MetacatClient class, exposes a number of core actions of the MetacatServlet as HTTP requests to the servlet. The current list of supported actions includes: login logout read squery insert update delete A number of additional actions, most of which are not critical but are convenient, are not supported by the interface: query export readinlinedata validate setaccess getaccesscontrol getprincipals getdoctypes getdtdschema getdataguide getlastdocid getrevisionanddoctype (Note: some of the methods listed above are obsolete and therefore should not be included in the extended interface. The complete list of methods that should be added to the interface is yet to be determined.) Some of the methods in the extended set are useful to other clients, such as Harvester. However, Harvester should interact exclusively with the MetacatClient, rather than directly with the Metacat servlet. Therefore, the current interface should be extended with the additional convenience methods so that they can be used by Harvester and potentially other clients. There are two ways that the current interface could be extended: (1) Add a new interface, 'MetacatExtendedInterface.java', to hold the definitions for the convenience methods. The MetacatClient class would implement both the current Metacat interface and the MetacatExtendedInterface. The advantage of this approach is that the current Metacat interface can be kept simple in that it will only define the core methods. (2) Add the new methods to the existing Metacat interface. The advantage of just adding these methods to Metacat.java is programmatically it's easier because you only have to cast to one interface (not two) to make method calls. We will determine in the next few days which of these two approaches to use. A secondary goal of this task is to improve the documentation for the methods in the extended interface. A future goal (outside the scope of this task) is to refactor Morpho and other code to use the MetacatClient to interact with MetacatServlet.
1.0
Extend Metacat Interface and Client - --- Author Name: **Duane Costa** (Duane Costa) Original Redmine Issue: 1217, https://projects.ecoinformatics.org/ecoinfo/issues/1217 Original Date: 2003-11-21 Original Assignee: Duane Costa --- The Metacat interface, implemented by the MetacatClient class, exposes a number of core actions of the MetacatServlet as HTTP requests to the servlet. The current list of supported actions includes: login logout read squery insert update delete A number of additional actions, most of which are not critical but are convenient, are not supported by the interface: query export readinlinedata validate setaccess getaccesscontrol getprincipals getdoctypes getdtdschema getdataguide getlastdocid getrevisionanddoctype (Note: some of the methods listed above are obsolete and therefore should not be included in the extended interface. The complete list of methods that should be added to the interface is yet to be determined.) Some of the methods in the extended set are useful to other clients, such as Harvester. However, Harvester should interact exclusively with the MetacatClient, rather than directly with the Metacat servlet. Therefore, the current interface should be extended with the additional convenience methods so that they can be used by Harvester and potentially other clients. There are two ways that the current interface could be extended: (1) Add a new interface, 'MetacatExtendedInterface.java', to hold the definitions for the convenience methods. The MetacatClient class would implement both the current Metacat interface and the MetacatExtendedInterface. The advantage of this approach is that the current Metacat interface can be kept simple in that it will only define the core methods. (2) Add the new methods to the existing Metacat interface. The advantage of just adding these methods to Metacat.java is programmatically it's easier because you only have to cast to one interface (not two) to make method calls. We will determine in the next few days which of these two approaches to use. A secondary goal of this task is to improve the documentation for the methods in the extended interface. A future goal (outside the scope of this task) is to refactor Morpho and other code to use the MetacatClient to interact with MetacatServlet.
priority
extend metacat interface and client author name duane costa duane costa original redmine issue original date original assignee duane costa the metacat interface implemented by the metacatclient class exposes a number of core actions of the metacatservlet as http requests to the servlet the current list of supported actions includes login logout read squery insert update delete a number of additional actions most of which are not critical but are convenient are not supported by the interface query export readinlinedata validate setaccess getaccesscontrol getprincipals getdoctypes getdtdschema getdataguide getlastdocid getrevisionanddoctype note some of the methods listed above are obsolete and therefore should not be included in the extended interface the complete list of methods that should be added to the interface is yet to be determined some of the methods in the extended set are useful to other clients such as harvester however harvester should interact exclusively with the metacatclient rather than directly with the metacat servlet therefore the current interface should be extended with the additional convenience methods so that they can be used by harvester and potentially other clients there are two ways that the current interface could be extended add a new interface metacatextendedinterface java to hold the definitions for the convenience methods the metacatclient class would implement both the current metacat interface and the metacatextendedinterface the advantage of this approach is that the current metacat interface can be kept simple in that it will only define the core methods add the new methods to the existing metacat interface the advantage of just adding these methods to metacat java is programmatically it s easier because you only have to cast to one interface not two to make method calls we will determine in the next few days which of these two approaches to use a secondary goal of this task is to improve the documentation for the methods in the extended interface a future goal outside the scope of this task is to refactor morpho and other code to use the metacatclient to interact with metacatservlet
1
57,960
8,215,766,987
IssuesEvent
2018-09-05 07:04:34
python-cmd2/cmd2
https://api.github.com/repos/python-cmd2/cmd2
closed
Unrestrict Sphinx version once their bug is fixed
documentation
Right now we are temporarily restricting the version of Sphinx used to run our documentation unit test due to the following Sphinx bug: https://github.com/sphinx-doc/sphinx/issues/5322 Once that bug is fixed, we should now longer restrict what version of Sphinx we are testing with.
1.0
Unrestrict Sphinx version once their bug is fixed - Right now we are temporarily restricting the version of Sphinx used to run our documentation unit test due to the following Sphinx bug: https://github.com/sphinx-doc/sphinx/issues/5322 Once that bug is fixed, we should now longer restrict what version of Sphinx we are testing with.
non_priority
unrestrict sphinx version once their bug is fixed right now we are temporarily restricting the version of sphinx used to run our documentation unit test due to the following sphinx bug once that bug is fixed we should now longer restrict what version of sphinx we are testing with
0
297,079
22,338,612,387
IssuesEvent
2022-06-14 21:14:09
aws-amplify/amplify-cli
https://api.github.com/repos/aws-amplify/amplify-cli
closed
amplify configure silently fails
documentation pending-close-response-required installation duplicate
### Before opening, please confirm: - [X] I have installed the latest version of the Amplify CLI (see above), and confirmed that the issue still persists. - [X] I have [searched for duplicate or closed issues](https://github.com/aws-amplify/amplify-cli/issues?q=is%3Aissue+). - [X] I have read the guide for [submitting bug reports](https://github.com/aws-amplify/amplify-cli/blob/master/CONTRIBUTING.md#bug-reports). - [X] I have done my best to include a minimal, self-contained set of instructions for consistently reproducing the issue. - [X] I have removed any sensitive information from my code snippets and submission. ### How did you install the Amplify CLI? npm install -g @aws-amplify/cli ### If applicable, what version of Node.js are you using? v16.15.1 ### Amplify CLI Version 8.5.0 ### What operating system are you using? MacOS Monterey ### Did you make any manual changes to the cloud resources managed by Amplify? Please describe the changes made. No ### Amplify Categories Not applicable ### Amplify Commands configure ### Describe the bug `amplify configure` or any combination of any other `amplify` command silently fails with no output whatsover. Within existing project: <img width="475" alt="image" src="https://user-images.githubusercontent.com/6605287/173685135-84c39670-31ac-47d5-a5d5-28da9e129e5c.png"> In any directory: <img width="475" alt="image" src="https://user-images.githubusercontent.com/6605287/173685225-200e6d54-f93d-418e-8d4c-35ed8f70e8e8.png"> <img width="475" alt="image" src="https://user-images.githubusercontent.com/6605287/173685243-1e151979-9e92-4e3e-b23d-2e51e2e111b8.png"> ### Expected behavior The cli should function as documented. ### Reproduction steps I am using a 2022 Macbook Pro base configuration 16" with M1 processor. Running native zsh shell and node / cli versions listed above. 1. Install node `brew install node@16` 2. Install cli `npm install -g @aws-amplify/cli` 3. Try above commands and observe the behaviour ### GraphQL schema(s) _No response_ ### Log output _No response_ ### Additional information _No response_
1.0
amplify configure silently fails - ### Before opening, please confirm: - [X] I have installed the latest version of the Amplify CLI (see above), and confirmed that the issue still persists. - [X] I have [searched for duplicate or closed issues](https://github.com/aws-amplify/amplify-cli/issues?q=is%3Aissue+). - [X] I have read the guide for [submitting bug reports](https://github.com/aws-amplify/amplify-cli/blob/master/CONTRIBUTING.md#bug-reports). - [X] I have done my best to include a minimal, self-contained set of instructions for consistently reproducing the issue. - [X] I have removed any sensitive information from my code snippets and submission. ### How did you install the Amplify CLI? npm install -g @aws-amplify/cli ### If applicable, what version of Node.js are you using? v16.15.1 ### Amplify CLI Version 8.5.0 ### What operating system are you using? MacOS Monterey ### Did you make any manual changes to the cloud resources managed by Amplify? Please describe the changes made. No ### Amplify Categories Not applicable ### Amplify Commands configure ### Describe the bug `amplify configure` or any combination of any other `amplify` command silently fails with no output whatsover. Within existing project: <img width="475" alt="image" src="https://user-images.githubusercontent.com/6605287/173685135-84c39670-31ac-47d5-a5d5-28da9e129e5c.png"> In any directory: <img width="475" alt="image" src="https://user-images.githubusercontent.com/6605287/173685225-200e6d54-f93d-418e-8d4c-35ed8f70e8e8.png"> <img width="475" alt="image" src="https://user-images.githubusercontent.com/6605287/173685243-1e151979-9e92-4e3e-b23d-2e51e2e111b8.png"> ### Expected behavior The cli should function as documented. ### Reproduction steps I am using a 2022 Macbook Pro base configuration 16" with M1 processor. Running native zsh shell and node / cli versions listed above. 1. Install node `brew install node@16` 2. Install cli `npm install -g @aws-amplify/cli` 3. Try above commands and observe the behaviour ### GraphQL schema(s) _No response_ ### Log output _No response_ ### Additional information _No response_
non_priority
amplify configure silently fails before opening please confirm i have installed the latest version of the amplify cli see above and confirmed that the issue still persists i have i have read the guide for i have done my best to include a minimal self contained set of instructions for consistently reproducing the issue i have removed any sensitive information from my code snippets and submission how did you install the amplify cli npm install g aws amplify cli if applicable what version of node js are you using amplify cli version what operating system are you using macos monterey did you make any manual changes to the cloud resources managed by amplify please describe the changes made no amplify categories not applicable amplify commands configure describe the bug amplify configure or any combination of any other amplify command silently fails with no output whatsover within existing project img width alt image src in any directory img width alt image src img width alt image src expected behavior the cli should function as documented reproduction steps i am using a macbook pro base configuration with processor running native zsh shell and node cli versions listed above install node brew install node install cli npm install g aws amplify cli try above commands and observe the behaviour graphql schema s no response log output no response additional information no response
0
784,819
27,585,653,600
IssuesEvent
2023-03-08 19:32:47
2DGD-F0TH/2DGD_F0TH
https://api.github.com/repos/2DGD-F0TH/2DGD_F0TH
closed
Tracking Issue: Example images
enhancement good first issue Priority 2 Content
I noticed that in certain parts, the book becomes really text-heavy, without many visual mediums to help break up the content and understand things better. Here are some ideas of images to add: - [x] Off-screen objects optimization - [x] Resource pools (instantiation, pull, return) - [x] Lookup tables - [x] Memoization - [x] Stretching/Rotating/Shearing matrices (examples with a square) - [x] Console/PC/Mobile/Web example images (maybe just as decoration?) - [x] Examples of 1/2/4/8/16 bit colors (more colors are useless) (resources) - [x] Images with less details on the background (resources) - [x] Images with muted colors in the background (resources) - [x] Elements contrasting with background (resources)
1.0
Tracking Issue: Example images - I noticed that in certain parts, the book becomes really text-heavy, without many visual mediums to help break up the content and understand things better. Here are some ideas of images to add: - [x] Off-screen objects optimization - [x] Resource pools (instantiation, pull, return) - [x] Lookup tables - [x] Memoization - [x] Stretching/Rotating/Shearing matrices (examples with a square) - [x] Console/PC/Mobile/Web example images (maybe just as decoration?) - [x] Examples of 1/2/4/8/16 bit colors (more colors are useless) (resources) - [x] Images with less details on the background (resources) - [x] Images with muted colors in the background (resources) - [x] Elements contrasting with background (resources)
priority
tracking issue example images i noticed that in certain parts the book becomes really text heavy without many visual mediums to help break up the content and understand things better here are some ideas of images to add off screen objects optimization resource pools instantiation pull return lookup tables memoization stretching rotating shearing matrices examples with a square console pc mobile web example images maybe just as decoration examples of bit colors more colors are useless resources images with less details on the background resources images with muted colors in the background resources elements contrasting with background resources
1
440,018
12,692,156,195
IssuesEvent
2020-06-21 20:50:12
microwavekonijn/ps2census
https://api.github.com/repos/microwavekonijn/ps2census
opened
Rebalance default DuplicateFilter
bug priority
The hashes generated internally require some attention to reduce the amount of false positives.
1.0
Rebalance default DuplicateFilter - The hashes generated internally require some attention to reduce the amount of false positives.
priority
rebalance default duplicatefilter the hashes generated internally require some attention to reduce the amount of false positives
1