Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
49,087
6,008,707,275
IssuesEvent
2017-06-06 08:38:02
owncloud/client
https://api.github.com/repos/owncloud/client
closed
Client asks for account password after certificate is rejected
bug ReadyToTest
When a SSL server certificate is presented to the user and rejected the client will not be able to communicate with that server. However, the client will still pop up a dialog asking for the user's password *after* the certificate was rejected. That is pointless and confusing. Steps: 1. Have a server with a self-signed certificate (or mitmproxy one) 2. Set up an account, log out of it, quit the client. 3. Remove the certificate authorization from the config file. (or restart mitmproxy) 4. Start the client. 5. Reject the server's certificate. 6. Notice that the "Please enter your password" dialog pops up for the account. @guruz
1.0
Client asks for account password after certificate is rejected - When a SSL server certificate is presented to the user and rejected the client will not be able to communicate with that server. However, the client will still pop up a dialog asking for the user's password *after* the certificate was rejected. That is pointless and confusing. Steps: 1. Have a server with a self-signed certificate (or mitmproxy one) 2. Set up an account, log out of it, quit the client. 3. Remove the certificate authorization from the config file. (or restart mitmproxy) 4. Start the client. 5. Reject the server's certificate. 6. Notice that the "Please enter your password" dialog pops up for the account. @guruz
test
client asks for account password after certificate is rejected when a ssl server certificate is presented to the user and rejected the client will not be able to communicate with that server however the client will still pop up a dialog asking for the user s password after the certificate was rejected that is pointless and confusing steps have a server with a self signed certificate or mitmproxy one set up an account log out of it quit the client remove the certificate authorization from the config file or restart mitmproxy start the client reject the server s certificate notice that the please enter your password dialog pops up for the account guruz
1
409,035
11,955,756,517
IssuesEvent
2020-04-04 06:39:38
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
closed
Synthesis failed for Google.Cloud.PubSub.V1
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate Google.Cloud.PubSub.V1. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1' Cloning into '/tmpfs/tmp/tmporf5fmop/googleapis'... Note: checking out '9d4a3ad084027c414a5670ff358a65917c81cdeb'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 9d4a3ad08 chore: set Ruby namespace in proto options Note: checking out 'd6cb4997910eda04c0c66c0f2fd043eeaa0f660d'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at d6cb4997 chore: enable gapic v2 and proto annotation for documentai API. Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-60' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:48:35,484 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index Building new version of Google.Cloud.PubSub.V1 for compatibility checking Changes in Google.Cloud.PubSub.V1: Diff level: Identical 2020-04-03 22:48:50,918 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/Google.Cloud.PubSub.V1/Pubsub.cs M apis/Google.Cloud.PubSub.V1/synth.metadata [autosynth-Google.Cloud.PubSub.V1-60 b9f5bffbe] ignored 2 files changed, 4 insertions(+), 3 deletions(-) HEAD is now at b9f5bffbe ignored Switched to branch 'autosynth-Google.Cloud.PubSub.V1' Previous HEAD position was d6cb4997 chore: enable gapic v2 and proto annotation for documentai API. HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes Note: checking out '98024617efce32982bd763ad14f00c9bc0819bea'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 98024617e Touch all the synth.metadata files again Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-0' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:48:51,462 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking 2020-04-03 22:48:57,493 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/synth.metadata HEAD is now at 98024617e Touch all the synth.metadata files again Switched to branch 'autosynth-Google.Cloud.PubSub.V1' Note: checking out '9d4a3ad084027c414a5670ff358a65917c81cdeb'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 9d4a3ad08 chore: set Ruby namespace in proto options Previous HEAD position was 7be2811a fix: Update gapic-generator version to pickup discogapic fixes HEAD is now at 17cfae00 Add a new AuthorizationType for Data Source Definition. Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-30' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:48:58,081 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking 2020-04-03 22:49:04,186 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/synth.metadata HEAD is now at 9d4a3ad08 chore: set Ruby namespace in proto options Switched to branch 'autosynth-Google.Cloud.PubSub.V1' Previous HEAD position was 17cfae00 Add a new AuthorizationType for Data Source Definition. HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes Note: checking out '5a41fb5f1c7f5329cc981b77cae1f4762d705002'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 5a41fb5f1 Dialogflow weekly v2 library update Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-15' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:49:04,714 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking 2020-04-03 22:49:10,840 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/synth.metadata HEAD is now at 5a41fb5f1 Dialogflow weekly v2 library update Switched to branch 'autosynth-Google.Cloud.PubSub.V1' HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes Note: checking out 'edebc2b7a22574b76412a9c1cf6832719b9c85e8'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at edebc2b7a fix Dataproc: add missing `REQUIRED` annotation. Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-7' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:49:11,371 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking 2020-04-03 22:49:17,464 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/synth.metadata HEAD is now at edebc2b7a fix Dataproc: add missing `REQUIRED` annotation. Switched to branch 'autosynth-Google.Cloud.PubSub.V1' HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes Note: checking out '38fb38aeb841fb08e12ff366c2159fb6669b45d8'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 38fb38aeb Touch synth.metadata to get SecretManager APIs building Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-3' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:49:18,004 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking 2020-04-03 22:49:24,391 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/synth.metadata HEAD is now at 38fb38aeb Touch synth.metadata to get SecretManager APIs building Switched to branch 'autosynth-Google.Cloud.PubSub.V1' HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes Note: checking out '0c88ce0da0e66910735643832b237c9f873b05e6'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 0c88ce0da Add a new AuthorizationType for Data Source Definition. Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-1' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:49:24,939 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking 2020-04-03 22:49:30,979 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/synth.metadata HEAD is now at 0c88ce0da Add a new AuthorizationType for Data Source Definition. Switched to branch 'autosynth-Google.Cloud.PubSub.V1' On branch autosynth-Google.Cloud.PubSub.V1 nothing to commit, working tree clean Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 484, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 370, in main return _inner_main(temp_dir) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 474, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 284, in synthesize_loop synthesize_range(toolbox, synthesizer) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 306, in synthesize_range toolbox.patch_merge_version(young) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 146, in patch_merge_version comment or self.versions[index].version.get_comment() File "/tmpfs/src/git/autosynth/autosynth/git.py", line 95, in commit_all_changes subprocess.check_call(["git", "commit", "-m", message]) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['git', 'commit', '-m', 'Add a new AuthorizationType for Data Source Definition.\n\nhttps://github.com/googleapis/google-cloud-dotnet/commit/0c88ce0da0e66910735643832b237c9f873b05e6\ncommit 0c88ce0da0e66910735643832b237c9f873b05e6\nAuthor: yoshi-automation <yoshi-automation@google.com>\nDate: Wed Apr 1 02:27:58 2020 -0700\n\n Add a new AuthorizationType for Data Source Definition.\n \n https://github.com/googleapis/googleapis/commit/17cfae00f2bb51cb1683f017da7e295a1b0f01a8\n commit 17cfae00f2bb51cb1683f017da7e295a1b0f01a8\n Author: Google APIs <noreply@google.com>\n Date: Tue Mar 31 10:21:11 2020 -0700\n \n Add a new AuthorizationType for Data Source Definition.\n \n PiperOrigin-RevId: 303992863']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](https://sponge/89ee1c7b-40f0-4f7a-a7cd-f68bdd6f31c1).
1.0
Synthesis failed for Google.Cloud.PubSub.V1 - Hello! Autosynth couldn't regenerate Google.Cloud.PubSub.V1. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1' Cloning into '/tmpfs/tmp/tmporf5fmop/googleapis'... Note: checking out '9d4a3ad084027c414a5670ff358a65917c81cdeb'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 9d4a3ad08 chore: set Ruby namespace in proto options Note: checking out 'd6cb4997910eda04c0c66c0f2fd043eeaa0f660d'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at d6cb4997 chore: enable gapic v2 and proto annotation for documentai API. Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-60' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:48:35,484 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index Building new version of Google.Cloud.PubSub.V1 for compatibility checking Changes in Google.Cloud.PubSub.V1: Diff level: Identical 2020-04-03 22:48:50,918 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/Google.Cloud.PubSub.V1/Pubsub.cs M apis/Google.Cloud.PubSub.V1/synth.metadata [autosynth-Google.Cloud.PubSub.V1-60 b9f5bffbe] ignored 2 files changed, 4 insertions(+), 3 deletions(-) HEAD is now at b9f5bffbe ignored Switched to branch 'autosynth-Google.Cloud.PubSub.V1' Previous HEAD position was d6cb4997 chore: enable gapic v2 and proto annotation for documentai API. HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes Note: checking out '98024617efce32982bd763ad14f00c9bc0819bea'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 98024617e Touch all the synth.metadata files again Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-0' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:48:51,462 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking 2020-04-03 22:48:57,493 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/synth.metadata HEAD is now at 98024617e Touch all the synth.metadata files again Switched to branch 'autosynth-Google.Cloud.PubSub.V1' Note: checking out '9d4a3ad084027c414a5670ff358a65917c81cdeb'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 9d4a3ad08 chore: set Ruby namespace in proto options Previous HEAD position was 7be2811a fix: Update gapic-generator version to pickup discogapic fixes HEAD is now at 17cfae00 Add a new AuthorizationType for Data Source Definition. Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-30' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:48:58,081 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking 2020-04-03 22:49:04,186 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/synth.metadata HEAD is now at 9d4a3ad08 chore: set Ruby namespace in proto options Switched to branch 'autosynth-Google.Cloud.PubSub.V1' Previous HEAD position was 17cfae00 Add a new AuthorizationType for Data Source Definition. HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes Note: checking out '5a41fb5f1c7f5329cc981b77cae1f4762d705002'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 5a41fb5f1 Dialogflow weekly v2 library update Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-15' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:49:04,714 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking 2020-04-03 22:49:10,840 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/synth.metadata HEAD is now at 5a41fb5f1 Dialogflow weekly v2 library update Switched to branch 'autosynth-Google.Cloud.PubSub.V1' HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes Note: checking out 'edebc2b7a22574b76412a9c1cf6832719b9c85e8'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at edebc2b7a fix Dataproc: add missing `REQUIRED` annotation. Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-7' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:49:11,371 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking 2020-04-03 22:49:17,464 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/synth.metadata HEAD is now at edebc2b7a fix Dataproc: add missing `REQUIRED` annotation. Switched to branch 'autosynth-Google.Cloud.PubSub.V1' HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes Note: checking out '38fb38aeb841fb08e12ff366c2159fb6669b45d8'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 38fb38aeb Touch synth.metadata to get SecretManager APIs building Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-3' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:49:18,004 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking 2020-04-03 22:49:24,391 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/synth.metadata HEAD is now at 38fb38aeb Touch synth.metadata to get SecretManager APIs building Switched to branch 'autosynth-Google.Cloud.PubSub.V1' HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes Note: checking out '0c88ce0da0e66910735643832b237c9f873b05e6'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 0c88ce0da Add a new AuthorizationType for Data Source Definition. Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-1' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-04-03 22:49:24,939 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py. Skipping microgenerator fetch/build: already built, and running on Kokoro Building existing version of Google.Cloud.PubSub.V1 for compatibility checking Generating Google.Cloud.PubSub.V1 Running mid-micro-generation script for Google.Cloud.PubSub.V1 Running post-generation script for Google.Cloud.PubSub.V1 Updated 1 path from the index Updated 1 path from the index git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking 2020-04-03 22:49:30,979 synthtool > Wrote metadata to synth.metadata. Changed files: M apis/Google.Cloud.PubSub.V1/synth.metadata HEAD is now at 0c88ce0da Add a new AuthorizationType for Data Source Definition. Switched to branch 'autosynth-Google.Cloud.PubSub.V1' On branch autosynth-Google.Cloud.PubSub.V1 nothing to commit, working tree clean Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 484, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 370, in main return _inner_main(temp_dir) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 474, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 284, in synthesize_loop synthesize_range(toolbox, synthesizer) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 306, in synthesize_range toolbox.patch_merge_version(young) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 146, in patch_merge_version comment or self.versions[index].version.get_comment() File "/tmpfs/src/git/autosynth/autosynth/git.py", line 95, in commit_all_changes subprocess.check_call(["git", "commit", "-m", message]) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['git', 'commit', '-m', 'Add a new AuthorizationType for Data Source Definition.\n\nhttps://github.com/googleapis/google-cloud-dotnet/commit/0c88ce0da0e66910735643832b237c9f873b05e6\ncommit 0c88ce0da0e66910735643832b237c9f873b05e6\nAuthor: yoshi-automation <yoshi-automation@google.com>\nDate: Wed Apr 1 02:27:58 2020 -0700\n\n Add a new AuthorizationType for Data Source Definition.\n \n https://github.com/googleapis/googleapis/commit/17cfae00f2bb51cb1683f017da7e295a1b0f01a8\n commit 17cfae00f2bb51cb1683f017da7e295a1b0f01a8\n Author: Google APIs <noreply@google.com>\n Date: Tue Mar 31 10:21:11 2020 -0700\n \n Add a new AuthorizationType for Data Source Definition.\n \n PiperOrigin-RevId: 303992863']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](https://sponge/89ee1c7b-40f0-4f7a-a7cd-f68bdd6f31c1).
non_test
synthesis failed for google cloud pubsub hello autosynth couldn t regenerate google cloud pubsub broken heart here s the output from running synth py cloning into working repo switched to a new branch autosynth google cloud pubsub cloning into tmpfs tmp googleapis note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at chore set ruby namespace in proto options note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at chore enable gapic and proto annotation for documentai api switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index building new version of google cloud pubsub for compatibility checking changes in google cloud pubsub diff level identical synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub google cloud pubsub pubsub cs m apis google cloud pubsub synth metadata ignored files changed insertions deletions head is now at ignored switched to branch autosynth google cloud pubsub previous head position was chore enable gapic and proto annotation for documentai api head is now at fix update gapic generator version to pickup discogapic fixes note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at touch all the synth metadata files again switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index git detects no change in google cloud pubsub skipping compatibility checking synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub synth metadata head is now at touch all the synth metadata files again switched to branch autosynth google cloud pubsub note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at chore set ruby namespace in proto options previous head position was fix update gapic generator version to pickup discogapic fixes head is now at add a new authorizationtype for data source definition switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index git detects no change in google cloud pubsub skipping compatibility checking synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub synth metadata head is now at chore set ruby namespace in proto options switched to branch autosynth google cloud pubsub previous head position was add a new authorizationtype for data source definition head is now at fix update gapic generator version to pickup discogapic fixes note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at dialogflow weekly library update switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index git detects no change in google cloud pubsub skipping compatibility checking synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub synth metadata head is now at dialogflow weekly library update switched to branch autosynth google cloud pubsub head is now at fix update gapic generator version to pickup discogapic fixes note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at fix dataproc add missing required annotation switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index git detects no change in google cloud pubsub skipping compatibility checking synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub synth metadata head is now at fix dataproc add missing required annotation switched to branch autosynth google cloud pubsub head is now at fix update gapic generator version to pickup discogapic fixes note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at touch synth metadata to get secretmanager apis building switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index git detects no change in google cloud pubsub skipping compatibility checking synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub synth metadata head is now at touch synth metadata to get secretmanager apis building switched to branch autosynth google cloud pubsub head is now at fix update gapic generator version to pickup discogapic fixes note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at add a new authorizationtype for data source definition switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index git detects no change in google cloud pubsub skipping compatibility checking synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub synth metadata head is now at add a new authorizationtype for data source definition switched to branch autosynth google cloud pubsub on branch autosynth google cloud pubsub nothing to commit working tree clean traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main return inner main temp dir file tmpfs src git autosynth autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src git autosynth autosynth synth py line in synthesize loop synthesize range toolbox synthesizer file tmpfs src git autosynth autosynth synth py line in synthesize range toolbox patch merge version young file tmpfs src git autosynth autosynth synth py line in patch merge version comment or self versions version get comment file tmpfs src git autosynth autosynth git py line in commit all changes subprocess check call file home kbuilder pyenv versions lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
477,059
13,754,824,790
IssuesEvent
2020-10-06 17:30:46
red-hat-storage/ocs-ci
https://api.github.com/repos/red-hat-storage/ocs-ci
closed
Ensure ocscilib module adheres to log level set by user
High Priority team/ecosystem
Currently the logger used in [ocscilib](https://github.com/red-hat-storage/ocs-ci/blob/master/ocs_ci/framework/pytest_customization/ocscilib.py) is only logging `WARNING` and greater log levels. This should adhere to the log level set by the user (default is INFO). I suspect this is due to how pytest loads and calls this module / these functions, but we should still find a way to ensure we can log properly here as this code is critical.
1.0
Ensure ocscilib module adheres to log level set by user - Currently the logger used in [ocscilib](https://github.com/red-hat-storage/ocs-ci/blob/master/ocs_ci/framework/pytest_customization/ocscilib.py) is only logging `WARNING` and greater log levels. This should adhere to the log level set by the user (default is INFO). I suspect this is due to how pytest loads and calls this module / these functions, but we should still find a way to ensure we can log properly here as this code is critical.
non_test
ensure ocscilib module adheres to log level set by user currently the logger used in is only logging warning and greater log levels this should adhere to the log level set by the user default is info i suspect this is due to how pytest loads and calls this module these functions but we should still find a way to ensure we can log properly here as this code is critical
0
344,710
10,348,813,595
IssuesEvent
2019-09-04 20:43:32
mozilla/addons-code-manager
https://api.github.com/repos/mozilla/addons-code-manager
opened
Add API for creating / updating a comment
priority: p3 qa: not needed
The thunk for managing a comment (https://github.com/mozilla/addons-code-manager/issues/988) needs some methods to work with the [API](https://addons-server.readthedocs.io/en/latest/topics/api/reviewers.html#post--api-v4-reviewers-addon-(int-addon_id)-versions-(int-version_id)-draft_comments-).
1.0
Add API for creating / updating a comment - The thunk for managing a comment (https://github.com/mozilla/addons-code-manager/issues/988) needs some methods to work with the [API](https://addons-server.readthedocs.io/en/latest/topics/api/reviewers.html#post--api-v4-reviewers-addon-(int-addon_id)-versions-(int-version_id)-draft_comments-).
non_test
add api for creating updating a comment the thunk for managing a comment needs some methods to work with the
0
280,143
8,678,339,895
IssuesEvent
2018-11-30 19:36:11
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
closed
Synthesis failed for redis
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate redis. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth-redis' synthtool > You are running the synthesis script directly, this will be disabled in a future release of Synthtool. Please use python3 -m synthtool instead. synthtool > Ensuring dependencies. synthtool > Pulling artman image. latest: Pulling from googleapis/artman Digest: sha256:2f6b261ee7fe1aedf238991c93a20b3820de37a343d0cacf3e3e9555c2aaf2ea Status: Image is up to date for googleapis/artman:latest synthtool > Cloning googleapis. synthtool > Running generator for google/cloud/redis/artman_redis_v1beta1.yaml. synthtool > Generated code into /home/kbuilder/.cache/synthtool/googleapis/artman-genfiles/python/redis-v1beta1. synthtool > Running generator for google/cloud/redis/artman_redis_v1.yaml. synthtool > Generated code into /home/kbuilder/.cache/synthtool/googleapis/artman-genfiles/python/redis-v1. synthtool > No replacements made in google/cloud/**/cloud_redis_client.py for pattern resources of the form:\n ``, maybe replacement is not longer needed? synthtool > No replacements made in google/cloud/**/cloud_redis_client.py for pattern parent \(str\): Required. The resource name of the instance location using the form: :: `projects/{project_id}/locations/{location_id}` where ``location_id`` refers to a GCP region, maybe replacement is not longer needed? synthtool > No replacements made in google/cloud/**/cloud_redis_client.py for pattern with the following restrictions: \* Must contain only lowercase letters, numbers, and hyphens\., maybe replacement is not longer needed? synthtool > No replacements made in google/cloud/**/cloud_redis_client.py for pattern name \(str\): Required. Redis instance resource name using the form: :: `projects/{project_id}/locations/{location_id}/instances/{instance_id}` where ``location_id`` refers to a GCP region, maybe replacement is not longer needed? synthtool > No replacements made in google/cloud/**/cloud_redis_client.py for pattern fields from ``Instance``: \* ``displayName`` \* ``labels`` \* ``memorySizeGb`` \* ``redisConfig``, maybe replacement is not longer needed? .coveragerc .flake8 MANIFEST.in noxfile.py.j2 setup.cfg Traceback (most recent call last): File "synth.py", line 101, in <module> s.shell.run(["nox", "-s", "blacken"], hide_output=False) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run with Popen(*popenargs, **kwargs) as process: File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 707, in __init__ restore_signals, start_new_session) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 1326, in _execute_child raise child_exception_type(errno_num, err_msg) FileNotFoundError: [Errno 2] No such file or directory: 'nox' synthtool > Cleaned up 2 temporary directories. Synthesis failed ``` Google internal developers can see the full log [here](https://sponge/3abb58ef-0652-4854-846e-e1d9be78e73e).
1.0
Synthesis failed for redis - Hello! Autosynth couldn't regenerate redis. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth-redis' synthtool > You are running the synthesis script directly, this will be disabled in a future release of Synthtool. Please use python3 -m synthtool instead. synthtool > Ensuring dependencies. synthtool > Pulling artman image. latest: Pulling from googleapis/artman Digest: sha256:2f6b261ee7fe1aedf238991c93a20b3820de37a343d0cacf3e3e9555c2aaf2ea Status: Image is up to date for googleapis/artman:latest synthtool > Cloning googleapis. synthtool > Running generator for google/cloud/redis/artman_redis_v1beta1.yaml. synthtool > Generated code into /home/kbuilder/.cache/synthtool/googleapis/artman-genfiles/python/redis-v1beta1. synthtool > Running generator for google/cloud/redis/artman_redis_v1.yaml. synthtool > Generated code into /home/kbuilder/.cache/synthtool/googleapis/artman-genfiles/python/redis-v1. synthtool > No replacements made in google/cloud/**/cloud_redis_client.py for pattern resources of the form:\n ``, maybe replacement is not longer needed? synthtool > No replacements made in google/cloud/**/cloud_redis_client.py for pattern parent \(str\): Required. The resource name of the instance location using the form: :: `projects/{project_id}/locations/{location_id}` where ``location_id`` refers to a GCP region, maybe replacement is not longer needed? synthtool > No replacements made in google/cloud/**/cloud_redis_client.py for pattern with the following restrictions: \* Must contain only lowercase letters, numbers, and hyphens\., maybe replacement is not longer needed? synthtool > No replacements made in google/cloud/**/cloud_redis_client.py for pattern name \(str\): Required. Redis instance resource name using the form: :: `projects/{project_id}/locations/{location_id}/instances/{instance_id}` where ``location_id`` refers to a GCP region, maybe replacement is not longer needed? synthtool > No replacements made in google/cloud/**/cloud_redis_client.py for pattern fields from ``Instance``: \* ``displayName`` \* ``labels`` \* ``memorySizeGb`` \* ``redisConfig``, maybe replacement is not longer needed? .coveragerc .flake8 MANIFEST.in noxfile.py.j2 setup.cfg Traceback (most recent call last): File "synth.py", line 101, in <module> s.shell.run(["nox", "-s", "blacken"], hide_output=False) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run with Popen(*popenargs, **kwargs) as process: File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 707, in __init__ restore_signals, start_new_session) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 1326, in _execute_child raise child_exception_type(errno_num, err_msg) FileNotFoundError: [Errno 2] No such file or directory: 'nox' synthtool > Cleaned up 2 temporary directories. Synthesis failed ``` Google internal developers can see the full log [here](https://sponge/3abb58ef-0652-4854-846e-e1d9be78e73e).
non_test
synthesis failed for redis hello autosynth couldn t regenerate redis broken heart here s the output from running synth py cloning into working repo switched to branch autosynth redis    are running the synthesis script directly this will be disabled in a future release of synthtool please use m synthtool instead    dependencies    artman image  latest pulling from googleapis artman digest status image is up to date for googleapis artman latest   googleapis    generator for google cloud redis artman redis yaml    code into home kbuilder cache synthtool googleapis artman genfiles python redis    generator for google cloud redis artman redis yaml    code into home kbuilder cache synthtool googleapis artman genfiles python redis    replacements made in google cloud cloud redis client py for pattern resources of the form n maybe replacement is not longer needed    replacements made in google cloud cloud redis client py for pattern parent str required the resource name of the instance location using the form projects project id locations location id where location id refers to a gcp region maybe replacement is not longer needed    replacements made in google cloud cloud redis client py for pattern with the following restrictions must contain only lowercase letters numbers and hyphens maybe replacement is not longer needed    replacements made in google cloud cloud redis client py for pattern name str required redis instance resource name using the form projects project id locations location id instances instance id where location id refers to a gcp region maybe replacement is not longer needed    replacements made in google cloud cloud redis client py for pattern fields from instance displayname labels memorysizegb redisconfig maybe replacement is not longer needed  coveragerc manifest in noxfile py setup cfg traceback most recent call last file synth py line in s shell run hide output false file tmpfs src git autosynth env lib site packages synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run with popen popenargs kwargs as process file home kbuilder pyenv versions lib subprocess py line in init restore signals start new session file home kbuilder pyenv versions lib subprocess py line in execute child raise child exception type errno num err msg filenotfounderror no such file or directory nox   up temporary directories  synthesis failed google internal developers can see the full log
0
342,744
30,636,743,374
IssuesEvent
2023-07-24 18:22:58
Loweredgames/Skyblock-Classic-Edition
https://api.github.com/repos/Loweredgames/Skyblock-Classic-Edition
closed
Aggiustata e migliorata la funzione Legacy per la 5.0.0
bug confirmed low test fix performance
- [x] Testare --------------------------------------------- sara l'ultima volta che verra aggiustata. vedere: #202
1.0
Aggiustata e migliorata la funzione Legacy per la 5.0.0 - - [x] Testare --------------------------------------------- sara l'ultima volta che verra aggiustata. vedere: #202
test
aggiustata e migliorata la funzione legacy per la testare sara l ultima volta che verra aggiustata vedere
1
313,456
23,475,716,681
IssuesEvent
2022-08-17 05:40:58
ToolJet/ToolJet
https://api.github.com/repos/ToolJet/ToolJet
opened
[docs]: White labelling - EE feature
documentation
### Summary Add documentation for White labelling - the feature added to EE in which customer can configure`CUSTOM_LOGO_URL` - url of custom logo. best viewed : maximum width 130px height: 26px ### Issue Type Documentation feature/update ### The entire URL of the documentation with the issue _No response_ ### Steps to reproduce the issue _No response_ ### Additional Information _No response_ ### If the issue is confirmed, would you be willing to submit a pull request? No ### Code of Conduct - [X] I agree to follow the ToolJet Code of Conduct
1.0
[docs]: White labelling - EE feature - ### Summary Add documentation for White labelling - the feature added to EE in which customer can configure`CUSTOM_LOGO_URL` - url of custom logo. best viewed : maximum width 130px height: 26px ### Issue Type Documentation feature/update ### The entire URL of the documentation with the issue _No response_ ### Steps to reproduce the issue _No response_ ### Additional Information _No response_ ### If the issue is confirmed, would you be willing to submit a pull request? No ### Code of Conduct - [X] I agree to follow the ToolJet Code of Conduct
non_test
white labelling ee feature summary add documentation for white labelling the feature added to ee in which customer can configure custom logo url url of custom logo best viewed maximum width height issue type documentation feature update the entire url of the documentation with the issue no response steps to reproduce the issue no response additional information no response if the issue is confirmed would you be willing to submit a pull request no code of conduct i agree to follow the tooljet code of conduct
0
289,403
24,987,099,255
IssuesEvent
2022-11-02 15:48:55
vegaprotocol/vega
https://api.github.com/repos/vegaprotocol/vega
opened
Update feature tests for 0005-COLL-001 for additional events
feature tests
As part of the spec reviews it was mentioned that for the AC 0005-COLL-001 we would add additional tests for other events as per: - https://github.com/vegaprotocol/specs/issues/1321
1.0
Update feature tests for 0005-COLL-001 for additional events - As part of the spec reviews it was mentioned that for the AC 0005-COLL-001 we would add additional tests for other events as per: - https://github.com/vegaprotocol/specs/issues/1321
test
update feature tests for coll for additional events as part of the spec reviews it was mentioned that for the ac coll we would add additional tests for other events as per
1
160,516
12,515,966,032
IssuesEvent
2020-06-03 08:36:48
dasch-swiss/knora-app
https://api.github.com/repos/dasch-swiss/knora-app
closed
Removing a search criteria e.g. property
bug user-testing
*Advanced Search, not logged in I created a long search with several selection properties. There seems to be a limit in criterias (see also image) Now I decide, that the first of the properties is not so important and I wish to remove it from my list (in this example im picture the property "Color - exists). There seems to be no possibility to remove that property individually but I have to remove all others afterwards in the list as well. ![image](https://user-images.githubusercontent.com/56250095/68293465-4cad4000-008e-11ea-826a-e9d512b2b861.png) **Desktop (please complete the following information):** * OS: Windows 10 Pro N, 1903 - 18362.418 * Browser Google Chrome, 78.0.3904.87 * Version Knora App, 1.6.2
1.0
Removing a search criteria e.g. property - *Advanced Search, not logged in I created a long search with several selection properties. There seems to be a limit in criterias (see also image) Now I decide, that the first of the properties is not so important and I wish to remove it from my list (in this example im picture the property "Color - exists). There seems to be no possibility to remove that property individually but I have to remove all others afterwards in the list as well. ![image](https://user-images.githubusercontent.com/56250095/68293465-4cad4000-008e-11ea-826a-e9d512b2b861.png) **Desktop (please complete the following information):** * OS: Windows 10 Pro N, 1903 - 18362.418 * Browser Google Chrome, 78.0.3904.87 * Version Knora App, 1.6.2
test
removing a search criteria e g property advanced search not logged in i created a long search with several selection properties there seems to be a limit in criterias see also image now i decide that the first of the properties is not so important and i wish to remove it from my list in this example im picture the property color exists there seems to be no possibility to remove that property individually but i have to remove all others afterwards in the list as well desktop please complete the following information os windows pro n browser google chrome version knora app
1
288,135
31,857,033,785
IssuesEvent
2023-09-15 08:14:04
nidhi7598/linux-4.19.72_CVE-2022-3564
https://api.github.com/repos/nidhi7598/linux-4.19.72_CVE-2022-3564
closed
CVE-2020-25704 (Medium) detected in linuxlinux-4.19.294 - autoclosed
Mend: dependency security vulnerability
## CVE-2020-25704 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72_CVE-2022-3564/commit/454c7dacf6fa9a6de86d4067f5a08f25cffa519b">454c7dacf6fa9a6de86d4067f5a08f25cffa519b</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A flaw memory leak in the Linux kernel performance monitoring subsystem was found in the way if using PERF_EVENT_IOC_SET_FILTER. A local user could use this flaw to starve the resources causing denial of service. <p>Publish Date: 2020-12-02 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25704>CVE-2020-25704</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25704">https://www.linuxkernelcves.com/cves/CVE-2020-25704</a></p> <p>Release Date: 2020-12-02</p> <p>Fix Resolution: v4.9.244, v4.14.207, v4.19.156, v5.4.76, v5.9.7, v5.10-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-25704 (Medium) detected in linuxlinux-4.19.294 - autoclosed - ## CVE-2020-25704 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72_CVE-2022-3564/commit/454c7dacf6fa9a6de86d4067f5a08f25cffa519b">454c7dacf6fa9a6de86d4067f5a08f25cffa519b</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A flaw memory leak in the Linux kernel performance monitoring subsystem was found in the way if using PERF_EVENT_IOC_SET_FILTER. A local user could use this flaw to starve the resources causing denial of service. <p>Publish Date: 2020-12-02 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25704>CVE-2020-25704</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25704">https://www.linuxkernelcves.com/cves/CVE-2020-25704</a></p> <p>Release Date: 2020-12-02</p> <p>Fix Resolution: v4.9.244, v4.14.207, v4.19.156, v5.4.76, v5.9.7, v5.10-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files vulnerability details a flaw memory leak in the linux kernel performance monitoring subsystem was found in the way if using perf event ioc set filter a local user could use this flaw to starve the resources causing denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
777,562
27,285,220,220
IssuesEvent
2023-02-23 13:03:00
wso2/api-manager
https://api.github.com/repos/wso2/api-manager
closed
No Server ALPNProcessors! error when running APIEndpointCertificateTestCase
Type/Bug Priority/Normal Component/APIM Affected/4.2.0-Beta
### Description - When running the `APIEndpointCertificateTestCase` in the integration test module, there is an error occuring in the form of `No Server ALPNProcessors!`. - When debugged this was occuring due to a startup failure in the `Wire Mock Server` library. ### Steps to Reproduce 1. Run the `APIEndpointCertificateTestCase` by enabling only it in the testng.xml ### Affected Component APIM ### Version 4.2.0 ### Environment Details (with versions) MacOS Ventura with M1 chip ### Relevant Log Output ```shell INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - Jacoco class file paths analyzed INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - -- repository/components/plugins INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - -- repository/deployment/server/webapps INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - -- lib/runtimes/cxf3 INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - Jacoco coverage HTML report path : /Users/dinithherath/Desktop/Work/Personal/product-apim/modules/integration/tests-integration/tests-backend/target/jacoco/coverage/index.html [ERROR] Tests run: 53, Failures: 4, Errors: 0, Skipped: 20, Time elapsed: 1,087.093 s <<< FAILURE! - in TestSuite [ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize Time elapsed: 349.106 s <<< FAILURE! java.lang.IllegalStateException: No Server ALPNProcessors! at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.startSecureEndpoint(APIEndpointCertificateTestCase.java:317) at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize(APIEndpointCertificateTestCase.java:110) [ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy Time elapsed: 349.107 s <<< FAILURE! org.wso2.am.integration.clients.publisher.api.ApiException: Missing the required parameter 'apiId' when calling deleteAPI(Async) at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy(APIEndpointCertificateTestCase.java:327) [ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize Time elapsed: 351.543 s <<< FAILURE! java.lang.IllegalStateException: No Server ALPNProcessors! at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.startSecureEndpoint(APIEndpointCertificateTestCase.java:317) at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize(APIEndpointCertificateTestCase.java:110) [ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy Time elapsed: 351.543 s <<< FAILURE! org.wso2.am.integration.clients.publisher.api.ApiException: Missing the required parameter 'apiId' when calling deleteAPI(Async) at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy(APIEndpointCertificateTestCase.java:327) ``` ### Related Issues _No response_ ### Suggested Labels _No response_
1.0
No Server ALPNProcessors! error when running APIEndpointCertificateTestCase - ### Description - When running the `APIEndpointCertificateTestCase` in the integration test module, there is an error occuring in the form of `No Server ALPNProcessors!`. - When debugged this was occuring due to a startup failure in the `Wire Mock Server` library. ### Steps to Reproduce 1. Run the `APIEndpointCertificateTestCase` by enabling only it in the testng.xml ### Affected Component APIM ### Version 4.2.0 ### Environment Details (with versions) MacOS Ventura with M1 chip ### Relevant Log Output ```shell INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - Jacoco class file paths analyzed INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - -- repository/components/plugins INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - -- repository/deployment/server/webapps INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - -- lib/runtimes/cxf3 INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - Jacoco coverage HTML report path : /Users/dinithherath/Desktop/Work/Personal/product-apim/modules/integration/tests-integration/tests-backend/target/jacoco/coverage/index.html [ERROR] Tests run: 53, Failures: 4, Errors: 0, Skipped: 20, Time elapsed: 1,087.093 s <<< FAILURE! - in TestSuite [ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize Time elapsed: 349.106 s <<< FAILURE! java.lang.IllegalStateException: No Server ALPNProcessors! at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.startSecureEndpoint(APIEndpointCertificateTestCase.java:317) at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize(APIEndpointCertificateTestCase.java:110) [ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy Time elapsed: 349.107 s <<< FAILURE! org.wso2.am.integration.clients.publisher.api.ApiException: Missing the required parameter 'apiId' when calling deleteAPI(Async) at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy(APIEndpointCertificateTestCase.java:327) [ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize Time elapsed: 351.543 s <<< FAILURE! java.lang.IllegalStateException: No Server ALPNProcessors! at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.startSecureEndpoint(APIEndpointCertificateTestCase.java:317) at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize(APIEndpointCertificateTestCase.java:110) [ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy Time elapsed: 351.543 s <<< FAILURE! org.wso2.am.integration.clients.publisher.api.ApiException: Missing the required parameter 'apiId' when calling deleteAPI(Async) at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy(APIEndpointCertificateTestCase.java:327) ``` ### Related Issues _No response_ ### Suggested Labels _No response_
non_test
no server alpnprocessors error when running apiendpointcertificatetestcase description when running the apiendpointcertificatetestcase in the integration test module there is an error occuring in the form of no server alpnprocessors when debugged this was occuring due to a startup failure in the wire mock server library steps to reproduce run the apiendpointcertificatetestcase by enabling only it in the testng xml affected component apim version environment details with versions macos ventura with chip relevant log output shell info jacoco class file paths analyzed info repository components plugins info repository deployment server webapps info lib runtimes info jacoco coverage html report path users dinithherath desktop work personal product apim modules integration tests integration tests backend target jacoco coverage index html tests run failures errors skipped time elapsed s failure in testsuite org am integration tests api lifecycle apiendpointcertificatetestcase initialize time elapsed s failure java lang illegalstateexception no server alpnprocessors at org am integration tests api lifecycle apiendpointcertificatetestcase startsecureendpoint apiendpointcertificatetestcase java at org am integration tests api lifecycle apiendpointcertificatetestcase initialize apiendpointcertificatetestcase java org am integration tests api lifecycle apiendpointcertificatetestcase destroy time elapsed s failure org am integration clients publisher api apiexception missing the required parameter apiid when calling deleteapi async at org am integration tests api lifecycle apiendpointcertificatetestcase destroy apiendpointcertificatetestcase java org am integration tests api lifecycle apiendpointcertificatetestcase initialize time elapsed s failure java lang illegalstateexception no server alpnprocessors at org am integration tests api lifecycle apiendpointcertificatetestcase startsecureendpoint apiendpointcertificatetestcase java at org am integration tests api lifecycle apiendpointcertificatetestcase initialize apiendpointcertificatetestcase java org am integration tests api lifecycle apiendpointcertificatetestcase destroy time elapsed s failure org am integration clients publisher api apiexception missing the required parameter apiid when calling deleteapi async at org am integration tests api lifecycle apiendpointcertificatetestcase destroy apiendpointcertificatetestcase java related issues no response suggested labels no response
0
84,289
7,909,990,545
IssuesEvent
2018-07-03 03:08:42
elastic/beats
https://api.github.com/repos/elastic/beats
opened
Test Filebeat modules from packages
:Testing Filebeat
We don't have any tests in place to detect if the Filebeat modules work from the packages we produce. I noticed this because they weren't included in the packages and none of our tests failed. I suggest adding some tests in these areas: - [ ] Check that packages have `modules` and `modules.d` at a minimum. Currently we check file permissions and ownership, but we don't assert they exist. I suggest adding a flag like `--modules` or `--dir` that causes the check to test that the directories exist in all packages. - [ ] Run some tests on the Filebeat modules from beats-tester. Maybe execute the enable commands.
1.0
Test Filebeat modules from packages - We don't have any tests in place to detect if the Filebeat modules work from the packages we produce. I noticed this because they weren't included in the packages and none of our tests failed. I suggest adding some tests in these areas: - [ ] Check that packages have `modules` and `modules.d` at a minimum. Currently we check file permissions and ownership, but we don't assert they exist. I suggest adding a flag like `--modules` or `--dir` that causes the check to test that the directories exist in all packages. - [ ] Run some tests on the Filebeat modules from beats-tester. Maybe execute the enable commands.
test
test filebeat modules from packages we don t have any tests in place to detect if the filebeat modules work from the packages we produce i noticed this because they weren t included in the packages and none of our tests failed i suggest adding some tests in these areas check that packages have modules and modules d at a minimum currently we check file permissions and ownership but we don t assert they exist i suggest adding a flag like modules or dir that causes the check to test that the directories exist in all packages run some tests on the filebeat modules from beats tester maybe execute the enable commands
1
27,269
12,535,511,669
IssuesEvent
2020-06-04 21:31:24
MicrosoftDocs/powerbi-docs
https://api.github.com/repos/MicrosoftDocs/powerbi-docs
closed
PostgreSQL requires gateway
assigned-to-author doc-bug powerbi-service/subsvc powerbi/svc pri1
The table is incorrect; PostgreSQL requires the data gateway to connect and refresh from the Power BI Service. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3c34c249-033f-4156-0abc-8d16cbe355a0 * Version Independent ID: 20dc29c2-9cb7-2322-e80b-3f801bc34cfe * Content: [Power BI data sources - Power BI](https://docs.microsoft.com/en-us/power-bi/power-bi-data-sources#feedback) * Content Source: [powerbi-docs/power-bi-data-sources.md](https://github.com/MicrosoftDocs/powerbi-docs/blob/live/powerbi-docs/power-bi-data-sources.md) * Service: **powerbi** * Sub-service: **powerbi-service** * GitHub Login: @kfollis * Microsoft Alias: **kfollis**
1.0
PostgreSQL requires gateway - The table is incorrect; PostgreSQL requires the data gateway to connect and refresh from the Power BI Service. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3c34c249-033f-4156-0abc-8d16cbe355a0 * Version Independent ID: 20dc29c2-9cb7-2322-e80b-3f801bc34cfe * Content: [Power BI data sources - Power BI](https://docs.microsoft.com/en-us/power-bi/power-bi-data-sources#feedback) * Content Source: [powerbi-docs/power-bi-data-sources.md](https://github.com/MicrosoftDocs/powerbi-docs/blob/live/powerbi-docs/power-bi-data-sources.md) * Service: **powerbi** * Sub-service: **powerbi-service** * GitHub Login: @kfollis * Microsoft Alias: **kfollis**
non_test
postgresql requires gateway the table is incorrect postgresql requires the data gateway to connect and refresh from the power bi service document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service powerbi sub service powerbi service github login kfollis microsoft alias kfollis
0
50,622
13,541,511,051
IssuesEvent
2020-09-16 15:58:43
oracle/helidon
https://api.github.com/repos/oracle/helidon
closed
Fix javadoc of EvictableCache
2.x P3 security
EvictableCache now implies some implementation details in its javadoc, such as backing it by a `ConcurrentHashMap` - this should not be done, as anybody can implement the interface. Fix javadoc of the class and its methods + add details to `factory` methods and builder that explain how the default implementation works. Examples of wrong docs: `Generic cache with eviction and max size` - the interface itself does not imply max size javadoc of `size` method.
True
Fix javadoc of EvictableCache - EvictableCache now implies some implementation details in its javadoc, such as backing it by a `ConcurrentHashMap` - this should not be done, as anybody can implement the interface. Fix javadoc of the class and its methods + add details to `factory` methods and builder that explain how the default implementation works. Examples of wrong docs: `Generic cache with eviction and max size` - the interface itself does not imply max size javadoc of `size` method.
non_test
fix javadoc of evictablecache evictablecache now implies some implementation details in its javadoc such as backing it by a concurrenthashmap this should not be done as anybody can implement the interface fix javadoc of the class and its methods add details to factory methods and builder that explain how the default implementation works examples of wrong docs generic cache with eviction and max size the interface itself does not imply max size javadoc of size method
0
204,164
7,084,874,615
IssuesEvent
2018-01-11 08:59:40
wulkano/kap
https://api.github.com/repos/wulkano/kap
opened
Blacklist redundant apps from the window selector.
Priority: Medium
**Platform:** MacOS **Kap Version:** 2.0.0-beta.4 (2.0.0-beta.4.556) ### Steps to reproduce Currently, a vast amount of redundant apps show up in the window selector that are currently not recordable. ### Workaround Suggestion: Create a blacklist of known redundant apps that shouldn't show up in the menu.
1.0
Blacklist redundant apps from the window selector. - **Platform:** MacOS **Kap Version:** 2.0.0-beta.4 (2.0.0-beta.4.556) ### Steps to reproduce Currently, a vast amount of redundant apps show up in the window selector that are currently not recordable. ### Workaround Suggestion: Create a blacklist of known redundant apps that shouldn't show up in the menu.
non_test
blacklist redundant apps from the window selector platform macos kap version beta beta steps to reproduce currently a vast amount of redundant apps show up in the window selector that are currently not recordable workaround suggestion create a blacklist of known redundant apps that shouldn t show up in the menu
0
61,155
25,383,311,305
IssuesEvent
2022-11-21 19:29:41
BCDevOps/developer-experience
https://api.github.com/repos/BCDevOps/developer-experience
closed
SSO realm merge planning - Rocketchat
ops and shared services
**Describe the issue** we will be merging the realm that RC uses to another realm in Gold. This ticket is to find out what's the impact and if the merge is doable for this service. **What is the plan? How will this get completed?** Discussion with Service Lead, testing, planning **Definition of done** - [x] identify service's SSO usage - [x] discuss if realm merge is doable - [x] come up with testing plan
1.0
SSO realm merge planning - Rocketchat - **Describe the issue** we will be merging the realm that RC uses to another realm in Gold. This ticket is to find out what's the impact and if the merge is doable for this service. **What is the plan? How will this get completed?** Discussion with Service Lead, testing, planning **Definition of done** - [x] identify service's SSO usage - [x] discuss if realm merge is doable - [x] come up with testing plan
non_test
sso realm merge planning rocketchat describe the issue we will be merging the realm that rc uses to another realm in gold this ticket is to find out what s the impact and if the merge is doable for this service what is the plan how will this get completed discussion with service lead testing planning definition of done identify service s sso usage discuss if realm merge is doable come up with testing plan
0
145,980
11,716,637,697
IssuesEvent
2020-03-09 15:56:57
waterloop/teamhub
https://api.github.com/repos/waterloop/teamhub
closed
Add config.json for tests running on circle-ci
backend testing
Need to add mongodb+srv url for when tests run on circle-ci.
1.0
Add config.json for tests running on circle-ci - Need to add mongodb+srv url for when tests run on circle-ci.
test
add config json for tests running on circle ci need to add mongodb srv url for when tests run on circle ci
1
292,120
25,202,059,142
IssuesEvent
2022-11-13 08:18:50
david3214/Team_2_CS_2450-601
https://api.github.com/repos/david3214/Team_2_CS_2450-601
closed
Functional/Non-Functional Testing
testing
According to the textbook: **Functional test**—Deals with features the application provides. These are generally listed in the requirements. **Nonfunctional test**—Studies application characteristics that aren’t related to specific functions the users will perform. For example, these tests might check performance under a heavy user load, with limited memory, or with missing network connections. These often identify minimal requirements.
1.0
Functional/Non-Functional Testing - According to the textbook: **Functional test**—Deals with features the application provides. These are generally listed in the requirements. **Nonfunctional test**—Studies application characteristics that aren’t related to specific functions the users will perform. For example, these tests might check performance under a heavy user load, with limited memory, or with missing network connections. These often identify minimal requirements.
test
functional non functional testing according to the textbook functional test —deals with features the application provides these are generally listed in the requirements nonfunctional test —studies application characteristics that aren’t related to specific functions the users will perform for example these tests might check performance under a heavy user load with limited memory or with missing network connections these often identify minimal requirements
1
256,971
22,139,086,764
IssuesEvent
2022-06-03 03:55:01
cmwall0605/cse112-sp22-group8
https://api.github.com/repos/cmwall0605/cse112-sp22-group8
opened
[Testing]: Add Overall Testing Suite for Cypress
difficult issue Priority: High testing task-11-testing
### Describe the feature to implement Create a comprehensive suite of testing for the user ### Breakdown of what needs to be completed - [ ] Create a test where a user enters the timer page, starts, and does not create a task. Ensure that the create task dialog pops up, and ensure that the timer runs for 30 minutes (25 work + 5 break). Ensure that after the break, that there is no task-item created in the task page. Ensure that the stats timer page has the correct info - [ ] Create a test where a user enters the timer page, starts, and does create a task. Ensure that the create task dialog pops up, and ensure that the timer runs for 30 minutes (25 work + 5 break). Ensure that after the break, that there is a task-item created in the task page with the correct title and the correct ratio. Ensure that the stats timer page has the correct info. - [ ] Create a test where a user enters the timer page, goes to the stats page, creates a task, and then goes to the timer page through the header (not by playing the task). Ensure that it goes through the same process as the first process. - [ ] Create a test where a user enters the timer page, goes to the stats page, creates a task, and then goes to the timer page through the task item (not through the header). Click the start and ensure that the create task does not show up. After the session, go to the task page and ensure that the task chosen updates its ratio and that the stats page is updated. - [ ] Create a test where the user edits the work, short and long break and then starts a test. Make sure that the work, short, and long sessions are the correct amount of time. - [ ] Enable auto timer and run through 4 work sessions, ensuring that that the system runs through them automatically and that no interaction needs to happen. ### Additional Context if Applicable
2.0
[Testing]: Add Overall Testing Suite for Cypress - ### Describe the feature to implement Create a comprehensive suite of testing for the user ### Breakdown of what needs to be completed - [ ] Create a test where a user enters the timer page, starts, and does not create a task. Ensure that the create task dialog pops up, and ensure that the timer runs for 30 minutes (25 work + 5 break). Ensure that after the break, that there is no task-item created in the task page. Ensure that the stats timer page has the correct info - [ ] Create a test where a user enters the timer page, starts, and does create a task. Ensure that the create task dialog pops up, and ensure that the timer runs for 30 minutes (25 work + 5 break). Ensure that after the break, that there is a task-item created in the task page with the correct title and the correct ratio. Ensure that the stats timer page has the correct info. - [ ] Create a test where a user enters the timer page, goes to the stats page, creates a task, and then goes to the timer page through the header (not by playing the task). Ensure that it goes through the same process as the first process. - [ ] Create a test where a user enters the timer page, goes to the stats page, creates a task, and then goes to the timer page through the task item (not through the header). Click the start and ensure that the create task does not show up. After the session, go to the task page and ensure that the task chosen updates its ratio and that the stats page is updated. - [ ] Create a test where the user edits the work, short and long break and then starts a test. Make sure that the work, short, and long sessions are the correct amount of time. - [ ] Enable auto timer and run through 4 work sessions, ensuring that that the system runs through them automatically and that no interaction needs to happen. ### Additional Context if Applicable
test
add overall testing suite for cypress describe the feature to implement create a comprehensive suite of testing for the user breakdown of what needs to be completed create a test where a user enters the timer page starts and does not create a task ensure that the create task dialog pops up and ensure that the timer runs for minutes work break ensure that after the break that there is no task item created in the task page ensure that the stats timer page has the correct info create a test where a user enters the timer page starts and does create a task ensure that the create task dialog pops up and ensure that the timer runs for minutes work break ensure that after the break that there is a task item created in the task page with the correct title and the correct ratio ensure that the stats timer page has the correct info create a test where a user enters the timer page goes to the stats page creates a task and then goes to the timer page through the header not by playing the task ensure that it goes through the same process as the first process create a test where a user enters the timer page goes to the stats page creates a task and then goes to the timer page through the task item not through the header click the start and ensure that the create task does not show up after the session go to the task page and ensure that the task chosen updates its ratio and that the stats page is updated create a test where the user edits the work short and long break and then starts a test make sure that the work short and long sessions are the correct amount of time enable auto timer and run through work sessions ensuring that that the system runs through them automatically and that no interaction needs to happen additional context if applicable
1
280,164
24,281,791,791
IssuesEvent
2022-09-28 18:05:35
Ascentroid/Ascentroid
https://api.github.com/repos/Ascentroid/Ascentroid
closed
Aftershock splash damage not killing sometimes, when it should
bug fixed needs testing
(In Shock Henge) Aftershock splash damage not killing sometimes, when it should
1.0
Aftershock splash damage not killing sometimes, when it should - (In Shock Henge) Aftershock splash damage not killing sometimes, when it should
test
aftershock splash damage not killing sometimes when it should in shock henge aftershock splash damage not killing sometimes when it should
1
165,785
12,879,872,091
IssuesEvent
2020-07-12 01:27:28
osquery/osquery
https://api.github.com/repos/osquery/osquery
closed
Create tests for table `shell_history`
FreeBSD Linux good-first-issue macOS test
## Create tests for table `shell_history` - Create header file for the table implementation, if one is not exists. - In test, query the table and check if retrieved columns (name and types) match the columns from table spec. - If there is any guarantee to number of rows (e.g. only 1 record in every query result, more than 3 records or something else) check it. - Test the implementation details of the table, if it possible. Table spec: `specs/posix/shell_history.table` Source files: - `osquery/tables/system/posix/shell_history.h` Table generating function: `genShellHistory()`
1.0
Create tests for table `shell_history` - ## Create tests for table `shell_history` - Create header file for the table implementation, if one is not exists. - In test, query the table and check if retrieved columns (name and types) match the columns from table spec. - If there is any guarantee to number of rows (e.g. only 1 record in every query result, more than 3 records or something else) check it. - Test the implementation details of the table, if it possible. Table spec: `specs/posix/shell_history.table` Source files: - `osquery/tables/system/posix/shell_history.h` Table generating function: `genShellHistory()`
test
create tests for table shell history create tests for table shell history create header file for the table implementation if one is not exists in test query the table and check if retrieved columns name and types match the columns from table spec if there is any guarantee to number of rows e g only record in every query result more than records or something else check it test the implementation details of the table if it possible table spec specs posix shell history table source files osquery tables system posix shell history h table generating function genshellhistory
1
315,940
27,120,548,247
IssuesEvent
2023-02-15 22:26:26
filecoin-project/ref-fvm
https://api.github.com/repos/filecoin-project/ref-fvm
opened
Eth API: Write full test case for EthGetStorageAt
Topic: Testing
Specifically, we need to test: 1. Looking up storage for a non-existent actor (returns 32 zero bytes). 2. Looking up storage for a non-evm actor (returns 32 zero bytes). 3. Looking up an unset slot (returns 32 zero bytes). 4. Looking up a set slot (returns the value). 5. Looking up a set slot _after_ deleting an actor (selfdestruct). We already have some test coverage here, but it's not thorough.
1.0
Eth API: Write full test case for EthGetStorageAt - Specifically, we need to test: 1. Looking up storage for a non-existent actor (returns 32 zero bytes). 2. Looking up storage for a non-evm actor (returns 32 zero bytes). 3. Looking up an unset slot (returns 32 zero bytes). 4. Looking up a set slot (returns the value). 5. Looking up a set slot _after_ deleting an actor (selfdestruct). We already have some test coverage here, but it's not thorough.
test
eth api write full test case for ethgetstorageat specifically we need to test looking up storage for a non existent actor returns zero bytes looking up storage for a non evm actor returns zero bytes looking up an unset slot returns zero bytes looking up a set slot returns the value looking up a set slot after deleting an actor selfdestruct we already have some test coverage here but it s not thorough
1
551,886
16,190,678,355
IssuesEvent
2021-05-04 08:02:10
epiphany-platform/epiphany
https://api.github.com/repos/epiphany-platform/epiphany
closed
[BUG] [Ubuntu] elasticsearch-curator in version 5.8.3 is not available from APT repo
area/logs priority/critical type/bug
**Describe the bug** Epicli installation fails while downloading requirements as the `elasticsearch-curator` package in version 5.8.3 is no longer available in the Curator APT repository. **How to reproduce** Steps to reproduce the behavior: 1. execute `epicli apply` (repository vm is enough) **Expected behavior** Repository has been set up properly. **Config files** **Environment** - OS: [Ubuntu] **epicli version**: [`epicli --version`] All epicli versions affected **Additional context** ``` root@ip-10-1-11-41:~# apt-cache policy elasticsearch-curator elasticsearch-curator: Installed: (none) Candidate: 5.8.4 Version table: 5.8.4 500 500 https://packages.elastic.co/curator/5/debian stable/main amd64 Packages 5.2.0-1 500 500 http://eu-west-1.ec2.archive.ubuntu.com/ubuntu bionic/universe amd64 Packages ``` --- **DoD checklist** * [x] Changelog updated (if affected version was released) * [x] COMPONENTS.md updated / doesn't need to be updated * [x] Automated tests passed (QA pipelines) * [x] apply * [ ] upgrade * [x] Case covered by automated test (if possible) * [ ] Idempotency tested * [x] Documentation updated / doesn't need to be updated * [x] All conversations in PR resolved * [x] Backport tasks created / doesn't need to be backported
1.0
[BUG] [Ubuntu] elasticsearch-curator in version 5.8.3 is not available from APT repo - **Describe the bug** Epicli installation fails while downloading requirements as the `elasticsearch-curator` package in version 5.8.3 is no longer available in the Curator APT repository. **How to reproduce** Steps to reproduce the behavior: 1. execute `epicli apply` (repository vm is enough) **Expected behavior** Repository has been set up properly. **Config files** **Environment** - OS: [Ubuntu] **epicli version**: [`epicli --version`] All epicli versions affected **Additional context** ``` root@ip-10-1-11-41:~# apt-cache policy elasticsearch-curator elasticsearch-curator: Installed: (none) Candidate: 5.8.4 Version table: 5.8.4 500 500 https://packages.elastic.co/curator/5/debian stable/main amd64 Packages 5.2.0-1 500 500 http://eu-west-1.ec2.archive.ubuntu.com/ubuntu bionic/universe amd64 Packages ``` --- **DoD checklist** * [x] Changelog updated (if affected version was released) * [x] COMPONENTS.md updated / doesn't need to be updated * [x] Automated tests passed (QA pipelines) * [x] apply * [ ] upgrade * [x] Case covered by automated test (if possible) * [ ] Idempotency tested * [x] Documentation updated / doesn't need to be updated * [x] All conversations in PR resolved * [x] Backport tasks created / doesn't need to be backported
non_test
elasticsearch curator in version is not available from apt repo describe the bug epicli installation fails while downloading requirements as the elasticsearch curator package in version is no longer available in the curator apt repository how to reproduce steps to reproduce the behavior execute epicli apply repository vm is enough expected behavior repository has been set up properly config files environment os epicli version all epicli versions affected additional context root ip apt cache policy elasticsearch curator elasticsearch curator installed none candidate version table stable main packages bionic universe packages dod checklist changelog updated if affected version was released components md updated doesn t need to be updated automated tests passed qa pipelines apply upgrade case covered by automated test if possible idempotency tested documentation updated doesn t need to be updated all conversations in pr resolved backport tasks created doesn t need to be backported
0
273,927
23,795,534,578
IssuesEvent
2022-09-02 19:11:01
pypa/pip
https://api.github.com/repos/pypa/pip
closed
Test Suite: Tests are not using the latest setuptools
type: enhancement C: tests
I'm struggling with running the tests while implementing #5407 (https://travis-ci.org/pfmoore/pip/jobs/420017719). The failure "error: 'egg_base' must be a directory name" is a bug in setuptools, that is fixed in the latest version, 40.2.0. But after a lot of debugging, I've found that the build environment used by `test_pip_wheel_with_pep518_build_reqs_no_isolation` has setuptools 39.1.0 installed. I've no idea why - there's no mention of that version in the pip sources, and as far as I can see that version isn't installed on my PC. So I'm somewhat baffled as to where the test suite is getting that version from. If anyone has any suggestions, I'd really appreciate it. Ultimately, this is a bug in the test suite as it doesn't seem to be using the latest version of setuptools, but in the short term my main concern is that I can't progress on the PEP 517 implementation until I get this sorted. @pypa/pip-committers any suggestions? Or anyone else?
1.0
Test Suite: Tests are not using the latest setuptools - I'm struggling with running the tests while implementing #5407 (https://travis-ci.org/pfmoore/pip/jobs/420017719). The failure "error: 'egg_base' must be a directory name" is a bug in setuptools, that is fixed in the latest version, 40.2.0. But after a lot of debugging, I've found that the build environment used by `test_pip_wheel_with_pep518_build_reqs_no_isolation` has setuptools 39.1.0 installed. I've no idea why - there's no mention of that version in the pip sources, and as far as I can see that version isn't installed on my PC. So I'm somewhat baffled as to where the test suite is getting that version from. If anyone has any suggestions, I'd really appreciate it. Ultimately, this is a bug in the test suite as it doesn't seem to be using the latest version of setuptools, but in the short term my main concern is that I can't progress on the PEP 517 implementation until I get this sorted. @pypa/pip-committers any suggestions? Or anyone else?
test
test suite tests are not using the latest setuptools i m struggling with running the tests while implementing the failure error egg base must be a directory name is a bug in setuptools that is fixed in the latest version but after a lot of debugging i ve found that the build environment used by test pip wheel with build reqs no isolation has setuptools installed i ve no idea why there s no mention of that version in the pip sources and as far as i can see that version isn t installed on my pc so i m somewhat baffled as to where the test suite is getting that version from if anyone has any suggestions i d really appreciate it ultimately this is a bug in the test suite as it doesn t seem to be using the latest version of setuptools but in the short term my main concern is that i can t progress on the pep implementation until i get this sorted pypa pip committers any suggestions or anyone else
1
308,040
26,575,727,506
IssuesEvent
2023-01-21 19:50:04
TheRenegadeCoder/sample-programs
https://api.github.com/repos/TheRenegadeCoder/sample-programs
closed
Add Dijkstra in Mathematica
enhancement needs tests dijkstra
To request a new code snippet, please fill out the following: Project name: Dijkstra Project link: https://sampleprograms.io/projects/dijkstra Language: Mathematica If this code snippet will also be the first in a new language, please fill out the following: Official Language Style Guide: Official Language Website: Official Language Docker Image: https://hub.docker.com/r/repository/image > The project link must be a link to a project on the [Sample Programs Website project list][sample-programs-project-list]. > If you would like to add a new project to the Sample Programs Website, > please first make a pull request to the [Sample Programs Website][sample-programs-website]. [sample-programs-website]: https://github.com/TheRenegadeCoder/sample-programs-website [sample-programs-project-list]: https://sampleprograms.io/projects/
1.0
Add Dijkstra in Mathematica - To request a new code snippet, please fill out the following: Project name: Dijkstra Project link: https://sampleprograms.io/projects/dijkstra Language: Mathematica If this code snippet will also be the first in a new language, please fill out the following: Official Language Style Guide: Official Language Website: Official Language Docker Image: https://hub.docker.com/r/repository/image > The project link must be a link to a project on the [Sample Programs Website project list][sample-programs-project-list]. > If you would like to add a new project to the Sample Programs Website, > please first make a pull request to the [Sample Programs Website][sample-programs-website]. [sample-programs-website]: https://github.com/TheRenegadeCoder/sample-programs-website [sample-programs-project-list]: https://sampleprograms.io/projects/
test
add dijkstra in mathematica to request a new code snippet please fill out the following project name dijkstra project link language mathematica if this code snippet will also be the first in a new language please fill out the following official language style guide official language website official language docker image the project link must be a link to a project on the if you would like to add a new project to the sample programs website please first make a pull request to the
1
309,491
23,297,202,429
IssuesEvent
2022-08-06 19:18:54
r-hyperspec/hyperSpec
https://api.github.com/repos/r-hyperspec/hyperSpec
opened
Review the examples with faux_cell if they make sense
Topic: documentation :blue_book:
After changing from `chondro` into `faux_cell` in examples and vignettes, the examples should be carefully reviewed to make sure that the results make sense. https://github.com/cbeleites/hyperSpec/issues/236
1.0
Review the examples with faux_cell if they make sense - After changing from `chondro` into `faux_cell` in examples and vignettes, the examples should be carefully reviewed to make sure that the results make sense. https://github.com/cbeleites/hyperSpec/issues/236
non_test
review the examples with faux cell if they make sense after changing from chondro into faux cell in examples and vignettes the examples should be carefully reviewed to make sure that the results make sense
0
100,794
8,755,136,686
IssuesEvent
2018-12-14 14:00:18
LiskHQ/lisk-hub
https://api.github.com/repos/LiskHQ/lisk-hub
closed
Fix unit tests for send transaction with HW wallet - Followup #1317
type: test
<!--- Provide a descriptive summary in the Title above --> ### Description of the problem <!-- A description of the problem you want to solve, including why you think this is a problem --> Create tests covering files below so test coverage is passing ```'src/components/passphraseSteps/index.js', 'src/components/resultBox/resultBox.js', 'src/components/send/steps/confirm/confirm.js', 'src/actions/peers.js', ,```
1.0
Fix unit tests for send transaction with HW wallet - Followup #1317 - <!--- Provide a descriptive summary in the Title above --> ### Description of the problem <!-- A description of the problem you want to solve, including why you think this is a problem --> Create tests covering files below so test coverage is passing ```'src/components/passphraseSteps/index.js', 'src/components/resultBox/resultBox.js', 'src/components/send/steps/confirm/confirm.js', 'src/actions/peers.js', ,```
test
fix unit tests for send transaction with hw wallet followup description of the problem create tests covering files below so test coverage is passing src components passphrasesteps index js src components resultbox resultbox js src components send steps confirm confirm js src actions peers js
1
114,266
14,544,025,415
IssuesEvent
2020-12-15 17:37:05
cloudfour/cloudfour.com-patterns
https://api.github.com/repos/cloudfour/cloudfour.com-patterns
closed
Grid/layout patterns
size:10 ⁉️ question ✋ blocked 🎨 design
Create a grid system that can be used to layout content. <img width="636" alt="grids" src="https://user-images.githubusercontent.com/42841342/74988623-1d2f0300-53f3-11ea-8f7d-1f1fd05db3ca.png">
1.0
Grid/layout patterns - Create a grid system that can be used to layout content. <img width="636" alt="grids" src="https://user-images.githubusercontent.com/42841342/74988623-1d2f0300-53f3-11ea-8f7d-1f1fd05db3ca.png">
non_test
grid layout patterns create a grid system that can be used to layout content img width alt grids src
0
222,518
7,433,136,136
IssuesEvent
2018-03-26 06:03:32
telerik/kendo-ui-core
https://api.github.com/repos/telerik/kendo-ui-core
opened
The popup edit form is not closed on the first update click if there was a invalid value before entering a valid one(SASS themes).
Bug C: Grid Kendo1 Priority 1 SEV: Low
### Bug report The popup edit form is not closed on the first update click if there was an invalid value before entering a valid one(SASS themes). ### Reproduction of the problem The issue can be observed in our demo: https://demos.telerik.com/kendo-ui/grid/editing-popup I made a video showcasing the issue: https://www.screencast.com/t/FiR6zWhPp ### Environment * **Kendo UI version:** 2018.2.221 * **Browser:** [all ]
1.0
The popup edit form is not closed on the first update click if there was a invalid value before entering a valid one(SASS themes). - ### Bug report The popup edit form is not closed on the first update click if there was an invalid value before entering a valid one(SASS themes). ### Reproduction of the problem The issue can be observed in our demo: https://demos.telerik.com/kendo-ui/grid/editing-popup I made a video showcasing the issue: https://www.screencast.com/t/FiR6zWhPp ### Environment * **Kendo UI version:** 2018.2.221 * **Browser:** [all ]
non_test
the popup edit form is not closed on the first update click if there was a invalid value before entering a valid one sass themes bug report the popup edit form is not closed on the first update click if there was an invalid value before entering a valid one sass themes reproduction of the problem the issue can be observed in our demo i made a video showcasing the issue environment kendo ui version browser
0
48,281
5,950,975,081
IssuesEvent
2017-05-26 18:12:55
openbmc/openbmc-test-automation
https://api.github.com/repos/openbmc/openbmc-test-automation
opened
[Network Test] Upload BMC work-around
SYS_TEST Test
The /tmp/images/ are specially reserved for code update. It doesnt' allow to upload unqualified file..
2.0
[Network Test] Upload BMC work-around - The /tmp/images/ are specially reserved for code update. It doesnt' allow to upload unqualified file..
test
upload bmc work around the tmp images are specially reserved for code update it doesnt allow to upload unqualified file
1
248,870
21,081,700,726
IssuesEvent
2022-04-03 01:33:19
Uuvana-Studios/longvinter-windows-client
https://api.github.com/repos/Uuvana-Studios/longvinter-windows-client
opened
Turrets not working even though electricity needs are met
Bug Not Tested
**Describe the bug** Turrets not working even though electricity needs are met. **To Reproduce** Steps to reproduce the behavior: 1. Fully upgrade house down sustainability path. 2. Install one turret. 4. Turret does not work during night. 5. Install second turret. 6. Second turret does not work during the day. **Expected behavior** Meeting electricity requirements would mean that turrets work. **Screenshots** ![image](https://user-images.githubusercontent.com/102885424/161407071-91be3f54-e8b5-4f91-8328-1bb84a021d84.png) Fully upgraded sustainable power path ![image](https://user-images.githubusercontent.com/102885424/161407203-768ad13b-314a-487c-b52e-3ea84b07cb3d.png) Turrets stop working at night ![image](https://user-images.githubusercontent.com/102885424/161407279-8a2aabc2-e8fc-47a0-afd9-ad632b0d7e9c.png) Daytime with one turret on, one turret off **Desktop (please complete the following information):** - OS: Windows - Game Version 1.0.2 - Steam Version **Additional context** Other solar powered houses have multiple fully working turrets.
1.0
Turrets not working even though electricity needs are met - **Describe the bug** Turrets not working even though electricity needs are met. **To Reproduce** Steps to reproduce the behavior: 1. Fully upgrade house down sustainability path. 2. Install one turret. 4. Turret does not work during night. 5. Install second turret. 6. Second turret does not work during the day. **Expected behavior** Meeting electricity requirements would mean that turrets work. **Screenshots** ![image](https://user-images.githubusercontent.com/102885424/161407071-91be3f54-e8b5-4f91-8328-1bb84a021d84.png) Fully upgraded sustainable power path ![image](https://user-images.githubusercontent.com/102885424/161407203-768ad13b-314a-487c-b52e-3ea84b07cb3d.png) Turrets stop working at night ![image](https://user-images.githubusercontent.com/102885424/161407279-8a2aabc2-e8fc-47a0-afd9-ad632b0d7e9c.png) Daytime with one turret on, one turret off **Desktop (please complete the following information):** - OS: Windows - Game Version 1.0.2 - Steam Version **Additional context** Other solar powered houses have multiple fully working turrets.
test
turrets not working even though electricity needs are met describe the bug turrets not working even though electricity needs are met to reproduce steps to reproduce the behavior fully upgrade house down sustainability path install one turret turret does not work during night install second turret second turret does not work during the day expected behavior meeting electricity requirements would mean that turrets work screenshots fully upgraded sustainable power path turrets stop working at night daytime with one turret on one turret off desktop please complete the following information os windows game version steam version additional context other solar powered houses have multiple fully working turrets
1
34,692
4,940,139,814
IssuesEvent
2016-11-29 16:06:44
flow123d/flow123d
https://api.github.com/repos/flow123d/flow123d
opened
Integration test update
New Tests
Go through all test input files: - Add header in form: # Test goal: # Author: # Source for reference: {text only, analytical, other SW, qualitative} - update description - [ ] Flow tests. - [ ] Transport tests. - [ ] Other tests.
1.0
Integration test update - Go through all test input files: - Add header in form: # Test goal: # Author: # Source for reference: {text only, analytical, other SW, qualitative} - update description - [ ] Flow tests. - [ ] Transport tests. - [ ] Other tests.
test
integration test update go through all test input files add header in form test goal author source for reference text only analytical other sw qualitative update description flow tests transport tests other tests
1
272,933
23,714,256,507
IssuesEvent
2022-08-30 10:22:58
marco-at-bit/test-quality-sdk
https://api.github.com/repos/marco-at-bit/test-quality-sdk
closed
Testing out status
bug TestQuality
## Precondition This condition must be met ## Steps to Reproduce: ### Step 1 <span style="color:#FCF0BA"> **[Retest]** </span> Step uno #### Expected Result Goes 1 ### Step 2 <span style="color:#58b880"> **[Pass]** </span> Second goes #### Expected Result Goes thwo ### Step 3 <span style="color:#58b880"> **[Pass]** </span> The classic third #### Expected Result goe three
1.0
Testing out status - ## Precondition This condition must be met ## Steps to Reproduce: ### Step 1 <span style="color:#FCF0BA"> **[Retest]** </span> Step uno #### Expected Result Goes 1 ### Step 2 <span style="color:#58b880"> **[Pass]** </span> Second goes #### Expected Result Goes thwo ### Step 3 <span style="color:#58b880"> **[Pass]** </span> The classic third #### Expected Result goe three
test
testing out status precondition this condition must be met steps to reproduce step step uno expected result goes step second goes expected result goes thwo step the classic third expected result goe three
1
91,760
10,727,992,907
IssuesEvent
2019-10-28 13:02:16
Timotheeee/PSIT3-HS19-IT18a_ZH-TeamIP2
https://api.github.com/repos/Timotheeee/PSIT3-HS19-IT18a_ZH-TeamIP2
closed
Lösungsarchitektur (M2)
documentation help wanted
− Use-Case-Modell − Zusätzliche Anforderungen − Domänenmodell − Softwarearchitektur − Design-Artefakte − Implementation − Projektmanagement − Glossar
1.0
Lösungsarchitektur (M2) - − Use-Case-Modell − Zusätzliche Anforderungen − Domänenmodell − Softwarearchitektur − Design-Artefakte − Implementation − Projektmanagement − Glossar
non_test
lösungsarchitektur − use case modell − zusätzliche anforderungen − domänenmodell − softwarearchitektur − design artefakte − implementation − projektmanagement − glossar
0
321,637
27,545,163,472
IssuesEvent
2023-03-07 11:13:49
giantswarm/roadmap
https://api.github.com/repos/giantswarm/roadmap
opened
Automated testing of CAPI management clusters
kind/epic area/kaas epic/capi component/mc-bootstrap topic/testing
## User Story - As a Giant Swarm Engineer I want to run tests on Giant Swarm management clusters so that I can ensure their quality and reliability. ## Details, Background Previous efforts from team Rocket around this topic can be tracked [here](https://github.com/giantswarm/roadmap/issues/685). ```[tasklist] - [ ] https://github.com/giantswarm/roadmap/issues/2118 ```
1.0
Automated testing of CAPI management clusters - ## User Story - As a Giant Swarm Engineer I want to run tests on Giant Swarm management clusters so that I can ensure their quality and reliability. ## Details, Background Previous efforts from team Rocket around this topic can be tracked [here](https://github.com/giantswarm/roadmap/issues/685). ```[tasklist] - [ ] https://github.com/giantswarm/roadmap/issues/2118 ```
test
automated testing of capi management clusters user story as a giant swarm engineer i want to run tests on giant swarm management clusters so that i can ensure their quality and reliability details background previous efforts from team rocket around this topic can be tracked
1
27,699
2,695,233,688
IssuesEvent
2015-04-02 02:48:45
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
opened
smooth centerline not perfect at edge
enhancement priority: high sct_straighten_spinalcord
Data: ~~~~ git clone https://github.com/neuropoly/sct_example_data.git ~~~~ Syntax: ~~~~ sct_straighten_spinalcord -i t2.nii.gz -c t2_seg.nii.gz -v 3 ~~~~ Output: ![screen shot 2015-04-01 at 20 27 46](https://cloud.githubusercontent.com/assets/2482071/6956781/39694286-d8c1-11e4-91d8-cf97860783ed.png)
1.0
smooth centerline not perfect at edge - Data: ~~~~ git clone https://github.com/neuropoly/sct_example_data.git ~~~~ Syntax: ~~~~ sct_straighten_spinalcord -i t2.nii.gz -c t2_seg.nii.gz -v 3 ~~~~ Output: ![screen shot 2015-04-01 at 20 27 46](https://cloud.githubusercontent.com/assets/2482071/6956781/39694286-d8c1-11e4-91d8-cf97860783ed.png)
non_test
smooth centerline not perfect at edge data git clone syntax sct straighten spinalcord i nii gz c seg nii gz v output
0
368,221
25,782,754,444
IssuesEvent
2022-12-09 17:21:54
Peruz/icsd
https://api.github.com/repos/Peruz/icsd
opened
Create config file for CI tests with python 3.9 and 3.10
documentation enhancement
Up to now, the coverage test using the [CI workflow](https://github.com/Peruz/icsd/blob/main/.github/workflows/tests_package.yml) is **only done for python 3.8 with os ubuntu**. We need to **extend it to 3.9 and 3.10 and to macos and windows os**. Adding the following two files in the test folder: - tests/envs/39-env.yaml ``` name: test channels: - conda-forge dependencies: - python=3.9 - ... ``` - tests/envs/310-env.yaml ``` name: test channels: - conda-forge dependencies: - python=3.10 - ... ``` And the following strategy command in the [CI workflow](https://github.com/Peruz/icsd/blob/main/.github/workflows/tests_package.yml) ``` strategy: fail-fast: false matrix: include: - os: ubuntu-latest env: tests/envs/38-env.yaml py: py3.8 - os: ubuntu-latest env: tests/envs/39-env.yaml py: py3.9 - os: ubuntu-latest env: tests/envs/310-env.yaml py: py3.10 - os: macos-latest env: tests/envs/310-env.yaml py: py3.10 - os: windows-latest env: tests/envs/310-env.yaml py: py3.10 ```
1.0
Create config file for CI tests with python 3.9 and 3.10 - Up to now, the coverage test using the [CI workflow](https://github.com/Peruz/icsd/blob/main/.github/workflows/tests_package.yml) is **only done for python 3.8 with os ubuntu**. We need to **extend it to 3.9 and 3.10 and to macos and windows os**. Adding the following two files in the test folder: - tests/envs/39-env.yaml ``` name: test channels: - conda-forge dependencies: - python=3.9 - ... ``` - tests/envs/310-env.yaml ``` name: test channels: - conda-forge dependencies: - python=3.10 - ... ``` And the following strategy command in the [CI workflow](https://github.com/Peruz/icsd/blob/main/.github/workflows/tests_package.yml) ``` strategy: fail-fast: false matrix: include: - os: ubuntu-latest env: tests/envs/38-env.yaml py: py3.8 - os: ubuntu-latest env: tests/envs/39-env.yaml py: py3.9 - os: ubuntu-latest env: tests/envs/310-env.yaml py: py3.10 - os: macos-latest env: tests/envs/310-env.yaml py: py3.10 - os: windows-latest env: tests/envs/310-env.yaml py: py3.10 ```
non_test
create config file for ci tests with python and up to now the coverage test using the is only done for python with os ubuntu we need to extend it to and and to macos and windows os adding the following two files in the test folder tests envs env yaml name test channels conda forge dependencies python tests envs env yaml name test channels conda forge dependencies python and the following strategy command in the strategy fail fast false matrix include os ubuntu latest env tests envs env yaml py os ubuntu latest env tests envs env yaml py os ubuntu latest env tests envs env yaml py os macos latest env tests envs env yaml py os windows latest env tests envs env yaml py
0
156,396
19,849,669,405
IssuesEvent
2022-01-21 10:50:13
billmcchesney1/jazz
https://api.github.com/repos/billmcchesney1/jazz
opened
CVE-2022-21704 (Medium) detected in log4js-0.6.38.tgz, log4js-4.5.1.tgz
security vulnerability
## CVE-2022-21704 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>log4js-0.6.38.tgz</b>, <b>log4js-4.5.1.tgz</b></p></summary> <p> <details><summary><b>log4js-0.6.38.tgz</b></p></summary> <p>Port of Log4js to work with node.</p> <p>Library home page: <a href="https://registry.npmjs.org/log4js/-/log4js-0.6.38.tgz">https://registry.npmjs.org/log4js/-/log4js-0.6.38.tgz</a></p> <p>Path to dependency file: /core/jazz_ui/src/app/primary-components/daterange-picker/ng2-datepicker/package.json</p> <p>Path to vulnerable library: /core/jazz_ui/src/app/primary-components/daterange-picker/ng2-datepicker/node_modules/log4js/package.json,/core/jazz_ui/node_modules/log4js/package.json</p> <p> Dependency Hierarchy: - karma-1.4.1.tgz (Root Library) - :x: **log4js-0.6.38.tgz** (Vulnerable Library) </details> <details><summary><b>log4js-4.5.1.tgz</b></p></summary> <p>Port of Log4js to work with node.</p> <p>Library home page: <a href="https://registry.npmjs.org/log4js/-/log4js-4.5.1.tgz">https://registry.npmjs.org/log4js/-/log4js-4.5.1.tgz</a></p> <p>Path to dependency file: /templates/angular-website-template/app/package.json</p> <p>Path to vulnerable library: /templates/angular-website-template/app/node_modules/log4js/package.json</p> <p> Dependency Hierarchy: - karma-4.0.1.tgz (Root Library) - :x: **log4js-4.5.1.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> log4js-node is a port of log4js to node.js. In affected versions default file permissions for log files created by the file, fileSync and dateFile appenders are world-readable (in unix). This could cause problems if log files contain sensitive information. This would affect any users that have not supplied their own permissions for the files via the mode parameter in the config. Users are advised to update. <p>Publish Date: 2022-01-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21704>CVE-2022-21704</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/log4js-node/log4js-node/security/advisories/GHSA-82v2-mx6x-wq7q">https://github.com/log4js-node/log4js-node/security/advisories/GHSA-82v2-mx6x-wq7q</a></p> <p>Release Date: 2022-01-19</p> <p>Fix Resolution: log4js - 6.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"log4js","packageVersion":"0.6.38","packageFilePaths":["/core/jazz_ui/src/app/primary-components/daterange-picker/ng2-datepicker/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma:1.4.1;log4js:0.6.38","isMinimumFixVersionAvailable":true,"minimumFixVersion":"log4js - 6.4.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"log4js","packageVersion":"4.5.1","packageFilePaths":["/templates/angular-website-template/app/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma:4.0.1;log4js:4.5.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"log4js - 6.4.0","isBinary":false}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2022-21704","vulnerabilityDetails":"log4js-node is a port of log4js to node.js. In affected versions default file permissions for log files created by the file, fileSync and dateFile appenders are world-readable (in unix). This could cause problems if log files contain sensitive information. This would affect any users that have not supplied their own permissions for the files via the mode parameter in the config. Users are advised to update.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21704","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2022-21704 (Medium) detected in log4js-0.6.38.tgz, log4js-4.5.1.tgz - ## CVE-2022-21704 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>log4js-0.6.38.tgz</b>, <b>log4js-4.5.1.tgz</b></p></summary> <p> <details><summary><b>log4js-0.6.38.tgz</b></p></summary> <p>Port of Log4js to work with node.</p> <p>Library home page: <a href="https://registry.npmjs.org/log4js/-/log4js-0.6.38.tgz">https://registry.npmjs.org/log4js/-/log4js-0.6.38.tgz</a></p> <p>Path to dependency file: /core/jazz_ui/src/app/primary-components/daterange-picker/ng2-datepicker/package.json</p> <p>Path to vulnerable library: /core/jazz_ui/src/app/primary-components/daterange-picker/ng2-datepicker/node_modules/log4js/package.json,/core/jazz_ui/node_modules/log4js/package.json</p> <p> Dependency Hierarchy: - karma-1.4.1.tgz (Root Library) - :x: **log4js-0.6.38.tgz** (Vulnerable Library) </details> <details><summary><b>log4js-4.5.1.tgz</b></p></summary> <p>Port of Log4js to work with node.</p> <p>Library home page: <a href="https://registry.npmjs.org/log4js/-/log4js-4.5.1.tgz">https://registry.npmjs.org/log4js/-/log4js-4.5.1.tgz</a></p> <p>Path to dependency file: /templates/angular-website-template/app/package.json</p> <p>Path to vulnerable library: /templates/angular-website-template/app/node_modules/log4js/package.json</p> <p> Dependency Hierarchy: - karma-4.0.1.tgz (Root Library) - :x: **log4js-4.5.1.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> log4js-node is a port of log4js to node.js. In affected versions default file permissions for log files created by the file, fileSync and dateFile appenders are world-readable (in unix). This could cause problems if log files contain sensitive information. This would affect any users that have not supplied their own permissions for the files via the mode parameter in the config. Users are advised to update. <p>Publish Date: 2022-01-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21704>CVE-2022-21704</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/log4js-node/log4js-node/security/advisories/GHSA-82v2-mx6x-wq7q">https://github.com/log4js-node/log4js-node/security/advisories/GHSA-82v2-mx6x-wq7q</a></p> <p>Release Date: 2022-01-19</p> <p>Fix Resolution: log4js - 6.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"log4js","packageVersion":"0.6.38","packageFilePaths":["/core/jazz_ui/src/app/primary-components/daterange-picker/ng2-datepicker/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma:1.4.1;log4js:0.6.38","isMinimumFixVersionAvailable":true,"minimumFixVersion":"log4js - 6.4.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"log4js","packageVersion":"4.5.1","packageFilePaths":["/templates/angular-website-template/app/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma:4.0.1;log4js:4.5.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"log4js - 6.4.0","isBinary":false}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2022-21704","vulnerabilityDetails":"log4js-node is a port of log4js to node.js. In affected versions default file permissions for log files created by the file, fileSync and dateFile appenders are world-readable (in unix). This could cause problems if log files contain sensitive information. This would affect any users that have not supplied their own permissions for the files via the mode parameter in the config. Users are advised to update.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21704","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in tgz tgz cve medium severity vulnerability vulnerable libraries tgz tgz tgz port of to work with node library home page a href path to dependency file core jazz ui src app primary components daterange picker datepicker package json path to vulnerable library core jazz ui src app primary components daterange picker datepicker node modules package json core jazz ui node modules package json dependency hierarchy karma tgz root library x tgz vulnerable library tgz port of to work with node library home page a href path to dependency file templates angular website template app package json path to vulnerable library templates angular website template app node modules package json dependency hierarchy karma tgz root library x tgz vulnerable library found in base branch develop vulnerability details node is a port of to node js in affected versions default file permissions for log files created by the file filesync and datefile appenders are world readable in unix this could cause problems if log files contain sensitive information this would affect any users that have not supplied their own permissions for the files via the mode parameter in the config users are advised to update publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree karma isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename packageversion packagefilepaths istransitivedependency true dependencytree karma isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails node is a port of to node js in affected versions default file permissions for log files created by the file filesync and datefile appenders are world readable in unix this could cause problems if log files contain sensitive information this would affect any users that have not supplied their own permissions for the files via the mode parameter in the config users are advised to update vulnerabilityurl
0
599,349
18,271,618,090
IssuesEvent
2021-10-04 14:21:12
ita-social-projects/TeachUA
https://api.github.com/repos/ita-social-projects/TeachUA
closed
[Advanced Search] "Вік дитини" field can accept unlimited letters via 'paste' function
bug Priority: Medium
**Environment:** Windows 10 21H1 64bit, Chrome Version 93.0.4577.63 (Official Build) (64-bit) **Reproducible:** always **Build found:** latest on dev and prod **Preconditions** https://speak-ukrainian.org.ua/dev/ **Steps to reproduce** 1. Go to https://speak-ukrainian.org.ua/dev/ 2. Click on advanced search button 3. Paste 300+ letters into "Вік дитини" field **Actual result** Unlimited letters can be paste into the "Вік дитини" field **Expected result** "Вік дитини" field should not allow to paste letters **Labels to be added** "Bug", Priority ("pri: mid"), Severity ("severity:low"), Type ("Functional"), "API" (for back-end bugs).
1.0
[Advanced Search] "Вік дитини" field can accept unlimited letters via 'paste' function - **Environment:** Windows 10 21H1 64bit, Chrome Version 93.0.4577.63 (Official Build) (64-bit) **Reproducible:** always **Build found:** latest on dev and prod **Preconditions** https://speak-ukrainian.org.ua/dev/ **Steps to reproduce** 1. Go to https://speak-ukrainian.org.ua/dev/ 2. Click on advanced search button 3. Paste 300+ letters into "Вік дитини" field **Actual result** Unlimited letters can be paste into the "Вік дитини" field **Expected result** "Вік дитини" field should not allow to paste letters **Labels to be added** "Bug", Priority ("pri: mid"), Severity ("severity:low"), Type ("Functional"), "API" (for back-end bugs).
non_test
вік дитини field can accept unlimited letters via paste function environment windows chrome version official build bit reproducible always build found latest on dev and prod preconditions steps to reproduce go to click on advanced search button paste letters into вік дитини field actual result unlimited letters can be paste into the вік дитини field expected result вік дитини field should not allow to paste letters labels to be added bug priority pri mid severity severity low type functional api for back end bugs
0
163,648
12,739,860,418
IssuesEvent
2020-06-26 00:29:25
jensengroup/propka
https://api.github.com/repos/jensengroup/propka
closed
report coverage from CI
continuous integration testing
With PR #62 merged, we should use coverage reporting and make it part of the CI reporting.
1.0
report coverage from CI - With PR #62 merged, we should use coverage reporting and make it part of the CI reporting.
test
report coverage from ci with pr merged we should use coverage reporting and make it part of the ci reporting
1
2,834
2,641,060,798
IssuesEvent
2015-03-11 15:49:51
bpmn-miwg/bpmn-miwg-test-suite
https://api.github.com/repos/bpmn-miwg/bpmn-miwg-test-suite
closed
Participant missing in collaboration
bug Test: A.4.0
The model A.4.0 has two processes, obviously collaborating. The collaboration element has only one participant. The specification has 3 relevant statements to be considered for this issue: - “One, and only one, Pool in a diagram MAY be presented without a boundary. If there is more than one Pool in the diagram, then the remaining Pools MUST have a boundary” - “A Collaboration can contain two (2) or more Pools (i.e., Participants). However, a Process that represents the work performed from the point of view of the modeler or the modeler’s organization can be considered “internal” and is NOT REQUIRED to be surrounded by the boundary of the Pool, while the other Pools in the Diagram MUST have their boundary (see Figure 9.5).” - "A Pool is the graphical representation of a Participant in a Collaboration." According to the aforementioned statements, the collaboration element should have two participants, but no BPMNShape representing the second participant as the process is considered “internal”.
1.0
Participant missing in collaboration - The model A.4.0 has two processes, obviously collaborating. The collaboration element has only one participant. The specification has 3 relevant statements to be considered for this issue: - “One, and only one, Pool in a diagram MAY be presented without a boundary. If there is more than one Pool in the diagram, then the remaining Pools MUST have a boundary” - “A Collaboration can contain two (2) or more Pools (i.e., Participants). However, a Process that represents the work performed from the point of view of the modeler or the modeler’s organization can be considered “internal” and is NOT REQUIRED to be surrounded by the boundary of the Pool, while the other Pools in the Diagram MUST have their boundary (see Figure 9.5).” - "A Pool is the graphical representation of a Participant in a Collaboration." According to the aforementioned statements, the collaboration element should have two participants, but no BPMNShape representing the second participant as the process is considered “internal”.
test
participant missing in collaboration the model a has two processes obviously collaborating the collaboration element has only one participant the specification has relevant statements to be considered for this issue “one and only one pool in a diagram may be presented without a boundary if there is more than one pool in the diagram then the remaining pools must have a boundary” “a collaboration can contain two or more pools i e participants however a process that represents the work performed from the point of view of the modeler or the modeler’s organization can be considered “internal” and is not required to be surrounded by the boundary of the pool while the other pools in the diagram must have their boundary see figure ” a pool is the graphical representation of a participant in a collaboration according to the aforementioned statements the collaboration element should have two participants but no bpmnshape representing the second participant as the process is considered “internal”
1
138,318
11,198,884,803
IssuesEvent
2020-01-03 17:06:28
Princeton-CDH/ppa-django
https://api.github.com/repos/Princeton-CDH/ppa-django
closed
As a content editor, I want to create linkable anchors in documents so that I can reference specific sections of my content on other pages.
awaiting testing
## testing notes This feature was adapted from mep-django, where it has already been tested and accepted. See testing notes and comments on https://github.com/Princeton-CDH/mep-django/issues/464 Probably useful to run through similar testing steps to make sure you're comfortable with it, and also check that the display for PPA looks ok (link styles are slightly different). Feel free to compare with the mep-django version of the feature in production; it's currently in use for the FAQ sections: https://shakespeareandco.princeton.edu/about/faq/ I thought this feature would be valuable to bring over to use with the most/least collected items listed in my essay, so that we can link directly to them and make them easier to refer to.
1.0
As a content editor, I want to create linkable anchors in documents so that I can reference specific sections of my content on other pages. - ## testing notes This feature was adapted from mep-django, where it has already been tested and accepted. See testing notes and comments on https://github.com/Princeton-CDH/mep-django/issues/464 Probably useful to run through similar testing steps to make sure you're comfortable with it, and also check that the display for PPA looks ok (link styles are slightly different). Feel free to compare with the mep-django version of the feature in production; it's currently in use for the FAQ sections: https://shakespeareandco.princeton.edu/about/faq/ I thought this feature would be valuable to bring over to use with the most/least collected items listed in my essay, so that we can link directly to them and make them easier to refer to.
test
as a content editor i want to create linkable anchors in documents so that i can reference specific sections of my content on other pages testing notes this feature was adapted from mep django where it has already been tested and accepted see testing notes and comments on probably useful to run through similar testing steps to make sure you re comfortable with it and also check that the display for ppa looks ok link styles are slightly different feel free to compare with the mep django version of the feature in production it s currently in use for the faq sections i thought this feature would be valuable to bring over to use with the most least collected items listed in my essay so that we can link directly to them and make them easier to refer to
1
42,783
22,839,710,948
IssuesEvent
2022-07-12 20:24:53
pulumi/pulumi
https://api.github.com/repos/pulumi/pulumi
closed
[Python SDK] slow performance with large CRD in helm chart.
kind/bug impact/performance language/python
<!--- Provide a general summary of the issue --> ## Expected Behavior <!--- Tell us what should happen --> `pulumi up` is fast just like in typescript ~5 seconds [Typescript SDK TRACE out.txt](https://github.com/pulumi/pulumi/files/5795889/out.txt) ## Current Behavior <!--- Tell us what happens instead of the expected behavior --> `pulumi up` with python is very very slow ~1 hour ! [Python SDK TRACE out.txt.zip](https://github.com/pulumi/pulumi/files/5796026/out.txt.zip) (I didn't wait and canceled the command `pulumi up`, very long wait) ## Steps to Reproduce <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> 1) Chart with a large CRD yaml size ~10k lines Example: https://github.com/VictoriaMetrics/helm-charts/blob/master/charts/victoria-metrics-operator/templates/crd.yaml https://github.com/VictoriaMetrics/helm-charts/tree/master/charts/victoria-metrics-operator 2) use python sdk to install chart: ``` from pulumi_kubernetes.helm import v3 from pulumi_kubernetes.helm.v3.helm import ChartOpts vm_chart = v3.Chart("vm", ChartOpts(chart="victoria-metrics-operator", repo="vm")) ``` ## Context (Environment) <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> <!--- Provide a general summary of the issue in the Title above --> ``` macOS Big Sur 11.1 Python 3.9.1 pulumi v2.17.0 pulumi-kubernetes==2.7.4 ```
True
[Python SDK] slow performance with large CRD in helm chart. - <!--- Provide a general summary of the issue --> ## Expected Behavior <!--- Tell us what should happen --> `pulumi up` is fast just like in typescript ~5 seconds [Typescript SDK TRACE out.txt](https://github.com/pulumi/pulumi/files/5795889/out.txt) ## Current Behavior <!--- Tell us what happens instead of the expected behavior --> `pulumi up` with python is very very slow ~1 hour ! [Python SDK TRACE out.txt.zip](https://github.com/pulumi/pulumi/files/5796026/out.txt.zip) (I didn't wait and canceled the command `pulumi up`, very long wait) ## Steps to Reproduce <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> 1) Chart with a large CRD yaml size ~10k lines Example: https://github.com/VictoriaMetrics/helm-charts/blob/master/charts/victoria-metrics-operator/templates/crd.yaml https://github.com/VictoriaMetrics/helm-charts/tree/master/charts/victoria-metrics-operator 2) use python sdk to install chart: ``` from pulumi_kubernetes.helm import v3 from pulumi_kubernetes.helm.v3.helm import ChartOpts vm_chart = v3.Chart("vm", ChartOpts(chart="victoria-metrics-operator", repo="vm")) ``` ## Context (Environment) <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> <!--- Provide a general summary of the issue in the Title above --> ``` macOS Big Sur 11.1 Python 3.9.1 pulumi v2.17.0 pulumi-kubernetes==2.7.4 ```
non_test
slow performance with large crd in helm chart expected behavior pulumi up is fast just like in typescript seconds current behavior pulumi up with python is very very slow hour i didn t wait and canceled the command pulumi up very long wait steps to reproduce chart with a large crd yaml size lines example use python sdk to install chart from pulumi kubernetes helm import from pulumi kubernetes helm helm import chartopts vm chart chart vm chartopts chart victoria metrics operator repo vm context environment macos big sur python pulumi pulumi kubernetes
0
773,516
27,160,398,636
IssuesEvent
2023-02-17 11:21:21
status-im/status-mobile
https://api.github.com/repos/status-im/status-mobile
closed
Some components disappear from the list of pinned messages after chat reopening
bug low-priority pinned messages
**Steps to reproduce:** 1. Go to the 1-1 chat 2. Send a message 3. Pin the message 4. Reopen the chat **Actual result:** The following info is dissapaered: - Owner and his chatkey - Timemark - Pin - Style for pinned messages https://user-images.githubusercontent.com/52490791/217331367-b65df057-1dd0-43ce-baeb-5cdddedd8dd8.mp4 **Expected result:** ![image](https://user-images.githubusercontent.com/52490791/217331508-07262846-f0d5-4985-b9da-929d9bd9776d.png) **ENV:** - Nightly 7 Feb 2023 [2619]
1.0
Some components disappear from the list of pinned messages after chat reopening - **Steps to reproduce:** 1. Go to the 1-1 chat 2. Send a message 3. Pin the message 4. Reopen the chat **Actual result:** The following info is dissapaered: - Owner and his chatkey - Timemark - Pin - Style for pinned messages https://user-images.githubusercontent.com/52490791/217331367-b65df057-1dd0-43ce-baeb-5cdddedd8dd8.mp4 **Expected result:** ![image](https://user-images.githubusercontent.com/52490791/217331508-07262846-f0d5-4985-b9da-929d9bd9776d.png) **ENV:** - Nightly 7 Feb 2023 [2619]
non_test
some components disappear from the list of pinned messages after chat reopening steps to reproduce go to the chat send a message pin the message reopen the chat actual result the following info is dissapaered owner and his chatkey timemark pin style for pinned messages expected result env nightly feb
0
797,730
28,153,593,236
IssuesEvent
2023-04-03 05:04:43
PrefectHQ/prefect
https://api.github.com/repos/PrefectHQ/prefect
closed
Subpaths and remote subdirectory paths can't be used on deployments with GCS storage block
bug status:accepted v2 priority:medium
### First check - [X] I added a descriptive title to this issue. - [X] I used the GitHub search to find a similar issue and didn't find it. - [X] I searched the Prefect documentation for this issue. - [X] I checked that this issue is related to Prefect and not one of its dependencies. ### Bug summary If you create a deployment with a storage block that has a directory as part of its "base path" _and_ a `path`, the files are not downloaded correctly from the storage and the flow run fails. (GCS) Storage Blocks can have a directory added [as part of their bucket path](https://docs.prefect.io/tutorials/storage/#create-a-storage-block). [As described in the tutorial](https://docs.prefect.io/tutorials/storage/#sharing-storage), you can also share Storage Blocks between deployments and put each deployment's file in its own subdirectory by making use of `path` on a deployment. --- I also tested by removing the directory from the "base path". Then it runs fine. When I add a `path`, it still runs fine. When I add another folder level, it fails (see Reproduction method 2). ### Reproduction There's two ways to get multiple dir levels as part of the flow storage. #### Method 1: I followed the tutorial (https://docs.prefect.io/tutorials/deployments/#setting-up). Differences: - I'm running inside a Poetry venv. - My GCS block has an additional directory as part of its base path: ![image](https://user-images.githubusercontent.com/6089256/191446769-a9d7960e-e6d2-470a-987a-72f37f51418e.png) Everything works fine if the deployment does not have a `path` specified: ``` poetry run prefect deployment build ./log_flow.py:log_flow -n log-simple -q test -sb gcs/poc-gcs-bucket-dev --apply ``` When adding a `path`, then running a flow from that deployment, fails. ``` poetry run prefect deployment build ./log_flow.py:log_flow -n log-simple -q test -sb gcs/poc-gcs-bucket-dev --path log-flow-directory --apply ``` Errors posted in error section. --- #### Method 2: Remove the base path from the storage block. Add more than one level to the path here: ``` poetry run prefect deployment build ./log_flow.py:log_flow -n log-simple -q test -sb gcs/poc-gcs-bucket-dev --path testing/log-flow-directory --apply ``` ### Error <details> <summary>On Windows</summary> ``` 09:37:13.657 | INFO | prefect.agent - Submitting flow run '926654d1-b5fc-4679-a0e6-01acf2bb70c2' 09:37:14.538 | INFO | prefect.infrastructure.process - Opening process 'omicron3-firiendfeld-point'... 09:37:14.573 | INFO | prefect.agent - Completed submission of flow run '926654d1-b5fc-4679-a0e6-01acf2bb70c2' 09:37:19.042 | ERROR | Flow run 'omicron3-firiendfeld-point' - Flow could not be retrieved from deployment. Traceback (most recent call last): File "<frozen importlib._bootstrap_external>", line 879, in exec_module File "<frozen importlib._bootstrap_external>", line 1016, in get_code File "<frozen importlib._bootstrap_external>", line 1073, in get_data FileNotFoundError: [Errno 2] No such file or directory: 'C:\\Users\\STALJA~1\\AppData\\Local\\Temp\\tmpal0n0yjbprefect\\log_flow.py' The above exception was the direct cause of the following exception: Traceback (most recent call last): File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\engine.py", line 256, in retrieve_flow_then_begin_flow_run flow = await load_flow_from_flow_run(flow_run, client=client) File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\client.py", line 103, in with_injected_client return await fn(*args, **kwargs) File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\deployments.py", line 69, in load_flow_from_flow_run flow = await run_sync_in_worker_thread(import_object, str(import_path)) File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\utilities\asyncutils.py", line 57, in run_sync_in_worker_thread return await anyio.to_thread.run_sync(call, cancellable=True) File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\anyio\to_thread.py", line 31, in run_sync return await get_asynclib().run_sync_in_worker_thread( File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\anyio\_backends\_asyncio.py", line 937, in run_sync_in_worker_thread return await future File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\anyio\_backends\_asyncio.py", line 867, in run result = context.run(func, *args) File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\utilities\importtools.py", line 193, in import_object module = load_script_as_module(script_path) File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\utilities\importtools.py", line 156, in load_script_as_module raise ScriptError(user_exc=exc, path=path) from exc prefect.exceptions.ScriptError: Script at 'log_flow.py' encountered an exception 09:37:20.008 | INFO | prefect.infrastructure.process - Process 'omicron3-firiendfeld-point' exited cleanly. ``` </details> <details> <summary>On Ubuntu</summary> ``` 08:46:12.139 | INFO | prefect.infrastructure.process - Opening process 'eta2-nivoch-nebula'... 08:46:12.143 | INFO | prefect.agent - Completed submission of flow run '8116942b-acfb-495b-8386-975d0aa6980e' 08:46:16.120 | ERROR | Flow run 'eta2-nivoch-nebula' - Flow could not be retrieved from deployment. Traceback (most recent call last): File "<frozen importlib._bootstrap_external>", line 879, in exec_module File "<frozen importlib._bootstrap_external>", line 1016, in get_code File "<frozen importlib._bootstrap_external>", line 1073, in get_data FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmp8lc3zuayprefect/log_flow.py' The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/engine.py", line 256, in retrieve_flow_then_begin_flow_run flow = await load_flow_from_flow_run(flow_run, client=client) File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/client.py", line 103, in with_injected_client return await fn(*args, **kwargs) File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/deployments.py", line 69, in load_flow_from_flow_run flow = await run_sync_in_worker_thread(import_object, str(import_path)) File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/utilities/asyncutils.py", line 57, in run_sync_in_worker_thread return await anyio.to_thread.run_sync(call, cancellable=True) File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/anyio/to_thread.py", line 31, in run_sync return await get_asynclib().run_sync_in_worker_thread( File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread return await future File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 867, in run result = context.run(func, *args) File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/utilities/importtools.py", line 193, in import_object module = load_script_as_module(script_path) File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/utilities/importtools.py", line 156, in load_script_as_module raise ScriptError(user_exc=exc, path=path) from exc prefect.exceptions.ScriptError: Script at 'log_flow.py' encountered an exception 08:46:17.028 | INFO | prefect.infrastructure.process - Process 'eta2-nivoch-nebula' exited cleanly. ``` </details> ### Versions <details> <summary>On Windows</summary> ``` Version: 2.4.0 API version: 0.8.0 Python version: 3.10.7 Git commit: 513639e8 Built: Tue, Sep 13, 2022 2:15 PM OS/Arch: win32/AMD64 Profile: eta-cloud Server type: cloud ``` </details> <details> <summary>On Ubuntu</summary> ``` Version: 2.4.0 API version: 0.8.0 Python version: 3.10.4 Git commit: 513639e8 Built: Tue, Sep 13, 2022 2:15 PM OS/Arch: linux/x86_64 Profile: cloud Server type: cloud ``` </details> ### Additional context I was only able to test using the GCS block. I tested on Windows and Ubuntu via WSL. I'm running everything inside the same Poetry venv.
1.0
Subpaths and remote subdirectory paths can't be used on deployments with GCS storage block - ### First check - [X] I added a descriptive title to this issue. - [X] I used the GitHub search to find a similar issue and didn't find it. - [X] I searched the Prefect documentation for this issue. - [X] I checked that this issue is related to Prefect and not one of its dependencies. ### Bug summary If you create a deployment with a storage block that has a directory as part of its "base path" _and_ a `path`, the files are not downloaded correctly from the storage and the flow run fails. (GCS) Storage Blocks can have a directory added [as part of their bucket path](https://docs.prefect.io/tutorials/storage/#create-a-storage-block). [As described in the tutorial](https://docs.prefect.io/tutorials/storage/#sharing-storage), you can also share Storage Blocks between deployments and put each deployment's file in its own subdirectory by making use of `path` on a deployment. --- I also tested by removing the directory from the "base path". Then it runs fine. When I add a `path`, it still runs fine. When I add another folder level, it fails (see Reproduction method 2). ### Reproduction There's two ways to get multiple dir levels as part of the flow storage. #### Method 1: I followed the tutorial (https://docs.prefect.io/tutorials/deployments/#setting-up). Differences: - I'm running inside a Poetry venv. - My GCS block has an additional directory as part of its base path: ![image](https://user-images.githubusercontent.com/6089256/191446769-a9d7960e-e6d2-470a-987a-72f37f51418e.png) Everything works fine if the deployment does not have a `path` specified: ``` poetry run prefect deployment build ./log_flow.py:log_flow -n log-simple -q test -sb gcs/poc-gcs-bucket-dev --apply ``` When adding a `path`, then running a flow from that deployment, fails. ``` poetry run prefect deployment build ./log_flow.py:log_flow -n log-simple -q test -sb gcs/poc-gcs-bucket-dev --path log-flow-directory --apply ``` Errors posted in error section. --- #### Method 2: Remove the base path from the storage block. Add more than one level to the path here: ``` poetry run prefect deployment build ./log_flow.py:log_flow -n log-simple -q test -sb gcs/poc-gcs-bucket-dev --path testing/log-flow-directory --apply ``` ### Error <details> <summary>On Windows</summary> ``` 09:37:13.657 | INFO | prefect.agent - Submitting flow run '926654d1-b5fc-4679-a0e6-01acf2bb70c2' 09:37:14.538 | INFO | prefect.infrastructure.process - Opening process 'omicron3-firiendfeld-point'... 09:37:14.573 | INFO | prefect.agent - Completed submission of flow run '926654d1-b5fc-4679-a0e6-01acf2bb70c2' 09:37:19.042 | ERROR | Flow run 'omicron3-firiendfeld-point' - Flow could not be retrieved from deployment. Traceback (most recent call last): File "<frozen importlib._bootstrap_external>", line 879, in exec_module File "<frozen importlib._bootstrap_external>", line 1016, in get_code File "<frozen importlib._bootstrap_external>", line 1073, in get_data FileNotFoundError: [Errno 2] No such file or directory: 'C:\\Users\\STALJA~1\\AppData\\Local\\Temp\\tmpal0n0yjbprefect\\log_flow.py' The above exception was the direct cause of the following exception: Traceback (most recent call last): File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\engine.py", line 256, in retrieve_flow_then_begin_flow_run flow = await load_flow_from_flow_run(flow_run, client=client) File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\client.py", line 103, in with_injected_client return await fn(*args, **kwargs) File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\deployments.py", line 69, in load_flow_from_flow_run flow = await run_sync_in_worker_thread(import_object, str(import_path)) File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\utilities\asyncutils.py", line 57, in run_sync_in_worker_thread return await anyio.to_thread.run_sync(call, cancellable=True) File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\anyio\to_thread.py", line 31, in run_sync return await get_asynclib().run_sync_in_worker_thread( File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\anyio\_backends\_asyncio.py", line 937, in run_sync_in_worker_thread return await future File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\anyio\_backends\_asyncio.py", line 867, in run result = context.run(func, *args) File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\utilities\importtools.py", line 193, in import_object module = load_script_as_module(script_path) File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\utilities\importtools.py", line 156, in load_script_as_module raise ScriptError(user_exc=exc, path=path) from exc prefect.exceptions.ScriptError: Script at 'log_flow.py' encountered an exception 09:37:20.008 | INFO | prefect.infrastructure.process - Process 'omicron3-firiendfeld-point' exited cleanly. ``` </details> <details> <summary>On Ubuntu</summary> ``` 08:46:12.139 | INFO | prefect.infrastructure.process - Opening process 'eta2-nivoch-nebula'... 08:46:12.143 | INFO | prefect.agent - Completed submission of flow run '8116942b-acfb-495b-8386-975d0aa6980e' 08:46:16.120 | ERROR | Flow run 'eta2-nivoch-nebula' - Flow could not be retrieved from deployment. Traceback (most recent call last): File "<frozen importlib._bootstrap_external>", line 879, in exec_module File "<frozen importlib._bootstrap_external>", line 1016, in get_code File "<frozen importlib._bootstrap_external>", line 1073, in get_data FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmp8lc3zuayprefect/log_flow.py' The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/engine.py", line 256, in retrieve_flow_then_begin_flow_run flow = await load_flow_from_flow_run(flow_run, client=client) File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/client.py", line 103, in with_injected_client return await fn(*args, **kwargs) File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/deployments.py", line 69, in load_flow_from_flow_run flow = await run_sync_in_worker_thread(import_object, str(import_path)) File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/utilities/asyncutils.py", line 57, in run_sync_in_worker_thread return await anyio.to_thread.run_sync(call, cancellable=True) File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/anyio/to_thread.py", line 31, in run_sync return await get_asynclib().run_sync_in_worker_thread( File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread return await future File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 867, in run result = context.run(func, *args) File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/utilities/importtools.py", line 193, in import_object module = load_script_as_module(script_path) File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/utilities/importtools.py", line 156, in load_script_as_module raise ScriptError(user_exc=exc, path=path) from exc prefect.exceptions.ScriptError: Script at 'log_flow.py' encountered an exception 08:46:17.028 | INFO | prefect.infrastructure.process - Process 'eta2-nivoch-nebula' exited cleanly. ``` </details> ### Versions <details> <summary>On Windows</summary> ``` Version: 2.4.0 API version: 0.8.0 Python version: 3.10.7 Git commit: 513639e8 Built: Tue, Sep 13, 2022 2:15 PM OS/Arch: win32/AMD64 Profile: eta-cloud Server type: cloud ``` </details> <details> <summary>On Ubuntu</summary> ``` Version: 2.4.0 API version: 0.8.0 Python version: 3.10.4 Git commit: 513639e8 Built: Tue, Sep 13, 2022 2:15 PM OS/Arch: linux/x86_64 Profile: cloud Server type: cloud ``` </details> ### Additional context I was only able to test using the GCS block. I tested on Windows and Ubuntu via WSL. I'm running everything inside the same Poetry venv.
non_test
subpaths and remote subdirectory paths can t be used on deployments with gcs storage block first check i added a descriptive title to this issue i used the github search to find a similar issue and didn t find it i searched the prefect documentation for this issue i checked that this issue is related to prefect and not one of its dependencies bug summary if you create a deployment with a storage block that has a directory as part of its base path and a path the files are not downloaded correctly from the storage and the flow run fails gcs storage blocks can have a directory added you can also share storage blocks between deployments and put each deployment s file in its own subdirectory by making use of path on a deployment i also tested by removing the directory from the base path then it runs fine when i add a path it still runs fine when i add another folder level it fails see reproduction method reproduction there s two ways to get multiple dir levels as part of the flow storage method i followed the tutorial differences i m running inside a poetry venv my gcs block has an additional directory as part of its base path everything works fine if the deployment does not have a path specified poetry run prefect deployment build log flow py log flow n log simple q test sb gcs poc gcs bucket dev apply when adding a path then running a flow from that deployment fails poetry run prefect deployment build log flow py log flow n log simple q test sb gcs poc gcs bucket dev path log flow directory apply errors posted in error section method remove the base path from the storage block add more than one level to the path here poetry run prefect deployment build log flow py log flow n log simple q test sb gcs poc gcs bucket dev path testing log flow directory apply error on windows info prefect agent submitting flow run info prefect infrastructure process opening process firiendfeld point info prefect agent completed submission of flow run error flow run firiendfeld point flow could not be retrieved from deployment traceback most recent call last file line in exec module file line in get code file line in get data filenotfounderror no such file or directory c users stalja appdata local temp log flow py the above exception was the direct cause of the following exception traceback most recent call last file c users staljaard source data integration poc venv lib site packages prefect engine py line in retrieve flow then begin flow run flow await load flow from flow run flow run client client file c users staljaard source data integration poc venv lib site packages prefect client py line in with injected client return await fn args kwargs file c users staljaard source data integration poc venv lib site packages prefect deployments py line in load flow from flow run flow await run sync in worker thread import object str import path file c users staljaard source data integration poc venv lib site packages prefect utilities asyncutils py line in run sync in worker thread return await anyio to thread run sync call cancellable true file c users staljaard source data integration poc venv lib site packages anyio to thread py line in run sync return await get asynclib run sync in worker thread file c users staljaard source data integration poc venv lib site packages anyio backends asyncio py line in run sync in worker thread return await future file c users staljaard source data integration poc venv lib site packages anyio backends asyncio py line in run result context run func args file c users staljaard source data integration poc venv lib site packages prefect utilities importtools py line in import object module load script as module script path file c users staljaard source data integration poc venv lib site packages prefect utilities importtools py line in load script as module raise scripterror user exc exc path path from exc prefect exceptions scripterror script at log flow py encountered an exception info prefect infrastructure process process firiendfeld point exited cleanly on ubuntu info prefect infrastructure process opening process nivoch nebula info prefect agent completed submission of flow run acfb error flow run nivoch nebula flow could not be retrieved from deployment traceback most recent call last file line in exec module file line in get code file line in get data filenotfounderror no such file or directory tmp log flow py the above exception was the direct cause of the following exception traceback most recent call last file home stephan di poc venv lib site packages prefect engine py line in retrieve flow then begin flow run flow await load flow from flow run flow run client client file home stephan di poc venv lib site packages prefect client py line in with injected client return await fn args kwargs file home stephan di poc venv lib site packages prefect deployments py line in load flow from flow run flow await run sync in worker thread import object str import path file home stephan di poc venv lib site packages prefect utilities asyncutils py line in run sync in worker thread return await anyio to thread run sync call cancellable true file home stephan di poc venv lib site packages anyio to thread py line in run sync return await get asynclib run sync in worker thread file home stephan di poc venv lib site packages anyio backends asyncio py line in run sync in worker thread return await future file home stephan di poc venv lib site packages anyio backends asyncio py line in run result context run func args file home stephan di poc venv lib site packages prefect utilities importtools py line in import object module load script as module script path file home stephan di poc venv lib site packages prefect utilities importtools py line in load script as module raise scripterror user exc exc path path from exc prefect exceptions scripterror script at log flow py encountered an exception info prefect infrastructure process process nivoch nebula exited cleanly versions on windows version api version python version git commit built tue sep pm os arch profile eta cloud server type cloud on ubuntu version api version python version git commit built tue sep pm os arch linux profile cloud server type cloud additional context i was only able to test using the gcs block i tested on windows and ubuntu via wsl i m running everything inside the same poetry venv
0
350,994
31,932,887,205
IssuesEvent
2023-09-19 08:37:08
bazelbuild/intellij
https://api.github.com/repos/bazelbuild/intellij
closed
"Illegal Character" trying to debug go_test, and passing tests don't show up when run.
type: bug P3 lang: go product: GoLand topic: testing
The problem occurs when run a test in Goland from the "Run/Debug configurations" using bazel test. If I just run, then the tests run and I can see the results on the console, but the test navigator says "No tests were found" (if they all pass) ![No Tests Found - Goland](https://user-images.githubusercontent.com/7515022/166160166-0681bf32-64a3-4e8d-a576-68e5a5574916.png) Note that I have "show passing tests" selected, so the passing tests are expected to show up. Interestingly, if I change one of the tests to fail, I get... ![Failing Test Shows Up](https://user-images.githubusercontent.com/7515022/166160244-be29538d-d55a-48cd-a94e-007d7ea13ade.png) so the failing test shows up and the passing ones do not. But the original issue I had is that I'm not able to debug. If I click the debug button I get: ![Illegal Character](https://user-images.githubusercontent.com/7515022/166160833-e8e7c1b3-465b-4064-830f-e9dae653d72f.png) I tried upgrading the bazel plugin (and Goland) since I saw that fixing debugging with `go_test` was on the list of recent fixes, but the issue persists after all of that, plus doing `bazel clean` and even nuking my output-user-root directory to force everything to rebuild. Also note that the illegal character remains at index 116 even if I change the test to have everything commented out, leaving only one empty test method. Throughout all of this, `bazel build` and `bazel test` continue to work from the terminal as expected. Versions: Goland: 2022.1 Bazelisk: 1.11.0 Bazel: 5.1.1 Bazel Plugin: 2022.04.28.0.0-api-version-221
1.0
"Illegal Character" trying to debug go_test, and passing tests don't show up when run. - The problem occurs when run a test in Goland from the "Run/Debug configurations" using bazel test. If I just run, then the tests run and I can see the results on the console, but the test navigator says "No tests were found" (if they all pass) ![No Tests Found - Goland](https://user-images.githubusercontent.com/7515022/166160166-0681bf32-64a3-4e8d-a576-68e5a5574916.png) Note that I have "show passing tests" selected, so the passing tests are expected to show up. Interestingly, if I change one of the tests to fail, I get... ![Failing Test Shows Up](https://user-images.githubusercontent.com/7515022/166160244-be29538d-d55a-48cd-a94e-007d7ea13ade.png) so the failing test shows up and the passing ones do not. But the original issue I had is that I'm not able to debug. If I click the debug button I get: ![Illegal Character](https://user-images.githubusercontent.com/7515022/166160833-e8e7c1b3-465b-4064-830f-e9dae653d72f.png) I tried upgrading the bazel plugin (and Goland) since I saw that fixing debugging with `go_test` was on the list of recent fixes, but the issue persists after all of that, plus doing `bazel clean` and even nuking my output-user-root directory to force everything to rebuild. Also note that the illegal character remains at index 116 even if I change the test to have everything commented out, leaving only one empty test method. Throughout all of this, `bazel build` and `bazel test` continue to work from the terminal as expected. Versions: Goland: 2022.1 Bazelisk: 1.11.0 Bazel: 5.1.1 Bazel Plugin: 2022.04.28.0.0-api-version-221
test
illegal character trying to debug go test and passing tests don t show up when run the problem occurs when run a test in goland from the run debug configurations using bazel test if i just run then the tests run and i can see the results on the console but the test navigator says no tests were found if they all pass note that i have show passing tests selected so the passing tests are expected to show up interestingly if i change one of the tests to fail i get so the failing test shows up and the passing ones do not but the original issue i had is that i m not able to debug if i click the debug button i get i tried upgrading the bazel plugin and goland since i saw that fixing debugging with go test was on the list of recent fixes but the issue persists after all of that plus doing bazel clean and even nuking my output user root directory to force everything to rebuild also note that the illegal character remains at index even if i change the test to have everything commented out leaving only one empty test method throughout all of this bazel build and bazel test continue to work from the terminal as expected versions goland bazelisk bazel bazel plugin api version
1
326,962
24,110,497,290
IssuesEvent
2022-09-20 10:56:05
CryptoBlades/cryptoblades
https://api.github.com/repos/CryptoBlades/cryptoblades
closed
Special (pickable) quests
documentation type-frontend type-solidity
Special quests are event-like offers we present to players when we want to create traffic for certain NFTs and resources. This extension involves a new pool of quests that can be freely picked to be done by any character regardless of reputation, as well as do quests that are tied only to wallets, without character involvement. The idea for wallet quests is to allow low barrier of entry options for users migrating from other projects (but possibly other use cases too). All wallet quests can be progressed simultaneously on a single wallet, it doesn't have to be one at a time. For quests hand picked for characters, it's the usual routine but no reputation requirement, and the user can choose the quest whenever the character is not on a quest already. This serves a similar purpose to wallet quests but we can tie it to character features (such as stamina, xp and rewarding reputation) Development is already underway (by me), character-pickable quests are done on the "pickableQuests" branch, wallet quests remain to be implemented still. Expected completion time (before testnet) is a few productive days away (but definitely before June 7th, ideally deploying to testnet by June 3rd to be tested on the weekend. This will need additional frontend work soon by someone with that skill set, to display the special quests on offer and hook in the buttons to pick them.
1.0
Special (pickable) quests - Special quests are event-like offers we present to players when we want to create traffic for certain NFTs and resources. This extension involves a new pool of quests that can be freely picked to be done by any character regardless of reputation, as well as do quests that are tied only to wallets, without character involvement. The idea for wallet quests is to allow low barrier of entry options for users migrating from other projects (but possibly other use cases too). All wallet quests can be progressed simultaneously on a single wallet, it doesn't have to be one at a time. For quests hand picked for characters, it's the usual routine but no reputation requirement, and the user can choose the quest whenever the character is not on a quest already. This serves a similar purpose to wallet quests but we can tie it to character features (such as stamina, xp and rewarding reputation) Development is already underway (by me), character-pickable quests are done on the "pickableQuests" branch, wallet quests remain to be implemented still. Expected completion time (before testnet) is a few productive days away (but definitely before June 7th, ideally deploying to testnet by June 3rd to be tested on the weekend. This will need additional frontend work soon by someone with that skill set, to display the special quests on offer and hook in the buttons to pick them.
non_test
special pickable quests special quests are event like offers we present to players when we want to create traffic for certain nfts and resources this extension involves a new pool of quests that can be freely picked to be done by any character regardless of reputation as well as do quests that are tied only to wallets without character involvement the idea for wallet quests is to allow low barrier of entry options for users migrating from other projects but possibly other use cases too all wallet quests can be progressed simultaneously on a single wallet it doesn t have to be one at a time for quests hand picked for characters it s the usual routine but no reputation requirement and the user can choose the quest whenever the character is not on a quest already this serves a similar purpose to wallet quests but we can tie it to character features such as stamina xp and rewarding reputation development is already underway by me character pickable quests are done on the pickablequests branch wallet quests remain to be implemented still expected completion time before testnet is a few productive days away but definitely before june ideally deploying to testnet by june to be tested on the weekend this will need additional frontend work soon by someone with that skill set to display the special quests on offer and hook in the buttons to pick them
0
14,033
3,372,449,058
IssuesEvent
2015-11-23 23:38:33
18F/doi-extractives-data
https://api.github.com/repos/18F/doi-extractives-data
closed
BUG: Header nav item not active for all subpages
type:bug workflow:testing
The header nav items are 'active' for direct subpages, but anything deeper (e.g. `/how-it-works/goverance`) does not activate the header nav item.
1.0
BUG: Header nav item not active for all subpages - The header nav items are 'active' for direct subpages, but anything deeper (e.g. `/how-it-works/goverance`) does not activate the header nav item.
test
bug header nav item not active for all subpages the header nav items are active for direct subpages but anything deeper e g how it works goverance does not activate the header nav item
1
53,321
6,310,091,615
IssuesEvent
2017-07-23 06:36:39
MajkiIT/polish-ads-filter
https://api.github.com/repos/MajkiIT/polish-ads-filter
closed
spidersweb.pl
reguły gotowe/testowanie reklama
szare boxy resztki http://www.spidersweb.pl/2017/07/komputer-do-cs-go.html http://www.spidersweb.pl/2015/06/apple-music-ceny-w-polsce.html ![screenshot from 2017-07-22 22-05-17](https://user-images.githubusercontent.com/15007183/28495858-07d4a4a2-6f2a-11e7-9236-8a0e75bcd26a.png) od czasu do czasu sa i na main page ![screenshot from 2017-07-22 22-06-14](https://user-images.githubusercontent.com/15007183/28495863-21d5f8ba-6f2a-11e7-84f4-12a9ff2b9d3f.png)
1.0
spidersweb.pl - szare boxy resztki http://www.spidersweb.pl/2017/07/komputer-do-cs-go.html http://www.spidersweb.pl/2015/06/apple-music-ceny-w-polsce.html ![screenshot from 2017-07-22 22-05-17](https://user-images.githubusercontent.com/15007183/28495858-07d4a4a2-6f2a-11e7-9236-8a0e75bcd26a.png) od czasu do czasu sa i na main page ![screenshot from 2017-07-22 22-06-14](https://user-images.githubusercontent.com/15007183/28495863-21d5f8ba-6f2a-11e7-84f4-12a9ff2b9d3f.png)
test
spidersweb pl szare boxy resztki od czasu do czasu sa i na main page
1
94,537
3,927,185,876
IssuesEvent
2016-04-23 11:47:07
MinetestForFun/server-minetestforfun-skyblock
https://api.github.com/repos/MinetestForFun/server-minetestforfun-skyblock
opened
Add PVP features
Priority: High
This days, many players ask us for a PVP system, we will include it but i think, in a way it will be acceptable for everyone, let me explain our roadmap : 1) Add swords, with linear increase damage with the ore you use to craft them. We will add them with modified textures from our "MFF_subgame", this default textures are 32px, well done and beautiful 2) Add the "PVP_button" mod Tweak/recode the mod to be compatible with "u_skins" inventory. Also we will add a button in your inventory to activate or deactivate PVP. We need to set a countdown for this activate/deactivate feature, well let's try 5 minutes for the beginning. 3) At last, activate damage on the server
1.0
Add PVP features - This days, many players ask us for a PVP system, we will include it but i think, in a way it will be acceptable for everyone, let me explain our roadmap : 1) Add swords, with linear increase damage with the ore you use to craft them. We will add them with modified textures from our "MFF_subgame", this default textures are 32px, well done and beautiful 2) Add the "PVP_button" mod Tweak/recode the mod to be compatible with "u_skins" inventory. Also we will add a button in your inventory to activate or deactivate PVP. We need to set a countdown for this activate/deactivate feature, well let's try 5 minutes for the beginning. 3) At last, activate damage on the server
non_test
add pvp features this days many players ask us for a pvp system we will include it but i think in a way it will be acceptable for everyone let me explain our roadmap add swords with linear increase damage with the ore you use to craft them we will add them with modified textures from our mff subgame this default textures are well done and beautiful add the pvp button mod tweak recode the mod to be compatible with u skins inventory also we will add a button in your inventory to activate or deactivate pvp we need to set a countdown for this activate deactivate feature well let s try minutes for the beginning at last activate damage on the server
0
328,967
28,143,202,572
IssuesEvent
2023-04-02 07:00:51
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix general.test_explicit_ivy_framework_handles
Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4558171043/jobs/8040729339" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4586003022/jobs/8098525241" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4575494898/jobs/8078405705" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4586062504/jobs/8098629047" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_general.py::test_explicit_ivy_framework_handles[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-04-02T00:50:01.0748115Z E AttributeError: module 'ivy' has no attribute 'get_backend' </details> <details> <summary>Not found</summary> Not found </details>
1.0
Fix general.test_explicit_ivy_framework_handles - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4558171043/jobs/8040729339" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4586003022/jobs/8098525241" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4575494898/jobs/8078405705" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4586062504/jobs/8098629047" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_general.py::test_explicit_ivy_framework_handles[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-04-02T00:50:01.0748115Z E AttributeError: module 'ivy' has no attribute 'get_backend' </details> <details> <summary>Not found</summary> Not found </details>
test
fix general test explicit ivy framework handles tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test functional test core test general py test explicit ivy framework handles e attributeerror module ivy has no attribute get backend not found not found
1
141,209
11,403,846,043
IssuesEvent
2020-01-31 08:23:36
ifb-git/ProteoRE
https://api.github.com/repos/ifb-git/ProteoRE
closed
New tool: Add_Protein_Features (Rattus) - uniprot
ToTest enhancement
Same specifications as "Add protein features (Mouse) [UniProt]" With Taxonomy ID: 10116 for Rattus norvegicus
1.0
New tool: Add_Protein_Features (Rattus) - uniprot - Same specifications as "Add protein features (Mouse) [UniProt]" With Taxonomy ID: 10116 for Rattus norvegicus
test
new tool add protein features rattus uniprot same specifications as add protein features mouse with taxonomy id for rattus norvegicus
1
283,531
8,719,830,902
IssuesEvent
2018-12-08 05:02:56
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
INSTALL_NOTES mentions installing from the dmg file and the tar.gz file, the tar.gz file is not on web page
bug likelihood medium priority reviewed severity low
The INSTALL_NOTES mentions that there are two ways to install on the Mac. One is with the dmg file and the other is with the tar.gz file. THe problem is that the tar.gz file is not on the download page (it is in the releases directory). We should either change the INSTALL_NOTES or add links for the tar.gz files. I think adding the tar.gz file to the download page is probably the way to go. I also noticed that the INSTALL_NOTES still mention 32 bit linux distributions, which we no longer provide. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1281 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: INSTALL_NOTES mentions installing from the dmg file and the tar.gz file, the tar.gz file is not on web page Assigned to: Eric Brugger Category: Target version: 2.6.1 Author: Eric Brugger Start: 12/18/2012 Due date: % Done: 100 Estimated time: 1.0 Created: 12/18/2012 12:10 pm Updated: 12/19/2012 05:15 pm Likelihood: 3 - Occasional Severity: 2 - Minor Irritation Found in version: 2.6.0 Impact: Expected Use: OS: All Support Group: Any Description: The INSTALL_NOTES mentions that there are two ways to install on the Mac. One is with the dmg file and the other is with the tar.gz file. THe problem is that the tar.gz file is not on the download page (it is in the releases directory). We should either change the INSTALL_NOTES or add links for the tar.gz files. I think adding the tar.gz file to the download page is probably the way to go. I also noticed that the INSTALL_NOTES still mention 32 bit linux distributions, which we no longer provide. Comments: After discussing this with Brad, Cyrus and Kathleen, we decided that we should remove it from the install notes rather than put the "tar.gz" file on the download page. I committed revisions 19872 and 19874 to the 2.6 RC and trunk with thefollowing change:1) I updated the INSTALL_NOTES so that only describes how to install the application bundle and no longer mentions installing it using the visit-install script. This resolves #1281.M INSTALL_NOTES
1.0
INSTALL_NOTES mentions installing from the dmg file and the tar.gz file, the tar.gz file is not on web page - The INSTALL_NOTES mentions that there are two ways to install on the Mac. One is with the dmg file and the other is with the tar.gz file. THe problem is that the tar.gz file is not on the download page (it is in the releases directory). We should either change the INSTALL_NOTES or add links for the tar.gz files. I think adding the tar.gz file to the download page is probably the way to go. I also noticed that the INSTALL_NOTES still mention 32 bit linux distributions, which we no longer provide. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1281 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: INSTALL_NOTES mentions installing from the dmg file and the tar.gz file, the tar.gz file is not on web page Assigned to: Eric Brugger Category: Target version: 2.6.1 Author: Eric Brugger Start: 12/18/2012 Due date: % Done: 100 Estimated time: 1.0 Created: 12/18/2012 12:10 pm Updated: 12/19/2012 05:15 pm Likelihood: 3 - Occasional Severity: 2 - Minor Irritation Found in version: 2.6.0 Impact: Expected Use: OS: All Support Group: Any Description: The INSTALL_NOTES mentions that there are two ways to install on the Mac. One is with the dmg file and the other is with the tar.gz file. THe problem is that the tar.gz file is not on the download page (it is in the releases directory). We should either change the INSTALL_NOTES or add links for the tar.gz files. I think adding the tar.gz file to the download page is probably the way to go. I also noticed that the INSTALL_NOTES still mention 32 bit linux distributions, which we no longer provide. Comments: After discussing this with Brad, Cyrus and Kathleen, we decided that we should remove it from the install notes rather than put the "tar.gz" file on the download page. I committed revisions 19872 and 19874 to the 2.6 RC and trunk with thefollowing change:1) I updated the INSTALL_NOTES so that only describes how to install the application bundle and no longer mentions installing it using the visit-install script. This resolves #1281.M INSTALL_NOTES
non_test
install notes mentions installing from the dmg file and the tar gz file the tar gz file is not on web page the install notes mentions that there are two ways to install on the mac one is with the dmg file and the other is with the tar gz file the problem is that the tar gz file is not on the download page it is in the releases directory we should either change the install notes or add links for the tar gz files i think adding the tar gz file to the download page is probably the way to go i also noticed that the install notes still mention bit linux distributions which we no longer provide redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject install notes mentions installing from the dmg file and the tar gz file the tar gz file is not on web page assigned to eric brugger category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity minor irritation found in version impact expected use os all support group any description the install notes mentions that there are two ways to install on the mac one is with the dmg file and the other is with the tar gz file the problem is that the tar gz file is not on the download page it is in the releases directory we should either change the install notes or add links for the tar gz files i think adding the tar gz file to the download page is probably the way to go i also noticed that the install notes still mention bit linux distributions which we no longer provide comments after discussing this with brad cyrus and kathleen we decided that we should remove it from the install notes rather than put the tar gz file on the download page i committed revisions and to the rc and trunk with thefollowing change i updated the install notes so that only describes how to install the application bundle and no longer mentions installing it using the visit install script this resolves m install notes
0
140,035
11,301,413,854
IssuesEvent
2020-01-17 15:32:49
stevenschader/kabanero-foundation
https://api.github.com/repos/stevenschader/kabanero-foundation
closed
SVT: TER: Kabanero Automation test execution master branch - Setup: svtcrc-91747-1
SVT Kabanero Test Execution Test Execution Record
Original logs: kabanero-svt01.fyre.ibm.com:/home/nest/kabanero-crc-logs/kabanero-auto-crc.sh.2019-10-24-05:45:33.setup.log PLAY [Create a one host Fyre stack] *********************************************************************************************** TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 05:45:37 -0700 (0:00:00.207) 0:00:00.207 ****** ok: [localhost] TASK [check status of the named cluster] ****************************************************************************************** Thursday 24 October 2019 05:45:37 -0700 (0:00:00.047) 0:00:00.255 ****** changed: [localhost] TASK [debug] ********************************************************************************************************************** Thursday 24 October 2019 05:45:38 -0700 (0:00:01.565) 0:00:01.821 ****** skipping: [localhost] TASK [get the public ssh id] ****************************************************************************************************** Thursday 24 October 2019 05:45:38 -0700 (0:00:00.057) 0:00:01.878 ****** changed: [localhost -> localhost] TASK [create input json file] ***************************************************************************************************** Thursday 24 October 2019 05:45:39 -0700 (0:00:00.306) 0:00:02.184 ****** changed: [localhost] TASK [Create Fyre stack] ********************************************************************************************************** Thursday 24 October 2019 05:45:40 -0700 (0:00:01.016) 0:00:03.201 ****** changed: [localhost] TASK [remove temp json] *********************************************************************************************************** Thursday 24 October 2019 05:45:41 -0700 (0:00:01.866) 0:00:05.068 ****** changed: [localhost -> localhost] TASK [check Fyre request status] ************************************************************************************************** Thursday 24 October 2019 05:45:42 -0700 (0:00:00.472) 0:00:05.540 ****** changed: [localhost] TASK [debug] ********************************************************************************************************************** Thursday 24 October 2019 05:45:43 -0700 (0:00:00.905) 0:00:06.445 ****** ok: [localhost] => { "fyreStatus.stdout": { "request": [ { "cluster_name": "", "created": "2019-10-24 05:45:41", "displayname": "steven r. schader", "error_details": "0", "id": "2420698", "status": "building", "type": "build", "updated": "2019-10-24 05:45:42", "username": "schader" } ] } } TASK [pause] ********************************************************************************************************************** Thursday 24 October 2019 05:45:43 -0700 (0:00:00.041) 0:00:06.487 ****** Pausing for 120 seconds (ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort) ok: [localhost] TASK [get Fyre request status] **************************************************************************************************** Thursday 24 October 2019 05:47:43 -0700 (0:02:00.047) 0:02:06.535 ****** changed: [localhost] TASK [check Fyre request status] ************************************************************************************************** Thursday 24 October 2019 05:47:44 -0700 (0:00:01.028) 0:02:07.564 ****** skipping: [localhost] TASK [check the host for a active ssh] ******************************************************************************************** Thursday 24 October 2019 05:47:44 -0700 (0:00:00.041) 0:02:07.605 ****** changed: [localhost -> localhost] TASK [remove new host from localhost known_hosts] ********************************************************************************* Thursday 24 October 2019 05:47:45 -0700 (0:00:00.582) 0:02:08.188 ****** changed: [localhost -> localhost] TASK [add host to known_hosts] **************************************************************************************************** Thursday 24 October 2019 05:47:45 -0700 (0:00:00.315) 0:02:08.503 ****** changed: [localhost -> localhost] TASK [create host inventory file] ************************************************************************************************* Thursday 24 October 2019 05:47:45 -0700 (0:00:00.378) 0:02:08.881 ****** changed: [localhost -> localhost] TASK [create plain file with host] ************************************************************************************************ Thursday 24 October 2019 05:47:46 -0700 (0:00:00.595) 0:02:09.476 ****** changed: [localhost -> localhost] TASK [create plain file with cluster name] **************************************************************************************** Thursday 24 October 2019 05:47:46 -0700 (0:00:00.563) 0:02:10.040 ****** changed: [localhost -> localhost] TASK [remove linked hosts.crc] **************************************************************************************************** Thursday 24 October 2019 05:47:47 -0700 (0:00:00.574) 0:02:10.615 ****** changed: [localhost -> localhost] TASK [link new host inventory file to hosts.crc] ********************************************************************************** Thursday 24 October 2019 05:47:47 -0700 (0:00:00.285) 0:02:10.901 ****** changed: [localhost -> localhost] PLAY RECAP ************************************************************************************************************************ localhost : ok=18 changed=15 unreachable=0 failed=0 skipped=2 rescued=0 ignored=0 Thursday 24 October 2019 05:47:48 -0700 (0:00:00.288) 0:02:11.189 ****** =============================================================================== pause -------------------------------------------------------------------------------------------------------------------- 120.05s Create Fyre stack ---------------------------------------------------------------------------------------------------------- 1.87s check status of the named cluster ------------------------------------------------------------------------------------------ 1.57s get Fyre request status ---------------------------------------------------------------------------------------------------- 1.03s create input json file ----------------------------------------------------------------------------------------------------- 1.02s check Fyre request status -------------------------------------------------------------------------------------------------- 0.91s create host inventory file ------------------------------------------------------------------------------------------------- 0.60s check the host for a active ssh -------------------------------------------------------------------------------------------- 0.58s create plain file with cluster name ---------------------------------------------------------------------------------------- 0.57s create plain file with host ------------------------------------------------------------------------------------------------ 0.56s remove temp json ----------------------------------------------------------------------------------------------------------- 0.47s add host to known_hosts ---------------------------------------------------------------------------------------------------- 0.38s remove new host from localhost known_hosts --------------------------------------------------------------------------------- 0.32s get the public ssh id ------------------------------------------------------------------------------------------------------ 0.31s link new host inventory file to hosts.crc ---------------------------------------------------------------------------------- 0.29s remove linked hosts.crc ---------------------------------------------------------------------------------------------------- 0.29s debug ---------------------------------------------------------------------------------------------------------------------- 0.06s set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s debug ---------------------------------------------------------------------------------------------------------------------- 0.04s check Fyre request status -------------------------------------------------------------------------------------------------- 0.04s fyrecrcvm complete. runtime 2 minutes PLAY [check RHEL 7 or greater x64 only] ******************************************************************************************* TASK [Gathering Facts] ************************************************************************************************************ Thursday 24 October 2019 05:47:50 -0700 (0:00:00.170) 0:00:00.170 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [Target host is RHEL7 ( or greater ) x86_64] ********************************************************************************* Thursday 24 October 2019 05:47:52 -0700 (0:00:01.803) 0:00:01.974 ****** ok: [svtcrc-91747-1.fyre.ibm.com] PLAY RECAP ************************************************************************************************************************ svtcrc-91747-1.fyre.ibm.com : ok=2 changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 Thursday 24 October 2019 05:47:52 -0700 (0:00:00.039) 0:00:02.014 ****** =============================================================================== Gathering Facts ------------------------------------------------------------------------------------------------------------ 1.80s Target host is RHEL7 ( or greater ) x86_64 --------------------------------------------------------------------------------- 0.04s PLAY [Create a User] ************************************************************************************************************** TASK [Generate User Password file] ************************************************************************************************ Thursday 24 October 2019 05:47:55 -0700 (0:00:00.183) 0:00:00.183 ****** ok: [svtcrc-91747-1.fyre.ibm.com -> localhost] TASK [Ensure User is Present] ***************************************************************************************************** Thursday 24 October 2019 05:47:56 -0700 (0:00:00.495) 0:00:00.678 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [authorized_key] ************************************************************************************************************* Thursday 24 October 2019 05:47:57 -0700 (0:00:01.163) 0:00:01.842 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [kevin sudoers] ************************************************************************************************************** Thursday 24 October 2019 05:47:58 -0700 (0:00:00.727) 0:00:02.569 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [kevin hard ulimit nofile] *************************************************************************************************** Thursday 24 October 2019 05:47:58 -0700 (0:00:00.750) 0:00:03.320 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [kevin soft ulimit nofile] *************************************************************************************************** Thursday 24 October 2019 05:47:59 -0700 (0:00:00.434) 0:00:03.754 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [add bin to path] ************************************************************************************************************ Thursday 24 October 2019 05:47:59 -0700 (0:00:00.437) 0:00:04.192 ****** changed: [svtcrc-91747-1.fyre.ibm.com] PLAY RECAP ************************************************************************************************************************ svtcrc-91747-1.fyre.ibm.com : ok=7 changed=6 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 Thursday 24 October 2019 05:48:00 -0700 (0:00:00.428) 0:00:04.620 ****** =============================================================================== Ensure User is Present ----------------------------------------------------------------------------------------------------- 1.16s kevin sudoers -------------------------------------------------------------------------------------------------------------- 0.75s authorized_key ------------------------------------------------------------------------------------------------------------- 0.73s Generate User Password file ------------------------------------------------------------------------------------------------ 0.50s kevin soft ulimit nofile --------------------------------------------------------------------------------------------------- 0.44s kevin hard ulimit nofile --------------------------------------------------------------------------------------------------- 0.43s add bin to path ------------------------------------------------------------------------------------------------------------ 0.43s PLAY [switch hosts.crc ansible_user] ********************************************************************************************** TASK [remove kevin ansible_user] ************************************************************************************************** Thursday 24 October 2019 05:48:02 -0700 (0:00:00.178) 0:00:00.178 ****** ok: [svtcrc-91747-1.fyre.ibm.com -> localhost] TASK [set kevin ansible_user] ***************************************************************************************************** Thursday 24 October 2019 05:48:03 -0700 (0:00:00.476) 0:00:00.654 ****** changed: [svtcrc-91747-1.fyre.ibm.com -> localhost] PLAY RECAP ************************************************************************************************************************ svtcrc-91747-1.fyre.ibm.com : ok=2 changed=1 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 Thursday 24 October 2019 05:48:03 -0700 (0:00:00.266) 0:00:00.921 ****** =============================================================================== remove kevin ansible_user -------------------------------------------------------------------------------------------------- 0.48s set kevin ansible_user ----------------------------------------------------------------------------------------------------- 0.27s PLAY [SVT - install OCP CodeReadyContainer binary] ******************************************************************************** TASK [Gathering Facts] ************************************************************************************************************ Thursday 24 October 2019 05:48:06 -0700 (0:00:00.171) 0:00:00.171 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [check if crc is installed] ************************************************************************************************** Thursday 24 October 2019 05:48:08 -0700 (0:00:01.862) 0:00:02.034 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [archive dir structure] ****************************************************************************************************** Thursday 24 October 2019 05:48:09 -0700 (0:00:00.653) 0:00:02.688 ****** changed: [svtcrc-91747-1.fyre.ibm.com -> localhost] TASK [debug] ********************************************************************************************************************** Thursday 24 October 2019 05:48:17 -0700 (0:00:08.477) 0:00:11.165 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "tarlist.stdout_lines": [ "crc-linux-1.0.0-amd64/", "crc-linux-1.0.0-amd64/LICENSE", "crc-linux-1.0.0-amd64/doc.pdf", "crc-linux-1.0.0-amd64/crc" ] } TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 05:48:17 -0700 (0:00:00.055) 0:00:11.221 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [copy crc archive] *********************************************************************************************************** Thursday 24 October 2019 05:48:17 -0700 (0:00:00.062) 0:00:11.284 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [install crc] **************************************************************************************************************** Thursday 24 October 2019 05:50:03 -0700 (0:01:46.017) 0:01:57.301 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [home bin] ******************************************************************************************************************* Thursday 24 October 2019 05:50:25 -0700 (0:00:22.091) 0:02:19.392 ****** changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/bin) TASK [link crc] ******************************************************************************************************************* Thursday 24 October 2019 05:50:26 -0700 (0:00:00.761) 0:02:20.154 ****** changed: [svtcrc-91747-1.fyre.ibm.com] PLAY RECAP ************************************************************************************************************************ svtcrc-91747-1.fyre.ibm.com : ok=9 changed=5 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 Thursday 24 October 2019 05:50:27 -0700 (0:00:00.483) 0:02:20.637 ****** =============================================================================== copy crc archive --------------------------------------------------------------------------------------------------------- 106.02s install crc --------------------------------------------------------------------------------------------------------------- 22.09s archive dir structure ------------------------------------------------------------------------------------------------------ 8.48s Gathering Facts ------------------------------------------------------------------------------------------------------------ 1.86s home bin ------------------------------------------------------------------------------------------------------------------- 0.76s check if crc is installed -------------------------------------------------------------------------------------------------- 0.65s link crc ------------------------------------------------------------------------------------------------------------------- 0.48s set_fact ------------------------------------------------------------------------------------------------------------------- 0.06s debug ---------------------------------------------------------------------------------------------------------------------- 0.06s crc-install complete. runtime 2 minutes PLAY [SVT - Kabanero on crc] ****************************************************************************************************** TASK [Gathering Facts] ************************************************************************************************************ Thursday 24 October 2019 05:50:31 -0700 (0:00:00.175) 0:00:00.175 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 05:50:33 -0700 (0:00:01.727) 0:00:01.902 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [Install NetworkManager FireFox] ********************************************************************************************* Thursday 24 October 2019 05:50:33 -0700 (0:00:00.060) 0:00:01.963 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [enable and start NetworkManager] ******************************************************************************************** Thursday 24 October 2019 05:51:47 -0700 (0:01:13.989) 0:01:15.952 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [check if crc is installed] ************************************************************************************************** Thursday 24 October 2019 05:51:48 -0700 (0:00:01.324) 0:01:17.276 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [fail] *********************************************************************************************************************** Thursday 24 October 2019 05:51:49 -0700 (0:00:00.713) 0:01:17.990 ****** skipping: [svtcrc-91747-1.fyre.ibm.com] TASK [prereq directories] ********************************************************************************************************* Thursday 24 October 2019 05:51:49 -0700 (0:00:00.059) 0:01:18.049 ****** changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/git) changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/testDir) changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/kabanero/logs) TASK [retrieve Kabanero Foundation master zip] ************************************************************************************ Thursday 24 October 2019 05:51:51 -0700 (0:00:01.507) 0:01:19.557 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [unzip Kabanerio Foundation master zip] ************************************************************************************** Thursday 24 October 2019 05:51:53 -0700 (0:00:02.402) 0:01:21.959 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [whoami] ********************************************************************************************************************* Thursday 24 October 2019 05:51:54 -0700 (0:00:01.073) 0:01:23.032 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [whoami stdout] ************************************************************************************************************** Thursday 24 October 2019 05:51:55 -0700 (0:00:00.689) 0:01:23.722 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "kevin" ] } TASK [check if .crc is created] *************************************************************************************************** Thursday 24 October 2019 05:51:55 -0700 (0:00:00.047) 0:01:23.770 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [crc stop] ******************************************************************************************************************* Thursday 24 October 2019 05:51:55 -0700 (0:00:00.481) 0:01:24.251 ****** skipping: [svtcrc-91747-1.fyre.ibm.com] TASK [crcstop stdout] ************************************************************************************************************* Thursday 24 October 2019 05:51:55 -0700 (0:00:00.053) 0:01:24.305 ****** skipping: [svtcrc-91747-1.fyre.ibm.com] TASK [crc delete] ***************************************************************************************************************** Thursday 24 October 2019 05:51:55 -0700 (0:00:00.058) 0:01:24.364 ****** skipping: [svtcrc-91747-1.fyre.ibm.com] TASK [crc delete stdout] ********************************************************************************************************** Thursday 24 October 2019 05:51:55 -0700 (0:00:00.059) 0:01:24.423 ****** skipping: [svtcrc-91747-1.fyre.ibm.com] TASK [crc setup] ****************************************************************************************************************** Thursday 24 October 2019 05:51:55 -0700 (0:00:00.050) 0:01:24.474 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [crc setup stdout] *********************************************************************************************************** Thursday 24 October 2019 05:52:34 -0700 (0:00:38.733) 0:02:03.207 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "level=info msg=\"Checking if running as non-root\"", "level=info msg=\"Caching oc binary\"", "level=info msg=\"Setting up virtualization\"", "level=info msg=\"Setting up KVM\"", "level=info msg=\"Installing libvirt service and dependencies\"", "level=info msg=\"Will use root access: install virtualization related packages\"", "level=info msg=\"Adding user to libvirt group\"", "level=info msg=\"Will use root access: add user to libvirt group\"", "level=info msg=\"Enabling libvirt\"", "level=info msg=\"Starting libvirt service\"", "level=info msg=\"Will use root access: start libvirtd service\"", "level=info msg=\"Checking if a supported libvirt version is installed\"", "level=info msg=\"Installing crc-driver-libvirt\"", "level=info msg=\"Removing older system-wide crc-driver-libvirt\"", "level=info msg=\"Setting up libvirt 'crc' network\"", "level=info msg=\"Starting libvirt 'crc' network\"", "level=info msg=\"Checking if NetworkManager is installed\"", "level=info msg=\"Checking if NetworkManager service is running\"", "level=info msg=\"Writing Network Manager config for crc\"", "level=info msg=\"Will use root access: write NetworkManager config in /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf\"", "level=info msg=\"Will use root access: execute systemctl daemon-reload command\"", "level=info msg=\"Will use root access: execute systemctl stop/start command\"", "level=info msg=\"Writing dnsmasq config for crc\"", "level=info msg=\"Will use root access: write dnsmasq configuration in /etc/NetworkManager/dnsmasq.d/crc.conf\"", "level=info msg=\"Will use root access: execute systemctl daemon-reload command\"", "level=info msg=\"Will use root access: execute systemctl stop/start command\"", "level=info msg=\"Unpacking bundle from the CRC binary\"", "Setup is complete, you can now run 'crc start' to start the OpenShift cluster" ] } TASK [crc pull secret] ************************************************************************************************************ Thursday 24 October 2019 05:52:34 -0700 (0:00:00.074) 0:02:03.281 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [crc start] ****************************************************************************************************************** Thursday 24 October 2019 05:52:35 -0700 (0:00:01.108) 0:02:04.390 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [crc start stdout] *********************************************************************************************************** Thursday 24 October 2019 06:03:32 -0700 (0:10:56.959) 0:13:01.349 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "level=info msg=\"Checking if running as non-root\"", "level=info msg=\"Checking if oc binary is cached\"", "level=info msg=\"Checking if Virtualization is enabled\"", "level=info msg=\"Checking if KVM is enabled\"", "level=info msg=\"Checking if libvirt is installed\"", "level=info msg=\"Checking if user is part of libvirt group\"", "level=info msg=\"Checking if libvirt is enabled\"", "level=info msg=\"Checking if libvirt daemon is running\"", "level=info msg=\"Checking if a supported libvirt version is installed\"", "level=info msg=\"Checking if crc-driver-libvirt is installed\"", "level=info msg=\"Checking if libvirt 'crc' network is available\"", "level=info msg=\"Checking if libvirt 'crc' network is active\"", "level=info msg=\"Checking if NetworkManager is installed\"", "level=info msg=\"Checking if NetworkManager service is running\"", "level=info msg=\"Checking if /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf exists\"", "level=info msg=\"Checking if /etc/NetworkManager/dnsmasq.d/crc.conf exists\"", "level=info msg=\"Extracting bundle: crc_libvirt_4.2.0.crcbundle ...\"", "level=info msg=\"Creating CodeReady Containers VM for OpenShift 4.2.0...\"", "level=info msg=\"Verifying validity of the cluster certificates ...\"", "level=info msg=\"Check internal and public DNS query ...\"", "level=info msg=\"Copying kubeconfig file to instance dir ...\"", "level=info msg=\"Adding user's pull secret and cluster ID ...\"", "level=info msg=\"Starting OpenShift cluster ... [waiting 3m]\"", "level=info", "level=info msg=\"To access the cluster, first set up your environment by following 'crc oc-env' instructions\"", "level=info msg=\"Then you can access it by running 'oc login -u developer -p developer https://api.crc.testing:6443'\"", "level=info msg=\"To login as an admin, username is 'kubeadmin' and password is xx-password-xx\"", "level=info", "level=info msg=\"You can now run 'crc console' and use these credentials to access the OpenShift web console\"", "The OpenShift cluster is running", "level=warning msg=\"The cluster might report a degraded or error state. This is expected since several operators have been disable to lower the resource usage. For more information, please consult the documentation\"" ] } TASK [oc in user path] ************************************************************************************************************ Thursday 24 October 2019 06:03:32 -0700 (0:00:00.066) 0:13:01.416 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [link oc kubectl] ************************************************************************************************************ Thursday 24 October 2019 06:03:33 -0700 (0:00:00.663) 0:13:02.079 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [crc creds] ****************************************************************************************************************** Thursday 24 October 2019 06:03:34 -0700 (0:00:00.438) 0:13:02.518 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [crc creds stdout] *********************************************************************************************************** Thursday 24 October 2019 06:03:34 -0700 (0:00:00.498) 0:13:03.016 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "To login as a regular user, username is 'developer' and password is 'developer'.", "To login as an admin, username is 'kubeadmin' and password is 'xx-password-xx'." ] } TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 06:03:34 -0700 (0:00:00.064) 0:13:03.081 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 06:03:34 -0700 (0:00:00.052) 0:13:03.134 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 06:03:34 -0700 (0:00:00.051) 0:13:03.185 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 06:03:34 -0700 (0:00:00.051) 0:13:03.236 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [password masking process] *************************************************************************************************** Thursday 24 October 2019 06:03:34 -0700 (0:00:00.076) 0:13:03.312 ****** changed: [svtcrc-91747-1.fyre.ibm.com -> localhost] TASK [set admin passwd] *********************************************************************************************************** Thursday 24 October 2019 06:03:35 -0700 (0:00:00.638) 0:13:03.951 ****** changed: [svtcrc-91747-1.fyre.ibm.com -> localhost] TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 06:03:35 -0700 (0:00:00.356) 0:13:04.307 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [oc login] ******************************************************************************************************************* Thursday 24 October 2019 06:03:35 -0700 (0:00:00.064) 0:13:04.371 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [oc login stdout] ************************************************************************************************************ Thursday 24 October 2019 06:03:37 -0700 (0:00:01.414) 0:13:05.786 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "Login successful.", "", "You have access to 51 projects, the list has been suppressed. You can list all projects with 'oc projects'", "", "Using project \"default\".", "Welcome! See 'oc help' to get started." ] } TASK [install kabanero-foundation] ************************************************************************************************ Thursday 24 October 2019 06:03:37 -0700 (0:00:00.062) 0:13:05.849 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [install kabanero-foundation stdout] ***************************************************************************************** Thursday 24 October 2019 06:14:36 -0700 (0:10:59.331) 0:24:05.180 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-ingress-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:default\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:prometheus\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-egressgateway-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-citadel-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-ingressgateway-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-cleanup-old-ca-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-mixer-post-install-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-mixer-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-pilot-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-sidecar-injector-service-account\"]", "clusterrole.rbac.authorization.k8s.io/cluster-admin added: \"istio-galley-service-account\"", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:cluster-local-gateway-service-account\"]", "customresourcedefinition.apiextensions.k8s.io/virtualservices.networking.istio.io created", "customresourcedefinition.apiextensions.k8s.io/destinationrules.networking.istio.io created", "customresourcedefinition.apiextensions.k8s.io/serviceentries.networking.istio.io created", "customresourcedefinition.apiextensions.k8s.io/gateways.networking.istio.io created", "customresourcedefinition.apiextensions.k8s.io/envoyfilters.networking.istio.io created", "customresourcedefinition.apiextensions.k8s.io/clusterrbacconfigs.rbac.istio.io created", "customresourcedefinition.apiextensions.k8s.io/policies.authentication.istio.io created", "customresourcedefinition.apiextensions.k8s.io/meshpolicies.authentication.istio.io created", "customresourcedefinition.apiextensions.k8s.io/httpapispecbindings.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/httpapispecs.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/quotaspecbindings.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/quotaspecs.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/rules.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/attributemanifests.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/bypasses.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/circonuses.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/deniers.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/fluentds.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/kubernetesenvs.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/listcheckers.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/memquotas.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/noops.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/opas.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/prometheuses.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/rbacs.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/redisquotas.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/signalfxs.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/solarwindses.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/stackdrivers.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/statsds.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/stdios.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/apikeys.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/authorizations.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/checknothings.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/kuberneteses.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/listentries.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/logentries.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/edges.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/metrics.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/quotas.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/reportnothings.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/tracespans.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/rbacconfigs.rbac.istio.io created", "customresourcedefinition.apiextensions.k8s.io/serviceroles.rbac.istio.io created", "customresourcedefinition.apiextensions.k8s.io/servicerolebindings.rbac.istio.io created", "customresourcedefinition.apiextensions.k8s.io/adapters.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/instances.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/templates.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/handlers.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/cloudwatches.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/dogstatsds.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/sidecars.networking.istio.io created", "customresourcedefinition.apiextensions.k8s.io/zipkins.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/clusterissuers.certmanager.k8s.io created", "customresourcedefinition.apiextensions.k8s.io/issuers.certmanager.k8s.io created", "customresourcedefinition.apiextensions.k8s.io/certificates.certmanager.k8s.io created", "customresourcedefinition.apiextensions.k8s.io/orders.certmanager.k8s.io created", "customresourcedefinition.apiextensions.k8s.io/challenges.certmanager.k8s.io created", "namespace/istio-system created", "customresourcedefinition.apiextensions.k8s.io/virtualservices.networking.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/destinationrules.networking.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/serviceentries.networking.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/gateways.networking.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/envoyfilters.networking.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/clusterrbacconfigs.rbac.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/policies.authentication.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/meshpolicies.authentication.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/httpapispecbindings.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/httpapispecs.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/quotaspecbindings.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/quotaspecs.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/rules.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/attributemanifests.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/bypasses.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/circonuses.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/deniers.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/fluentds.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/kubernetesenvs.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/listcheckers.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/memquotas.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/noops.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/opas.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/prometheuses.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/rbacs.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/redisquotas.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/signalfxs.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/solarwindses.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/stackdrivers.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/statsds.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/stdios.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/apikeys.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/authorizations.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/checknothings.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/kuberneteses.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/listentries.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/logentries.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/edges.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/metrics.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/quotas.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/reportnothings.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/tracespans.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/rbacconfigs.rbac.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/serviceroles.rbac.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/servicerolebindings.rbac.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/adapters.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/instances.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/templates.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/handlers.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/cloudwatches.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/dogstatsds.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/sidecars.networking.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/zipkins.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/clusterissuers.certmanager.k8s.io unchanged", "customresourcedefinition.apiextensions.k8s.io/issuers.certmanager.k8s.io unchanged", "customresourcedefinition.apiextensions.k8s.io/orders.certmanager.k8s.io unchanged", "customresourcedefinition.apiextensions.k8s.io/challenges.certmanager.k8s.io unchanged", "secret/kiali created", "configmap/istio-galley-configuration created", "configmap/istio-grafana-custom-resources created", "configmap/istio-grafana-configuration-dashboards-galley-dashboard created", "configmap/istio-grafana-configuration-dashboards-istio-mesh-dashboard created", "configmap/istio-grafana-configuration-dashboards-istio-performance-dashboard created", "configmap/istio-grafana-configuration-dashboards-istio-service-dashboard created", "configmap/istio-grafana-configuration-dashboards-istio-workload-dashboard created", "configmap/istio-grafana-configuration-dashboards-mixer-dashboard created", "configmap/istio-grafana-configuration-dashboards-pilot-dashboard created", "configmap/istio-grafana created", "configmap/kiali created", "configmap/prometheus created", "configmap/istio-security-custom-resources created", "configmap/istio created", "configmap/istio-sidecar-injector created", "serviceaccount/istio-galley-service-account created", "serviceaccount/istio-egressgateway-service-account created", "serviceaccount/istio-ingressgateway-service-account created", "serviceaccount/istio-grafana-post-install-account created", "clusterrole.rbac.authorization.k8s.io/istio-grafana-post-install-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-grafana-post-install-role-binding-istio-system created", "job.batch/istio-grafana-post-install-1.1.7 created", "serviceaccount/kiali-service-account created", "serviceaccount/istio-mixer-service-account created", "serviceaccount/istio-pilot-service-account created", "serviceaccount/prometheus created", "serviceaccount/istio-cleanup-secrets-service-account created", "clusterrole.rbac.authorization.k8s.io/istio-cleanup-secrets-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-cleanup-secrets-istio-system created", "job.batch/istio-cleanup-secrets-1.1.7 created", "serviceaccount/istio-security-post-install-account created", "clusterrole.rbac.authorization.k8s.io/istio-security-post-install-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-security-post-install-role-binding-istio-system created", "job.batch/istio-security-post-install-1.1.7 created", "serviceaccount/istio-citadel-service-account created", "serviceaccount/istio-sidecar-injector-service-account created", "serviceaccount/istio-multi created", "clusterrole.rbac.authorization.k8s.io/istio-galley-istio-system created", "clusterrole.rbac.authorization.k8s.io/istio-egressgateway-istio-system created", "clusterrole.rbac.authorization.k8s.io/istio-ingressgateway-istio-system created", "clusterrole.rbac.authorization.k8s.io/kiali created", "clusterrole.rbac.authorization.k8s.io/kiali-viewer created", "clusterrole.rbac.authorization.k8s.io/istio-mixer-istio-system created", "clusterrole.rbac.authorization.k8s.io/istio-pilot-istio-system created", "clusterrole.rbac.authorization.k8s.io/prometheus-istio-system created", "clusterrole.rbac.authorization.k8s.io/istio-citadel-istio-system created", "clusterrole.rbac.authorization.k8s.io/istio-sidecar-injector-istio-system created", "clusterrole.rbac.authorization.k8s.io/istio-reader created", "clusterrolebinding.rbac.authorization.k8s.io/istio-galley-admin-role-binding-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-egressgateway-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-ingressgateway-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-kiali-admin-role-binding-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-mixer-admin-role-binding-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-pilot-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/prometheus-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-citadel-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-sidecar-injector-admin-role-binding-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-multi created", "role.rbac.authorization.k8s.io/istio-ingressgateway-sds created", "rolebinding.rbac.authorization.k8s.io/istio-ingressgateway-sds created", "service/istio-galley created", "service/istio-egressgateway created", "service/istio-ingressgateway created", "service/grafana created", "service/kiali created", "service/istio-policy created", "service/istio-telemetry created", "service/istio-pilot created", "service/prometheus created", "service/istio-citadel created", "service/istio-sidecar-injector created", "deployment.extensions/istio-galley created", "deployment.extensions/istio-egressgateway created", "deployment.extensions/istio-ingressgateway created", "deployment.extensions/grafana created", "deployment.extensions/kiali created", "deployment.extensions/istio-policy created", "deployment.extensions/istio-telemetry created", "deployment.extensions/istio-pilot created", "deployment.extensions/prometheus created", "deployment.extensions/istio-citadel created", "deployment.extensions/istio-sidecar-injector created", "deployment.extensions/istio-tracing created", "horizontalpodautoscaler.autoscaling/istio-egressgateway created", "horizontalpodautoscaler.autoscaling/istio-ingressgateway created", "horizontalpodautoscaler.autoscaling/istio-policy created", "horizontalpodautoscaler.autoscaling/istio-telemetry created", "horizontalpodautoscaler.autoscaling/istio-pilot created", "service/jaeger-query created", "service/jaeger-collector created", "service/jaeger-agent created", "service/zipkin created", "service/tracing created", "mutatingwebhookconfiguration.admissionregistration.k8s.io/istio-sidecar-injector created", "poddisruptionbudget.policy/istio-galley created", "poddisruptionbudget.policy/istio-egressgateway created", "poddisruptionbudget.policy/istio-ingressgateway created", "poddisruptionbudget.policy/istio-policy created", "poddisruptionbudget.policy/istio-telemetry created", "poddisruptionbudget.policy/istio-pilot created", "attributemanifest.config.istio.io/istioproxy created", "attributemanifest.config.istio.io/kubernetes created", "handler.config.istio.io/stdio created", "logentry.config.istio.io/accesslog created", "logentry.config.istio.io/tcpaccesslog created", "rule.config.istio.io/stdio created", "rule.config.istio.io/stdiotcp created", "metric.config.istio.io/requestcount created", "metric.config.istio.io/requestduration created", "metric.config.istio.io/requestsize created", "metric.config.istio.io/responsesize created", "metric.config.istio.io/tcpbytesent created", "metric.config.istio.io/tcpbytereceived created", "metric.config.istio.io/tcpconnectionsopened created", "metric.config.istio.io/tcpconnectionsclosed created", "handler.config.istio.io/prometheus created", "rule.config.istio.io/promhttp created", "rule.config.istio.io/promtcp created", "rule.config.istio.io/promtcpconnectionopen created", "rule.config.istio.io/promtcpconnectionclosed created", "handler.config.istio.io/kubernetesenv created", "rule.config.istio.io/kubeattrgenrulerule created", "rule.config.istio.io/tcpkubeattrgenrulerule created", "kubernetes.config.istio.io/attributes created", "destinationrule.networking.istio.io/istio-policy created", "destinationrule.networking.istio.io/istio-telemetry created", "Now using project \"kabanero\" on server \"https://api.crc.testing:6443\".", "", "You can add applications to this project with the 'new-app' command. For example, try:", "", " oc new-app django-psql-example", "", "to build a new example application in Python. Or use kubectl to deploy a simple Kubernetes application:", "", " kubectl create deployment hello-node --image=gcr.io/hello-minikube-zero-install/hello-node", "", "namespace/kabanero configured", "deployment.apps/knative-eventing-operator created", "role.rbac.authorization.k8s.io/knative-eventing-operator created", "clusterrole.rbac.authorization.k8s.io/knative-eventing-operator created", "rolebinding.rbac.authorization.k8s.io/knative-eventing-operator created", "clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-operator created", "serviceaccount/knative-eventing-operator created", "customresourcedefinition.apiextensions.k8s.io/knativeeventings.eventing.knative.dev created", "deployment.apps/knative-serving-operator created", "role.rbac.authorization.k8s.io/knative-serving-operator created", "clusterrole.rbac.authorization.k8s.io/knative-serving-operator created", "rolebinding.rbac.authorization.k8s.io/knative-serving-operator created", "clusterrolebinding.rbac.authorization.k8s.io/knative-serving-operator created", "serviceaccount/knative-serving-operator created", "customresourcedefinition.apiextensions.k8s.io/knativeservings.serving.knative.dev created", "namespace/knative-sources created", "serviceaccount/controller-manager created", "clusterrole.rbac.authorization.k8s.io/eventing-sources-controller created", "clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-controller created", "clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-addressable-resolver created", "customresourcedefinition.apiextensions.k8s.io/githubsources.sources.eventing.knative.dev created", "service/controller created", "statefulset.apps/controller-manager created", "deployment.apps/openshift-pipelines-operator created", "clusterrole.rbac.authorization.k8s.io/openshift-pipelines-operator created", "clusterrolebinding.rbac.authorization.k8s.io/openshift-pipelines-operator created", "serviceaccount/openshift-pipelines-operator created", "namespace/openshift-pipelines created", "customresourcedefinition.apiextensions.k8s.io/config.operator.tekton.dev created", "clusterrole.rbac.authorization.k8s.io/knative-serving-operator configured", "clusterrole.rbac.authorization.k8s.io/knative-eventing-operator configured", "customresourcedefinition.apiextensions.k8s.io/kabaneros.kabanero.io created", "customresourcedefinition.apiextensions.k8s.io/collections.kabanero.io created", "serviceaccount/kabanero-operator created", "deployment.apps/kabanero-operator created", "clusterrole.rbac.authorization.k8s.io/kabanero-operator created", "clusterrolebinding.rbac.authorization.k8s.io/kabanero-operator created", "namespace/kabanero unchanged", "deployment.apps/knative-eventing-operator configured", "role.rbac.authorization.k8s.io/knative-eventing-operator configured", "clusterrole.rbac.authorization.k8s.io/knative-eventing-operator configured", "rolebinding.rbac.authorization.k8s.io/knative-eventing-operator unchanged", "clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-operator unchanged", "serviceaccount/knative-eventing-operator unchanged", "customresourcedefinition.apiextensions.k8s.io/knativeeventings.eventing.knative.dev unchanged", "deployment.apps/knative-serving-operator configured", "role.rbac.authorization.k8s.io/knative-serving-operator configured", "clusterrole.rbac.authorization.k8s.io/knative-serving-operator configured", "rolebinding.rbac.authorization.k8s.io/knative-serving-operator unchanged", "clusterrolebinding.rbac.authorization.k8s.io/knative-serving-operator unchanged", "serviceaccount/knative-serving-operator unchanged", "customresourcedefinition.apiextensions.k8s.io/knativeservings.serving.knative.dev unchanged", "namespace/knative-sources unchanged", "serviceaccount/controller-manager unchanged", "clusterrole.rbac.authorization.k8s.io/eventing-sources-controller unchanged", "clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-controller unchanged", "clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-addressable-resolver unchanged", "customresourcedefinition.apiextensions.k8s.io/githubsources.sources.eventing.knative.dev unchanged", "service/controller unchanged", "statefulset.apps/controller-manager configured", "deployment.apps/openshift-pipelines-operator configured", "clusterrole.rbac.authorization.k8s.io/openshift-pipelines-operator configured", "clusterrolebinding.rbac.authorization.k8s.io/openshift-pipelines-operator unchanged", "serviceaccount/openshift-pipelines-operator unchanged", "namespace/openshift-pipelines unchanged", "customresourcedefinition.apiextensions.k8s.io/config.operator.tekton.dev unchanged", "config.operator.tekton.dev/cluster created", "clusterrole.rbac.authorization.k8s.io/knative-serving-operator configured", "clusterrole.rbac.authorization.k8s.io/knative-eventing-operator configured", "customresourcedefinition.apiextensions.k8s.io/kabaneros.kabanero.io configured", "customresourcedefinition.apiextensions.k8s.io/collections.kabanero.io configured", "serviceaccount/kabanero-operator unchanged", "deployment.apps/kabanero-operator configured", "clusterrole.rbac.authorization.k8s.io/kabanero-operator configured", "clusterrolebinding.rbac.authorization.k8s.io/kabanero-operator unchanged", "clusterrole.rbac.authorization.k8s.io/cluster-admin added: \"kabanero-operator\"", "NAME CREATED AT", "services.serving.knative.dev 2019-10-24T13:12:56Z", "NAME CREATED AT", "clustertasks.tekton.dev 2019-10-24T13:12:23Z", "config.operator.tekton.dev 2019-10-24T13:04:17Z", "pipelineresources.tekton.dev 2019-10-24T13:12:30Z", "pipelineruns.tekton.dev 2019-10-24T13:12:37Z", "pipelines.tekton.dev 2019-10-24T13:12:23Z", "taskruns.tekton.dev 2019-10-24T13:12:31Z", "tasks.tekton.dev 2019-10-24T13:12:37Z", "serviceaccount/tekton-webhooks-extension created", "clusterrole.rbac.authorization.k8s.io/tekton-webhooks-extension-minimal created", "clusterrolebinding.rbac.authorization.k8s.io/tekton-webhooks-extension-minimal created", "deployment.apps/webhooks-extension created", "service/webhooks-extension created", "service.serving.knative.dev/webhooks-extension-sink created", "task.tekton.dev/monitor-result-task created", "serviceaccount/tekton-dashboard created", "customresourcedefinition.apiextensions.k8s.io/extensions.dashboard.tekton.dev created", "clusterrole.rbac.authorization.k8s.io/tekton-dashboard-minimal created", "clusterrolebinding.rbac.authorization.k8s.io/tekton-dashboard-minimal created", "deployment.apps/tekton-dashboard created", "route.route.openshift.io/tekton-dashboard created", "service/tekton-dashboard created", "pipeline.tekton.dev/pipeline0 created", "task.tekton.dev/pipeline0-task created", "NAME AGE", "tekton-dashboard-minimal 1s", "clusterrole.rbac.authorization.k8s.io/tekton-dashboard-minimal patched", "deployment.extensions/tekton-dashboard scaled", "deployment.extensions/tekton-dashboard scaled", "configmap/config-domain patched", "NAME CREATED AT", "extensions.dashboard.tekton.dev 2019-10-24T13:14:28Z", "The installation script is complete. You can now create an instance", "of the Kabanero CR. If you have cloned and curated a collection set,", "apply the Kabanero CR that you created. Or, to create the default ", "instance:", "oc apply -n kabanero -f https://raw.githubusercontent.com/kabanero-io/kabanero-operator/0.2.0/config/samples/default.yaml" ] } TASK [install kabanero-foundation return code] ************************************************************************************ Thursday 24 October 2019 06:14:36 -0700 (0:00:00.085) 0:24:05.265 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": "0" } TASK [install kabanero foundation errors] ***************************************************************************************** Thursday 24 October 2019 06:14:36 -0700 (0:00:00.056) 0:24:05.321 ****** skipping: [svtcrc-91747-1.fyre.ibm.com] TASK [enable monitoring, alerting, and telemetry services] ************************************************************************ Thursday 24 October 2019 06:14:36 -0700 (0:00:00.058) 0:24:05.380 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [enable monitoring, alerting, and telemetry services stdout] ***************************************************************** Thursday 24 October 2019 06:14:38 -0700 (0:00:01.407) 0:24:06.787 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "statefulset.apps/alertmanager-main scaled", "statefulset.apps/prometheus-k8s scaled", "deployment.extensions/cluster-monitoring-operator scaled", "deployment.extensions/grafana scaled", "deployment.extensions/kube-state-metrics scaled", "deployment.extensions/openshift-state-metrics scaled", "deployment.extensions/prometheus-adapter scaled", "deployment.extensions/prometheus-operator scaled", "deployment.extensions/telemeter-client scaled" ] } TASK [oc version] ***************************************************************************************************************** Thursday 24 October 2019 06:14:38 -0700 (0:00:00.061) 0:24:06.848 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [oc version stdout] ********************************************************************************************************** Thursday 24 October 2019 06:14:39 -0700 (0:00:01.037) 0:24:07.886 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "Client Version: v4.3.0", "Server Version: 4.2.0", "Kubernetes Version: v1.14.6+2e5ed54" ] } PLAY RECAP ************************************************************************************************************************ svtcrc-91747-1.fyre.ibm.com : ok=36 changed=18 unreachable=0 failed=0 skipped=6 rescued=0 ignored=0 Thursday 24 October 2019 06:14:39 -0700 (0:00:00.029) 0:24:07.916 ****** =============================================================================== install kabanero-foundation ---------------------------------------------------------------------------------------------- 659.33s crc start ---------------------------------------------------------------------------------------------------------------- 656.96s Install NetworkManager FireFox -------------------------------------------------------------------------------------------- 73.99s crc setup ----------------------------------------------------------------------------------------------------------------- 38.73s retrieve Kabanero Foundation master zip ------------------------------------------------------------------------------------ 2.40s Gathering Facts ------------------------------------------------------------------------------------------------------------ 1.73s prereq directories --------------------------------------------------------------------------------------------------------- 1.51s oc login ------------------------------------------------------------------------------------------------------------------- 1.41s enable monitoring, alerting, and telemetry services ------------------------------------------------------------------------ 1.41s enable and start NetworkManager -------------------------------------------------------------------------------------------- 1.32s crc pull secret ------------------------------------------------------------------------------------------------------------ 1.11s unzip Kabanerio Foundation master zip -------------------------------------------------------------------------------------- 1.07s oc version ----------------------------------------------------------------------------------------------------------------- 1.04s check if crc is installed -------------------------------------------------------------------------------------------------- 0.71s whoami --------------------------------------------------------------------------------------------------------------------- 0.69s oc in user path ------------------------------------------------------------------------------------------------------------ 0.66s password masking process --------------------------------------------------------------------------------------------------- 0.64s crc creds ------------------------------------------------------------------------------------------------------------------ 0.50s check if .crc is created --------------------------------------------------------------------------------------------------- 0.48s link oc kubectl ------------------------------------------------------------------------------------------------------------ 0.44s set admin passwd ----------------------------------------------------------------------------------------------------------- 0.36s install kabanero-foundation stdout ----------------------------------------------------------------------------------------- 0.09s set_fact ------------------------------------------------------------------------------------------------------------------- 0.08s crc setup stdout ----------------------------------------------------------------------------------------------------------- 0.07s crc start stdout ----------------------------------------------------------------------------------------------------------- 0.07s crc creds stdout ----------------------------------------------------------------------------------------------------------- 0.06s set_fact ------------------------------------------------------------------------------------------------------------------- 0.06s oc login stdout ------------------------------------------------------------------------------------------------------------ 0.06s enable monitoring, alerting, and telemetry services stdout ----------------------------------------------------------------- 0.06s set_fact ------------------------------------------------------------------------------------------------------------------- 0.06s fail ----------------------------------------------------------------------------------------------------------------------- 0.06s crc delete ----------------------------------------------------------------------------------------------------------------- 0.06s crcstop stdout ------------------------------------------------------------------------------------------------------------- 0.06s install kabanero foundation errors ----------------------------------------------------------------------------------------- 0.06s install kabanero-foundation return code ------------------------------------------------------------------------------------ 0.06s crc stop ------------------------------------------------------------------------------------------------------------------- 0.05s set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s crc delete stdout ---------------------------------------------------------------------------------------------------------- 0.05s whoami stdout -------------------------------------------------------------------------------------------------------------- 0.05s oc version stdout ---------------------------------------------------------------------------------------------------------- 0.03s kabanero-crc complete. runtime 24 minutes Setup complete. overall runtime 29 minutes
2.0
SVT: TER: Kabanero Automation test execution master branch - Setup: svtcrc-91747-1 - Original logs: kabanero-svt01.fyre.ibm.com:/home/nest/kabanero-crc-logs/kabanero-auto-crc.sh.2019-10-24-05:45:33.setup.log PLAY [Create a one host Fyre stack] *********************************************************************************************** TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 05:45:37 -0700 (0:00:00.207) 0:00:00.207 ****** ok: [localhost] TASK [check status of the named cluster] ****************************************************************************************** Thursday 24 October 2019 05:45:37 -0700 (0:00:00.047) 0:00:00.255 ****** changed: [localhost] TASK [debug] ********************************************************************************************************************** Thursday 24 October 2019 05:45:38 -0700 (0:00:01.565) 0:00:01.821 ****** skipping: [localhost] TASK [get the public ssh id] ****************************************************************************************************** Thursday 24 October 2019 05:45:38 -0700 (0:00:00.057) 0:00:01.878 ****** changed: [localhost -> localhost] TASK [create input json file] ***************************************************************************************************** Thursday 24 October 2019 05:45:39 -0700 (0:00:00.306) 0:00:02.184 ****** changed: [localhost] TASK [Create Fyre stack] ********************************************************************************************************** Thursday 24 October 2019 05:45:40 -0700 (0:00:01.016) 0:00:03.201 ****** changed: [localhost] TASK [remove temp json] *********************************************************************************************************** Thursday 24 October 2019 05:45:41 -0700 (0:00:01.866) 0:00:05.068 ****** changed: [localhost -> localhost] TASK [check Fyre request status] ************************************************************************************************** Thursday 24 October 2019 05:45:42 -0700 (0:00:00.472) 0:00:05.540 ****** changed: [localhost] TASK [debug] ********************************************************************************************************************** Thursday 24 October 2019 05:45:43 -0700 (0:00:00.905) 0:00:06.445 ****** ok: [localhost] => { "fyreStatus.stdout": { "request": [ { "cluster_name": "", "created": "2019-10-24 05:45:41", "displayname": "steven r. schader", "error_details": "0", "id": "2420698", "status": "building", "type": "build", "updated": "2019-10-24 05:45:42", "username": "schader" } ] } } TASK [pause] ********************************************************************************************************************** Thursday 24 October 2019 05:45:43 -0700 (0:00:00.041) 0:00:06.487 ****** Pausing for 120 seconds (ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort) ok: [localhost] TASK [get Fyre request status] **************************************************************************************************** Thursday 24 October 2019 05:47:43 -0700 (0:02:00.047) 0:02:06.535 ****** changed: [localhost] TASK [check Fyre request status] ************************************************************************************************** Thursday 24 October 2019 05:47:44 -0700 (0:00:01.028) 0:02:07.564 ****** skipping: [localhost] TASK [check the host for a active ssh] ******************************************************************************************** Thursday 24 October 2019 05:47:44 -0700 (0:00:00.041) 0:02:07.605 ****** changed: [localhost -> localhost] TASK [remove new host from localhost known_hosts] ********************************************************************************* Thursday 24 October 2019 05:47:45 -0700 (0:00:00.582) 0:02:08.188 ****** changed: [localhost -> localhost] TASK [add host to known_hosts] **************************************************************************************************** Thursday 24 October 2019 05:47:45 -0700 (0:00:00.315) 0:02:08.503 ****** changed: [localhost -> localhost] TASK [create host inventory file] ************************************************************************************************* Thursday 24 October 2019 05:47:45 -0700 (0:00:00.378) 0:02:08.881 ****** changed: [localhost -> localhost] TASK [create plain file with host] ************************************************************************************************ Thursday 24 October 2019 05:47:46 -0700 (0:00:00.595) 0:02:09.476 ****** changed: [localhost -> localhost] TASK [create plain file with cluster name] **************************************************************************************** Thursday 24 October 2019 05:47:46 -0700 (0:00:00.563) 0:02:10.040 ****** changed: [localhost -> localhost] TASK [remove linked hosts.crc] **************************************************************************************************** Thursday 24 October 2019 05:47:47 -0700 (0:00:00.574) 0:02:10.615 ****** changed: [localhost -> localhost] TASK [link new host inventory file to hosts.crc] ********************************************************************************** Thursday 24 October 2019 05:47:47 -0700 (0:00:00.285) 0:02:10.901 ****** changed: [localhost -> localhost] PLAY RECAP ************************************************************************************************************************ localhost : ok=18 changed=15 unreachable=0 failed=0 skipped=2 rescued=0 ignored=0 Thursday 24 October 2019 05:47:48 -0700 (0:00:00.288) 0:02:11.189 ****** =============================================================================== pause -------------------------------------------------------------------------------------------------------------------- 120.05s Create Fyre stack ---------------------------------------------------------------------------------------------------------- 1.87s check status of the named cluster ------------------------------------------------------------------------------------------ 1.57s get Fyre request status ---------------------------------------------------------------------------------------------------- 1.03s create input json file ----------------------------------------------------------------------------------------------------- 1.02s check Fyre request status -------------------------------------------------------------------------------------------------- 0.91s create host inventory file ------------------------------------------------------------------------------------------------- 0.60s check the host for a active ssh -------------------------------------------------------------------------------------------- 0.58s create plain file with cluster name ---------------------------------------------------------------------------------------- 0.57s create plain file with host ------------------------------------------------------------------------------------------------ 0.56s remove temp json ----------------------------------------------------------------------------------------------------------- 0.47s add host to known_hosts ---------------------------------------------------------------------------------------------------- 0.38s remove new host from localhost known_hosts --------------------------------------------------------------------------------- 0.32s get the public ssh id ------------------------------------------------------------------------------------------------------ 0.31s link new host inventory file to hosts.crc ---------------------------------------------------------------------------------- 0.29s remove linked hosts.crc ---------------------------------------------------------------------------------------------------- 0.29s debug ---------------------------------------------------------------------------------------------------------------------- 0.06s set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s debug ---------------------------------------------------------------------------------------------------------------------- 0.04s check Fyre request status -------------------------------------------------------------------------------------------------- 0.04s fyrecrcvm complete. runtime 2 minutes PLAY [check RHEL 7 or greater x64 only] ******************************************************************************************* TASK [Gathering Facts] ************************************************************************************************************ Thursday 24 October 2019 05:47:50 -0700 (0:00:00.170) 0:00:00.170 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [Target host is RHEL7 ( or greater ) x86_64] ********************************************************************************* Thursday 24 October 2019 05:47:52 -0700 (0:00:01.803) 0:00:01.974 ****** ok: [svtcrc-91747-1.fyre.ibm.com] PLAY RECAP ************************************************************************************************************************ svtcrc-91747-1.fyre.ibm.com : ok=2 changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 Thursday 24 October 2019 05:47:52 -0700 (0:00:00.039) 0:00:02.014 ****** =============================================================================== Gathering Facts ------------------------------------------------------------------------------------------------------------ 1.80s Target host is RHEL7 ( or greater ) x86_64 --------------------------------------------------------------------------------- 0.04s PLAY [Create a User] ************************************************************************************************************** TASK [Generate User Password file] ************************************************************************************************ Thursday 24 October 2019 05:47:55 -0700 (0:00:00.183) 0:00:00.183 ****** ok: [svtcrc-91747-1.fyre.ibm.com -> localhost] TASK [Ensure User is Present] ***************************************************************************************************** Thursday 24 October 2019 05:47:56 -0700 (0:00:00.495) 0:00:00.678 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [authorized_key] ************************************************************************************************************* Thursday 24 October 2019 05:47:57 -0700 (0:00:01.163) 0:00:01.842 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [kevin sudoers] ************************************************************************************************************** Thursday 24 October 2019 05:47:58 -0700 (0:00:00.727) 0:00:02.569 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [kevin hard ulimit nofile] *************************************************************************************************** Thursday 24 October 2019 05:47:58 -0700 (0:00:00.750) 0:00:03.320 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [kevin soft ulimit nofile] *************************************************************************************************** Thursday 24 October 2019 05:47:59 -0700 (0:00:00.434) 0:00:03.754 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [add bin to path] ************************************************************************************************************ Thursday 24 October 2019 05:47:59 -0700 (0:00:00.437) 0:00:04.192 ****** changed: [svtcrc-91747-1.fyre.ibm.com] PLAY RECAP ************************************************************************************************************************ svtcrc-91747-1.fyre.ibm.com : ok=7 changed=6 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 Thursday 24 October 2019 05:48:00 -0700 (0:00:00.428) 0:00:04.620 ****** =============================================================================== Ensure User is Present ----------------------------------------------------------------------------------------------------- 1.16s kevin sudoers -------------------------------------------------------------------------------------------------------------- 0.75s authorized_key ------------------------------------------------------------------------------------------------------------- 0.73s Generate User Password file ------------------------------------------------------------------------------------------------ 0.50s kevin soft ulimit nofile --------------------------------------------------------------------------------------------------- 0.44s kevin hard ulimit nofile --------------------------------------------------------------------------------------------------- 0.43s add bin to path ------------------------------------------------------------------------------------------------------------ 0.43s PLAY [switch hosts.crc ansible_user] ********************************************************************************************** TASK [remove kevin ansible_user] ************************************************************************************************** Thursday 24 October 2019 05:48:02 -0700 (0:00:00.178) 0:00:00.178 ****** ok: [svtcrc-91747-1.fyre.ibm.com -> localhost] TASK [set kevin ansible_user] ***************************************************************************************************** Thursday 24 October 2019 05:48:03 -0700 (0:00:00.476) 0:00:00.654 ****** changed: [svtcrc-91747-1.fyre.ibm.com -> localhost] PLAY RECAP ************************************************************************************************************************ svtcrc-91747-1.fyre.ibm.com : ok=2 changed=1 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 Thursday 24 October 2019 05:48:03 -0700 (0:00:00.266) 0:00:00.921 ****** =============================================================================== remove kevin ansible_user -------------------------------------------------------------------------------------------------- 0.48s set kevin ansible_user ----------------------------------------------------------------------------------------------------- 0.27s PLAY [SVT - install OCP CodeReadyContainer binary] ******************************************************************************** TASK [Gathering Facts] ************************************************************************************************************ Thursday 24 October 2019 05:48:06 -0700 (0:00:00.171) 0:00:00.171 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [check if crc is installed] ************************************************************************************************** Thursday 24 October 2019 05:48:08 -0700 (0:00:01.862) 0:00:02.034 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [archive dir structure] ****************************************************************************************************** Thursday 24 October 2019 05:48:09 -0700 (0:00:00.653) 0:00:02.688 ****** changed: [svtcrc-91747-1.fyre.ibm.com -> localhost] TASK [debug] ********************************************************************************************************************** Thursday 24 October 2019 05:48:17 -0700 (0:00:08.477) 0:00:11.165 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "tarlist.stdout_lines": [ "crc-linux-1.0.0-amd64/", "crc-linux-1.0.0-amd64/LICENSE", "crc-linux-1.0.0-amd64/doc.pdf", "crc-linux-1.0.0-amd64/crc" ] } TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 05:48:17 -0700 (0:00:00.055) 0:00:11.221 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [copy crc archive] *********************************************************************************************************** Thursday 24 October 2019 05:48:17 -0700 (0:00:00.062) 0:00:11.284 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [install crc] **************************************************************************************************************** Thursday 24 October 2019 05:50:03 -0700 (0:01:46.017) 0:01:57.301 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [home bin] ******************************************************************************************************************* Thursday 24 October 2019 05:50:25 -0700 (0:00:22.091) 0:02:19.392 ****** changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/bin) TASK [link crc] ******************************************************************************************************************* Thursday 24 October 2019 05:50:26 -0700 (0:00:00.761) 0:02:20.154 ****** changed: [svtcrc-91747-1.fyre.ibm.com] PLAY RECAP ************************************************************************************************************************ svtcrc-91747-1.fyre.ibm.com : ok=9 changed=5 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 Thursday 24 October 2019 05:50:27 -0700 (0:00:00.483) 0:02:20.637 ****** =============================================================================== copy crc archive --------------------------------------------------------------------------------------------------------- 106.02s install crc --------------------------------------------------------------------------------------------------------------- 22.09s archive dir structure ------------------------------------------------------------------------------------------------------ 8.48s Gathering Facts ------------------------------------------------------------------------------------------------------------ 1.86s home bin ------------------------------------------------------------------------------------------------------------------- 0.76s check if crc is installed -------------------------------------------------------------------------------------------------- 0.65s link crc ------------------------------------------------------------------------------------------------------------------- 0.48s set_fact ------------------------------------------------------------------------------------------------------------------- 0.06s debug ---------------------------------------------------------------------------------------------------------------------- 0.06s crc-install complete. runtime 2 minutes PLAY [SVT - Kabanero on crc] ****************************************************************************************************** TASK [Gathering Facts] ************************************************************************************************************ Thursday 24 October 2019 05:50:31 -0700 (0:00:00.175) 0:00:00.175 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 05:50:33 -0700 (0:00:01.727) 0:00:01.902 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [Install NetworkManager FireFox] ********************************************************************************************* Thursday 24 October 2019 05:50:33 -0700 (0:00:00.060) 0:00:01.963 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [enable and start NetworkManager] ******************************************************************************************** Thursday 24 October 2019 05:51:47 -0700 (0:01:13.989) 0:01:15.952 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [check if crc is installed] ************************************************************************************************** Thursday 24 October 2019 05:51:48 -0700 (0:00:01.324) 0:01:17.276 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [fail] *********************************************************************************************************************** Thursday 24 October 2019 05:51:49 -0700 (0:00:00.713) 0:01:17.990 ****** skipping: [svtcrc-91747-1.fyre.ibm.com] TASK [prereq directories] ********************************************************************************************************* Thursday 24 October 2019 05:51:49 -0700 (0:00:00.059) 0:01:18.049 ****** changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/git) changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/testDir) changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/kabanero/logs) TASK [retrieve Kabanero Foundation master zip] ************************************************************************************ Thursday 24 October 2019 05:51:51 -0700 (0:00:01.507) 0:01:19.557 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [unzip Kabanerio Foundation master zip] ************************************************************************************** Thursday 24 October 2019 05:51:53 -0700 (0:00:02.402) 0:01:21.959 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [whoami] ********************************************************************************************************************* Thursday 24 October 2019 05:51:54 -0700 (0:00:01.073) 0:01:23.032 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [whoami stdout] ************************************************************************************************************** Thursday 24 October 2019 05:51:55 -0700 (0:00:00.689) 0:01:23.722 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "kevin" ] } TASK [check if .crc is created] *************************************************************************************************** Thursday 24 October 2019 05:51:55 -0700 (0:00:00.047) 0:01:23.770 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [crc stop] ******************************************************************************************************************* Thursday 24 October 2019 05:51:55 -0700 (0:00:00.481) 0:01:24.251 ****** skipping: [svtcrc-91747-1.fyre.ibm.com] TASK [crcstop stdout] ************************************************************************************************************* Thursday 24 October 2019 05:51:55 -0700 (0:00:00.053) 0:01:24.305 ****** skipping: [svtcrc-91747-1.fyre.ibm.com] TASK [crc delete] ***************************************************************************************************************** Thursday 24 October 2019 05:51:55 -0700 (0:00:00.058) 0:01:24.364 ****** skipping: [svtcrc-91747-1.fyre.ibm.com] TASK [crc delete stdout] ********************************************************************************************************** Thursday 24 October 2019 05:51:55 -0700 (0:00:00.059) 0:01:24.423 ****** skipping: [svtcrc-91747-1.fyre.ibm.com] TASK [crc setup] ****************************************************************************************************************** Thursday 24 October 2019 05:51:55 -0700 (0:00:00.050) 0:01:24.474 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [crc setup stdout] *********************************************************************************************************** Thursday 24 October 2019 05:52:34 -0700 (0:00:38.733) 0:02:03.207 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "level=info msg=\"Checking if running as non-root\"", "level=info msg=\"Caching oc binary\"", "level=info msg=\"Setting up virtualization\"", "level=info msg=\"Setting up KVM\"", "level=info msg=\"Installing libvirt service and dependencies\"", "level=info msg=\"Will use root access: install virtualization related packages\"", "level=info msg=\"Adding user to libvirt group\"", "level=info msg=\"Will use root access: add user to libvirt group\"", "level=info msg=\"Enabling libvirt\"", "level=info msg=\"Starting libvirt service\"", "level=info msg=\"Will use root access: start libvirtd service\"", "level=info msg=\"Checking if a supported libvirt version is installed\"", "level=info msg=\"Installing crc-driver-libvirt\"", "level=info msg=\"Removing older system-wide crc-driver-libvirt\"", "level=info msg=\"Setting up libvirt 'crc' network\"", "level=info msg=\"Starting libvirt 'crc' network\"", "level=info msg=\"Checking if NetworkManager is installed\"", "level=info msg=\"Checking if NetworkManager service is running\"", "level=info msg=\"Writing Network Manager config for crc\"", "level=info msg=\"Will use root access: write NetworkManager config in /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf\"", "level=info msg=\"Will use root access: execute systemctl daemon-reload command\"", "level=info msg=\"Will use root access: execute systemctl stop/start command\"", "level=info msg=\"Writing dnsmasq config for crc\"", "level=info msg=\"Will use root access: write dnsmasq configuration in /etc/NetworkManager/dnsmasq.d/crc.conf\"", "level=info msg=\"Will use root access: execute systemctl daemon-reload command\"", "level=info msg=\"Will use root access: execute systemctl stop/start command\"", "level=info msg=\"Unpacking bundle from the CRC binary\"", "Setup is complete, you can now run 'crc start' to start the OpenShift cluster" ] } TASK [crc pull secret] ************************************************************************************************************ Thursday 24 October 2019 05:52:34 -0700 (0:00:00.074) 0:02:03.281 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [crc start] ****************************************************************************************************************** Thursday 24 October 2019 05:52:35 -0700 (0:00:01.108) 0:02:04.390 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [crc start stdout] *********************************************************************************************************** Thursday 24 October 2019 06:03:32 -0700 (0:10:56.959) 0:13:01.349 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "level=info msg=\"Checking if running as non-root\"", "level=info msg=\"Checking if oc binary is cached\"", "level=info msg=\"Checking if Virtualization is enabled\"", "level=info msg=\"Checking if KVM is enabled\"", "level=info msg=\"Checking if libvirt is installed\"", "level=info msg=\"Checking if user is part of libvirt group\"", "level=info msg=\"Checking if libvirt is enabled\"", "level=info msg=\"Checking if libvirt daemon is running\"", "level=info msg=\"Checking if a supported libvirt version is installed\"", "level=info msg=\"Checking if crc-driver-libvirt is installed\"", "level=info msg=\"Checking if libvirt 'crc' network is available\"", "level=info msg=\"Checking if libvirt 'crc' network is active\"", "level=info msg=\"Checking if NetworkManager is installed\"", "level=info msg=\"Checking if NetworkManager service is running\"", "level=info msg=\"Checking if /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf exists\"", "level=info msg=\"Checking if /etc/NetworkManager/dnsmasq.d/crc.conf exists\"", "level=info msg=\"Extracting bundle: crc_libvirt_4.2.0.crcbundle ...\"", "level=info msg=\"Creating CodeReady Containers VM for OpenShift 4.2.0...\"", "level=info msg=\"Verifying validity of the cluster certificates ...\"", "level=info msg=\"Check internal and public DNS query ...\"", "level=info msg=\"Copying kubeconfig file to instance dir ...\"", "level=info msg=\"Adding user's pull secret and cluster ID ...\"", "level=info msg=\"Starting OpenShift cluster ... [waiting 3m]\"", "level=info", "level=info msg=\"To access the cluster, first set up your environment by following 'crc oc-env' instructions\"", "level=info msg=\"Then you can access it by running 'oc login -u developer -p developer https://api.crc.testing:6443'\"", "level=info msg=\"To login as an admin, username is 'kubeadmin' and password is xx-password-xx\"", "level=info", "level=info msg=\"You can now run 'crc console' and use these credentials to access the OpenShift web console\"", "The OpenShift cluster is running", "level=warning msg=\"The cluster might report a degraded or error state. This is expected since several operators have been disable to lower the resource usage. For more information, please consult the documentation\"" ] } TASK [oc in user path] ************************************************************************************************************ Thursday 24 October 2019 06:03:32 -0700 (0:00:00.066) 0:13:01.416 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [link oc kubectl] ************************************************************************************************************ Thursday 24 October 2019 06:03:33 -0700 (0:00:00.663) 0:13:02.079 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [crc creds] ****************************************************************************************************************** Thursday 24 October 2019 06:03:34 -0700 (0:00:00.438) 0:13:02.518 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [crc creds stdout] *********************************************************************************************************** Thursday 24 October 2019 06:03:34 -0700 (0:00:00.498) 0:13:03.016 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "To login as a regular user, username is 'developer' and password is 'developer'.", "To login as an admin, username is 'kubeadmin' and password is 'xx-password-xx'." ] } TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 06:03:34 -0700 (0:00:00.064) 0:13:03.081 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 06:03:34 -0700 (0:00:00.052) 0:13:03.134 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 06:03:34 -0700 (0:00:00.051) 0:13:03.185 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 06:03:34 -0700 (0:00:00.051) 0:13:03.236 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [password masking process] *************************************************************************************************** Thursday 24 October 2019 06:03:34 -0700 (0:00:00.076) 0:13:03.312 ****** changed: [svtcrc-91747-1.fyre.ibm.com -> localhost] TASK [set admin passwd] *********************************************************************************************************** Thursday 24 October 2019 06:03:35 -0700 (0:00:00.638) 0:13:03.951 ****** changed: [svtcrc-91747-1.fyre.ibm.com -> localhost] TASK [set_fact] ******************************************************************************************************************* Thursday 24 October 2019 06:03:35 -0700 (0:00:00.356) 0:13:04.307 ****** ok: [svtcrc-91747-1.fyre.ibm.com] TASK [oc login] ******************************************************************************************************************* Thursday 24 October 2019 06:03:35 -0700 (0:00:00.064) 0:13:04.371 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [oc login stdout] ************************************************************************************************************ Thursday 24 October 2019 06:03:37 -0700 (0:00:01.414) 0:13:05.786 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "Login successful.", "", "You have access to 51 projects, the list has been suppressed. You can list all projects with 'oc projects'", "", "Using project \"default\".", "Welcome! See 'oc help' to get started." ] } TASK [install kabanero-foundation] ************************************************************************************************ Thursday 24 October 2019 06:03:37 -0700 (0:00:00.062) 0:13:05.849 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [install kabanero-foundation stdout] ***************************************************************************************** Thursday 24 October 2019 06:14:36 -0700 (0:10:59.331) 0:24:05.180 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-ingress-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:default\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:prometheus\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-egressgateway-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-citadel-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-ingressgateway-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-cleanup-old-ca-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-mixer-post-install-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-mixer-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-pilot-service-account\"]", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-sidecar-injector-service-account\"]", "clusterrole.rbac.authorization.k8s.io/cluster-admin added: \"istio-galley-service-account\"", "securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:cluster-local-gateway-service-account\"]", "customresourcedefinition.apiextensions.k8s.io/virtualservices.networking.istio.io created", "customresourcedefinition.apiextensions.k8s.io/destinationrules.networking.istio.io created", "customresourcedefinition.apiextensions.k8s.io/serviceentries.networking.istio.io created", "customresourcedefinition.apiextensions.k8s.io/gateways.networking.istio.io created", "customresourcedefinition.apiextensions.k8s.io/envoyfilters.networking.istio.io created", "customresourcedefinition.apiextensions.k8s.io/clusterrbacconfigs.rbac.istio.io created", "customresourcedefinition.apiextensions.k8s.io/policies.authentication.istio.io created", "customresourcedefinition.apiextensions.k8s.io/meshpolicies.authentication.istio.io created", "customresourcedefinition.apiextensions.k8s.io/httpapispecbindings.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/httpapispecs.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/quotaspecbindings.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/quotaspecs.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/rules.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/attributemanifests.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/bypasses.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/circonuses.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/deniers.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/fluentds.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/kubernetesenvs.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/listcheckers.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/memquotas.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/noops.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/opas.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/prometheuses.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/rbacs.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/redisquotas.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/signalfxs.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/solarwindses.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/stackdrivers.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/statsds.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/stdios.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/apikeys.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/authorizations.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/checknothings.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/kuberneteses.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/listentries.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/logentries.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/edges.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/metrics.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/quotas.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/reportnothings.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/tracespans.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/rbacconfigs.rbac.istio.io created", "customresourcedefinition.apiextensions.k8s.io/serviceroles.rbac.istio.io created", "customresourcedefinition.apiextensions.k8s.io/servicerolebindings.rbac.istio.io created", "customresourcedefinition.apiextensions.k8s.io/adapters.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/instances.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/templates.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/handlers.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/cloudwatches.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/dogstatsds.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/sidecars.networking.istio.io created", "customresourcedefinition.apiextensions.k8s.io/zipkins.config.istio.io created", "customresourcedefinition.apiextensions.k8s.io/clusterissuers.certmanager.k8s.io created", "customresourcedefinition.apiextensions.k8s.io/issuers.certmanager.k8s.io created", "customresourcedefinition.apiextensions.k8s.io/certificates.certmanager.k8s.io created", "customresourcedefinition.apiextensions.k8s.io/orders.certmanager.k8s.io created", "customresourcedefinition.apiextensions.k8s.io/challenges.certmanager.k8s.io created", "namespace/istio-system created", "customresourcedefinition.apiextensions.k8s.io/virtualservices.networking.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/destinationrules.networking.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/serviceentries.networking.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/gateways.networking.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/envoyfilters.networking.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/clusterrbacconfigs.rbac.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/policies.authentication.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/meshpolicies.authentication.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/httpapispecbindings.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/httpapispecs.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/quotaspecbindings.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/quotaspecs.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/rules.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/attributemanifests.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/bypasses.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/circonuses.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/deniers.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/fluentds.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/kubernetesenvs.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/listcheckers.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/memquotas.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/noops.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/opas.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/prometheuses.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/rbacs.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/redisquotas.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/signalfxs.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/solarwindses.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/stackdrivers.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/statsds.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/stdios.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/apikeys.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/authorizations.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/checknothings.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/kuberneteses.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/listentries.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/logentries.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/edges.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/metrics.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/quotas.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/reportnothings.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/tracespans.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/rbacconfigs.rbac.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/serviceroles.rbac.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/servicerolebindings.rbac.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/adapters.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/instances.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/templates.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/handlers.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/cloudwatches.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/dogstatsds.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/sidecars.networking.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/zipkins.config.istio.io unchanged", "customresourcedefinition.apiextensions.k8s.io/clusterissuers.certmanager.k8s.io unchanged", "customresourcedefinition.apiextensions.k8s.io/issuers.certmanager.k8s.io unchanged", "customresourcedefinition.apiextensions.k8s.io/orders.certmanager.k8s.io unchanged", "customresourcedefinition.apiextensions.k8s.io/challenges.certmanager.k8s.io unchanged", "secret/kiali created", "configmap/istio-galley-configuration created", "configmap/istio-grafana-custom-resources created", "configmap/istio-grafana-configuration-dashboards-galley-dashboard created", "configmap/istio-grafana-configuration-dashboards-istio-mesh-dashboard created", "configmap/istio-grafana-configuration-dashboards-istio-performance-dashboard created", "configmap/istio-grafana-configuration-dashboards-istio-service-dashboard created", "configmap/istio-grafana-configuration-dashboards-istio-workload-dashboard created", "configmap/istio-grafana-configuration-dashboards-mixer-dashboard created", "configmap/istio-grafana-configuration-dashboards-pilot-dashboard created", "configmap/istio-grafana created", "configmap/kiali created", "configmap/prometheus created", "configmap/istio-security-custom-resources created", "configmap/istio created", "configmap/istio-sidecar-injector created", "serviceaccount/istio-galley-service-account created", "serviceaccount/istio-egressgateway-service-account created", "serviceaccount/istio-ingressgateway-service-account created", "serviceaccount/istio-grafana-post-install-account created", "clusterrole.rbac.authorization.k8s.io/istio-grafana-post-install-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-grafana-post-install-role-binding-istio-system created", "job.batch/istio-grafana-post-install-1.1.7 created", "serviceaccount/kiali-service-account created", "serviceaccount/istio-mixer-service-account created", "serviceaccount/istio-pilot-service-account created", "serviceaccount/prometheus created", "serviceaccount/istio-cleanup-secrets-service-account created", "clusterrole.rbac.authorization.k8s.io/istio-cleanup-secrets-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-cleanup-secrets-istio-system created", "job.batch/istio-cleanup-secrets-1.1.7 created", "serviceaccount/istio-security-post-install-account created", "clusterrole.rbac.authorization.k8s.io/istio-security-post-install-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-security-post-install-role-binding-istio-system created", "job.batch/istio-security-post-install-1.1.7 created", "serviceaccount/istio-citadel-service-account created", "serviceaccount/istio-sidecar-injector-service-account created", "serviceaccount/istio-multi created", "clusterrole.rbac.authorization.k8s.io/istio-galley-istio-system created", "clusterrole.rbac.authorization.k8s.io/istio-egressgateway-istio-system created", "clusterrole.rbac.authorization.k8s.io/istio-ingressgateway-istio-system created", "clusterrole.rbac.authorization.k8s.io/kiali created", "clusterrole.rbac.authorization.k8s.io/kiali-viewer created", "clusterrole.rbac.authorization.k8s.io/istio-mixer-istio-system created", "clusterrole.rbac.authorization.k8s.io/istio-pilot-istio-system created", "clusterrole.rbac.authorization.k8s.io/prometheus-istio-system created", "clusterrole.rbac.authorization.k8s.io/istio-citadel-istio-system created", "clusterrole.rbac.authorization.k8s.io/istio-sidecar-injector-istio-system created", "clusterrole.rbac.authorization.k8s.io/istio-reader created", "clusterrolebinding.rbac.authorization.k8s.io/istio-galley-admin-role-binding-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-egressgateway-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-ingressgateway-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-kiali-admin-role-binding-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-mixer-admin-role-binding-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-pilot-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/prometheus-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-citadel-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-sidecar-injector-admin-role-binding-istio-system created", "clusterrolebinding.rbac.authorization.k8s.io/istio-multi created", "role.rbac.authorization.k8s.io/istio-ingressgateway-sds created", "rolebinding.rbac.authorization.k8s.io/istio-ingressgateway-sds created", "service/istio-galley created", "service/istio-egressgateway created", "service/istio-ingressgateway created", "service/grafana created", "service/kiali created", "service/istio-policy created", "service/istio-telemetry created", "service/istio-pilot created", "service/prometheus created", "service/istio-citadel created", "service/istio-sidecar-injector created", "deployment.extensions/istio-galley created", "deployment.extensions/istio-egressgateway created", "deployment.extensions/istio-ingressgateway created", "deployment.extensions/grafana created", "deployment.extensions/kiali created", "deployment.extensions/istio-policy created", "deployment.extensions/istio-telemetry created", "deployment.extensions/istio-pilot created", "deployment.extensions/prometheus created", "deployment.extensions/istio-citadel created", "deployment.extensions/istio-sidecar-injector created", "deployment.extensions/istio-tracing created", "horizontalpodautoscaler.autoscaling/istio-egressgateway created", "horizontalpodautoscaler.autoscaling/istio-ingressgateway created", "horizontalpodautoscaler.autoscaling/istio-policy created", "horizontalpodautoscaler.autoscaling/istio-telemetry created", "horizontalpodautoscaler.autoscaling/istio-pilot created", "service/jaeger-query created", "service/jaeger-collector created", "service/jaeger-agent created", "service/zipkin created", "service/tracing created", "mutatingwebhookconfiguration.admissionregistration.k8s.io/istio-sidecar-injector created", "poddisruptionbudget.policy/istio-galley created", "poddisruptionbudget.policy/istio-egressgateway created", "poddisruptionbudget.policy/istio-ingressgateway created", "poddisruptionbudget.policy/istio-policy created", "poddisruptionbudget.policy/istio-telemetry created", "poddisruptionbudget.policy/istio-pilot created", "attributemanifest.config.istio.io/istioproxy created", "attributemanifest.config.istio.io/kubernetes created", "handler.config.istio.io/stdio created", "logentry.config.istio.io/accesslog created", "logentry.config.istio.io/tcpaccesslog created", "rule.config.istio.io/stdio created", "rule.config.istio.io/stdiotcp created", "metric.config.istio.io/requestcount created", "metric.config.istio.io/requestduration created", "metric.config.istio.io/requestsize created", "metric.config.istio.io/responsesize created", "metric.config.istio.io/tcpbytesent created", "metric.config.istio.io/tcpbytereceived created", "metric.config.istio.io/tcpconnectionsopened created", "metric.config.istio.io/tcpconnectionsclosed created", "handler.config.istio.io/prometheus created", "rule.config.istio.io/promhttp created", "rule.config.istio.io/promtcp created", "rule.config.istio.io/promtcpconnectionopen created", "rule.config.istio.io/promtcpconnectionclosed created", "handler.config.istio.io/kubernetesenv created", "rule.config.istio.io/kubeattrgenrulerule created", "rule.config.istio.io/tcpkubeattrgenrulerule created", "kubernetes.config.istio.io/attributes created", "destinationrule.networking.istio.io/istio-policy created", "destinationrule.networking.istio.io/istio-telemetry created", "Now using project \"kabanero\" on server \"https://api.crc.testing:6443\".", "", "You can add applications to this project with the 'new-app' command. For example, try:", "", " oc new-app django-psql-example", "", "to build a new example application in Python. Or use kubectl to deploy a simple Kubernetes application:", "", " kubectl create deployment hello-node --image=gcr.io/hello-minikube-zero-install/hello-node", "", "namespace/kabanero configured", "deployment.apps/knative-eventing-operator created", "role.rbac.authorization.k8s.io/knative-eventing-operator created", "clusterrole.rbac.authorization.k8s.io/knative-eventing-operator created", "rolebinding.rbac.authorization.k8s.io/knative-eventing-operator created", "clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-operator created", "serviceaccount/knative-eventing-operator created", "customresourcedefinition.apiextensions.k8s.io/knativeeventings.eventing.knative.dev created", "deployment.apps/knative-serving-operator created", "role.rbac.authorization.k8s.io/knative-serving-operator created", "clusterrole.rbac.authorization.k8s.io/knative-serving-operator created", "rolebinding.rbac.authorization.k8s.io/knative-serving-operator created", "clusterrolebinding.rbac.authorization.k8s.io/knative-serving-operator created", "serviceaccount/knative-serving-operator created", "customresourcedefinition.apiextensions.k8s.io/knativeservings.serving.knative.dev created", "namespace/knative-sources created", "serviceaccount/controller-manager created", "clusterrole.rbac.authorization.k8s.io/eventing-sources-controller created", "clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-controller created", "clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-addressable-resolver created", "customresourcedefinition.apiextensions.k8s.io/githubsources.sources.eventing.knative.dev created", "service/controller created", "statefulset.apps/controller-manager created", "deployment.apps/openshift-pipelines-operator created", "clusterrole.rbac.authorization.k8s.io/openshift-pipelines-operator created", "clusterrolebinding.rbac.authorization.k8s.io/openshift-pipelines-operator created", "serviceaccount/openshift-pipelines-operator created", "namespace/openshift-pipelines created", "customresourcedefinition.apiextensions.k8s.io/config.operator.tekton.dev created", "clusterrole.rbac.authorization.k8s.io/knative-serving-operator configured", "clusterrole.rbac.authorization.k8s.io/knative-eventing-operator configured", "customresourcedefinition.apiextensions.k8s.io/kabaneros.kabanero.io created", "customresourcedefinition.apiextensions.k8s.io/collections.kabanero.io created", "serviceaccount/kabanero-operator created", "deployment.apps/kabanero-operator created", "clusterrole.rbac.authorization.k8s.io/kabanero-operator created", "clusterrolebinding.rbac.authorization.k8s.io/kabanero-operator created", "namespace/kabanero unchanged", "deployment.apps/knative-eventing-operator configured", "role.rbac.authorization.k8s.io/knative-eventing-operator configured", "clusterrole.rbac.authorization.k8s.io/knative-eventing-operator configured", "rolebinding.rbac.authorization.k8s.io/knative-eventing-operator unchanged", "clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-operator unchanged", "serviceaccount/knative-eventing-operator unchanged", "customresourcedefinition.apiextensions.k8s.io/knativeeventings.eventing.knative.dev unchanged", "deployment.apps/knative-serving-operator configured", "role.rbac.authorization.k8s.io/knative-serving-operator configured", "clusterrole.rbac.authorization.k8s.io/knative-serving-operator configured", "rolebinding.rbac.authorization.k8s.io/knative-serving-operator unchanged", "clusterrolebinding.rbac.authorization.k8s.io/knative-serving-operator unchanged", "serviceaccount/knative-serving-operator unchanged", "customresourcedefinition.apiextensions.k8s.io/knativeservings.serving.knative.dev unchanged", "namespace/knative-sources unchanged", "serviceaccount/controller-manager unchanged", "clusterrole.rbac.authorization.k8s.io/eventing-sources-controller unchanged", "clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-controller unchanged", "clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-addressable-resolver unchanged", "customresourcedefinition.apiextensions.k8s.io/githubsources.sources.eventing.knative.dev unchanged", "service/controller unchanged", "statefulset.apps/controller-manager configured", "deployment.apps/openshift-pipelines-operator configured", "clusterrole.rbac.authorization.k8s.io/openshift-pipelines-operator configured", "clusterrolebinding.rbac.authorization.k8s.io/openshift-pipelines-operator unchanged", "serviceaccount/openshift-pipelines-operator unchanged", "namespace/openshift-pipelines unchanged", "customresourcedefinition.apiextensions.k8s.io/config.operator.tekton.dev unchanged", "config.operator.tekton.dev/cluster created", "clusterrole.rbac.authorization.k8s.io/knative-serving-operator configured", "clusterrole.rbac.authorization.k8s.io/knative-eventing-operator configured", "customresourcedefinition.apiextensions.k8s.io/kabaneros.kabanero.io configured", "customresourcedefinition.apiextensions.k8s.io/collections.kabanero.io configured", "serviceaccount/kabanero-operator unchanged", "deployment.apps/kabanero-operator configured", "clusterrole.rbac.authorization.k8s.io/kabanero-operator configured", "clusterrolebinding.rbac.authorization.k8s.io/kabanero-operator unchanged", "clusterrole.rbac.authorization.k8s.io/cluster-admin added: \"kabanero-operator\"", "NAME CREATED AT", "services.serving.knative.dev 2019-10-24T13:12:56Z", "NAME CREATED AT", "clustertasks.tekton.dev 2019-10-24T13:12:23Z", "config.operator.tekton.dev 2019-10-24T13:04:17Z", "pipelineresources.tekton.dev 2019-10-24T13:12:30Z", "pipelineruns.tekton.dev 2019-10-24T13:12:37Z", "pipelines.tekton.dev 2019-10-24T13:12:23Z", "taskruns.tekton.dev 2019-10-24T13:12:31Z", "tasks.tekton.dev 2019-10-24T13:12:37Z", "serviceaccount/tekton-webhooks-extension created", "clusterrole.rbac.authorization.k8s.io/tekton-webhooks-extension-minimal created", "clusterrolebinding.rbac.authorization.k8s.io/tekton-webhooks-extension-minimal created", "deployment.apps/webhooks-extension created", "service/webhooks-extension created", "service.serving.knative.dev/webhooks-extension-sink created", "task.tekton.dev/monitor-result-task created", "serviceaccount/tekton-dashboard created", "customresourcedefinition.apiextensions.k8s.io/extensions.dashboard.tekton.dev created", "clusterrole.rbac.authorization.k8s.io/tekton-dashboard-minimal created", "clusterrolebinding.rbac.authorization.k8s.io/tekton-dashboard-minimal created", "deployment.apps/tekton-dashboard created", "route.route.openshift.io/tekton-dashboard created", "service/tekton-dashboard created", "pipeline.tekton.dev/pipeline0 created", "task.tekton.dev/pipeline0-task created", "NAME AGE", "tekton-dashboard-minimal 1s", "clusterrole.rbac.authorization.k8s.io/tekton-dashboard-minimal patched", "deployment.extensions/tekton-dashboard scaled", "deployment.extensions/tekton-dashboard scaled", "configmap/config-domain patched", "NAME CREATED AT", "extensions.dashboard.tekton.dev 2019-10-24T13:14:28Z", "The installation script is complete. You can now create an instance", "of the Kabanero CR. If you have cloned and curated a collection set,", "apply the Kabanero CR that you created. Or, to create the default ", "instance:", "oc apply -n kabanero -f https://raw.githubusercontent.com/kabanero-io/kabanero-operator/0.2.0/config/samples/default.yaml" ] } TASK [install kabanero-foundation return code] ************************************************************************************ Thursday 24 October 2019 06:14:36 -0700 (0:00:00.085) 0:24:05.265 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": "0" } TASK [install kabanero foundation errors] ***************************************************************************************** Thursday 24 October 2019 06:14:36 -0700 (0:00:00.056) 0:24:05.321 ****** skipping: [svtcrc-91747-1.fyre.ibm.com] TASK [enable monitoring, alerting, and telemetry services] ************************************************************************ Thursday 24 October 2019 06:14:36 -0700 (0:00:00.058) 0:24:05.380 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [enable monitoring, alerting, and telemetry services stdout] ***************************************************************** Thursday 24 October 2019 06:14:38 -0700 (0:00:01.407) 0:24:06.787 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "statefulset.apps/alertmanager-main scaled", "statefulset.apps/prometheus-k8s scaled", "deployment.extensions/cluster-monitoring-operator scaled", "deployment.extensions/grafana scaled", "deployment.extensions/kube-state-metrics scaled", "deployment.extensions/openshift-state-metrics scaled", "deployment.extensions/prometheus-adapter scaled", "deployment.extensions/prometheus-operator scaled", "deployment.extensions/telemeter-client scaled" ] } TASK [oc version] ***************************************************************************************************************** Thursday 24 October 2019 06:14:38 -0700 (0:00:00.061) 0:24:06.848 ****** changed: [svtcrc-91747-1.fyre.ibm.com] TASK [oc version stdout] ********************************************************************************************************** Thursday 24 October 2019 06:14:39 -0700 (0:00:01.037) 0:24:07.886 ****** ok: [svtcrc-91747-1.fyre.ibm.com] => { "msg": [ "Client Version: v4.3.0", "Server Version: 4.2.0", "Kubernetes Version: v1.14.6+2e5ed54" ] } PLAY RECAP ************************************************************************************************************************ svtcrc-91747-1.fyre.ibm.com : ok=36 changed=18 unreachable=0 failed=0 skipped=6 rescued=0 ignored=0 Thursday 24 October 2019 06:14:39 -0700 (0:00:00.029) 0:24:07.916 ****** =============================================================================== install kabanero-foundation ---------------------------------------------------------------------------------------------- 659.33s crc start ---------------------------------------------------------------------------------------------------------------- 656.96s Install NetworkManager FireFox -------------------------------------------------------------------------------------------- 73.99s crc setup ----------------------------------------------------------------------------------------------------------------- 38.73s retrieve Kabanero Foundation master zip ------------------------------------------------------------------------------------ 2.40s Gathering Facts ------------------------------------------------------------------------------------------------------------ 1.73s prereq directories --------------------------------------------------------------------------------------------------------- 1.51s oc login ------------------------------------------------------------------------------------------------------------------- 1.41s enable monitoring, alerting, and telemetry services ------------------------------------------------------------------------ 1.41s enable and start NetworkManager -------------------------------------------------------------------------------------------- 1.32s crc pull secret ------------------------------------------------------------------------------------------------------------ 1.11s unzip Kabanerio Foundation master zip -------------------------------------------------------------------------------------- 1.07s oc version ----------------------------------------------------------------------------------------------------------------- 1.04s check if crc is installed -------------------------------------------------------------------------------------------------- 0.71s whoami --------------------------------------------------------------------------------------------------------------------- 0.69s oc in user path ------------------------------------------------------------------------------------------------------------ 0.66s password masking process --------------------------------------------------------------------------------------------------- 0.64s crc creds ------------------------------------------------------------------------------------------------------------------ 0.50s check if .crc is created --------------------------------------------------------------------------------------------------- 0.48s link oc kubectl ------------------------------------------------------------------------------------------------------------ 0.44s set admin passwd ----------------------------------------------------------------------------------------------------------- 0.36s install kabanero-foundation stdout ----------------------------------------------------------------------------------------- 0.09s set_fact ------------------------------------------------------------------------------------------------------------------- 0.08s crc setup stdout ----------------------------------------------------------------------------------------------------------- 0.07s crc start stdout ----------------------------------------------------------------------------------------------------------- 0.07s crc creds stdout ----------------------------------------------------------------------------------------------------------- 0.06s set_fact ------------------------------------------------------------------------------------------------------------------- 0.06s oc login stdout ------------------------------------------------------------------------------------------------------------ 0.06s enable monitoring, alerting, and telemetry services stdout ----------------------------------------------------------------- 0.06s set_fact ------------------------------------------------------------------------------------------------------------------- 0.06s fail ----------------------------------------------------------------------------------------------------------------------- 0.06s crc delete ----------------------------------------------------------------------------------------------------------------- 0.06s crcstop stdout ------------------------------------------------------------------------------------------------------------- 0.06s install kabanero foundation errors ----------------------------------------------------------------------------------------- 0.06s install kabanero-foundation return code ------------------------------------------------------------------------------------ 0.06s crc stop ------------------------------------------------------------------------------------------------------------------- 0.05s set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s crc delete stdout ---------------------------------------------------------------------------------------------------------- 0.05s whoami stdout -------------------------------------------------------------------------------------------------------------- 0.05s oc version stdout ---------------------------------------------------------------------------------------------------------- 0.03s kabanero-crc complete. runtime 24 minutes Setup complete. overall runtime 29 minutes
test
svt ter kabanero automation test execution master branch setup svtcrc original logs kabanero fyre ibm com home nest kabanero crc logs kabanero auto crc sh setup log play task thursday october ok task thursday october changed task thursday october skipping task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october ok fyrestatus stdout request cluster name created displayname steven r schader error details id status building type build updated username schader task thursday october pausing for seconds ctrl c then c continue early ctrl c then a abort ok task thursday october changed task thursday october skipping task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed play recap localhost ok changed unreachable failed skipped rescued ignored thursday october pause create fyre stack check status of the named cluster get fyre request status create input json file check fyre request status create host inventory file check the host for a active ssh create plain file with cluster name create plain file with host remove temp json add host to known hosts remove new host from localhost known hosts get the public ssh id link new host inventory file to hosts crc remove linked hosts crc debug set fact debug check fyre request status fyrecrcvm complete runtime minutes play task thursday october ok task thursday october ok play recap svtcrc fyre ibm com ok changed unreachable failed skipped rescued ignored thursday october gathering facts target host is or greater play task thursday october ok task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed play recap svtcrc fyre ibm com ok changed unreachable failed skipped rescued ignored thursday october ensure user is present kevin sudoers authorized key generate user password file kevin soft ulimit nofile kevin hard ulimit nofile add bin to path play task thursday october ok task thursday october changed play recap svtcrc fyre ibm com ok changed unreachable failed skipped rescued ignored thursday october remove kevin ansible user set kevin ansible user play task thursday october ok task thursday october ok task thursday october changed task thursday october ok tarlist stdout lines crc linux crc linux license crc linux doc pdf crc linux crc task thursday october ok task thursday october changed task thursday october changed task thursday october changed item bin task thursday october changed play recap svtcrc fyre ibm com ok changed unreachable failed skipped rescued ignored thursday october copy crc archive install crc archive dir structure gathering facts home bin check if crc is installed link crc set fact debug crc install complete runtime minutes play task thursday october ok task thursday october ok task thursday october changed task thursday october changed task thursday october ok task thursday october skipping task thursday october changed item git changed item testdir changed item kabanero logs task thursday october changed task thursday october changed task thursday october changed task thursday october ok msg kevin task thursday october ok task thursday october skipping task thursday october skipping task thursday october skipping task thursday october skipping task thursday october changed task thursday october ok msg level info msg checking if running as non root level info msg caching oc binary level info msg setting up virtualization level info msg setting up kvm level info msg installing libvirt service and dependencies level info msg will use root access install virtualization related packages level info msg adding user to libvirt group level info msg will use root access add user to libvirt group level info msg enabling libvirt level info msg starting libvirt service level info msg will use root access start libvirtd service level info msg checking if a supported libvirt version is installed level info msg installing crc driver libvirt level info msg removing older system wide crc driver libvirt level info msg setting up libvirt crc network level info msg starting libvirt crc network level info msg checking if networkmanager is installed level info msg checking if networkmanager service is running level info msg writing network manager config for crc level info msg will use root access write networkmanager config in etc networkmanager conf d crc nm dnsmasq conf level info msg will use root access execute systemctl daemon reload command level info msg will use root access execute systemctl stop start command level info msg writing dnsmasq config for crc level info msg will use root access write dnsmasq configuration in etc networkmanager dnsmasq d crc conf level info msg will use root access execute systemctl daemon reload command level info msg will use root access execute systemctl stop start command level info msg unpacking bundle from the crc binary setup is complete you can now run crc start to start the openshift cluster task thursday october changed task thursday october changed task thursday october ok msg level info msg checking if running as non root level info msg checking if oc binary is cached level info msg checking if virtualization is enabled level info msg checking if kvm is enabled level info msg checking if libvirt is installed level info msg checking if user is part of libvirt group level info msg checking if libvirt is enabled level info msg checking if libvirt daemon is running level info msg checking if a supported libvirt version is installed level info msg checking if crc driver libvirt is installed level info msg checking if libvirt crc network is available level info msg checking if libvirt crc network is active level info msg checking if networkmanager is installed level info msg checking if networkmanager service is running level info msg checking if etc networkmanager conf d crc nm dnsmasq conf exists level info msg checking if etc networkmanager dnsmasq d crc conf exists level info msg extracting bundle crc libvirt crcbundle level info msg creating codeready containers vm for openshift level info msg verifying validity of the cluster certificates level info msg check internal and public dns query level info msg copying kubeconfig file to instance dir level info msg adding user s pull secret and cluster id level info msg starting openshift cluster level info level info msg to access the cluster first set up your environment by following crc oc env instructions level info msg then you can access it by running oc login u developer p developer level info msg to login as an admin username is kubeadmin and password is xx password xx level info level info msg you can now run crc console and use these credentials to access the openshift web console the openshift cluster is running level warning msg the cluster might report a degraded or error state this is expected since several operators have been disable to lower the resource usage for more information please consult the documentation task thursday october changed task thursday october changed task thursday october changed task thursday october ok msg to login as a regular user username is developer and password is developer to login as an admin username is kubeadmin and password is xx password xx task thursday october ok task thursday october ok task thursday october ok task thursday october ok task thursday october changed task thursday october changed task thursday october ok task thursday october changed task thursday october ok msg login successful you have access to projects the list has been suppressed you can list all projects with oc projects using project default welcome see oc help to get started task thursday october changed task thursday october ok msg securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to clusterrole rbac authorization io cluster admin added istio galley service account securitycontextconstraints security openshift io anyuid added to customresourcedefinition apiextensions io virtualservices networking istio io created customresourcedefinition apiextensions io destinationrules networking istio io created customresourcedefinition apiextensions io serviceentries networking istio io created customresourcedefinition apiextensions io gateways networking istio io created customresourcedefinition apiextensions io envoyfilters networking istio io created customresourcedefinition apiextensions io clusterrbacconfigs rbac istio io created customresourcedefinition apiextensions io policies authentication istio io created customresourcedefinition apiextensions io meshpolicies authentication istio io created customresourcedefinition apiextensions io httpapispecbindings config istio io created customresourcedefinition apiextensions io httpapispecs config istio io created customresourcedefinition apiextensions io quotaspecbindings config istio io created customresourcedefinition apiextensions io quotaspecs config istio io created customresourcedefinition apiextensions io rules config istio io created customresourcedefinition apiextensions io attributemanifests config istio io created customresourcedefinition apiextensions io bypasses config istio io created customresourcedefinition apiextensions io circonuses config istio io created customresourcedefinition apiextensions io deniers config istio io created customresourcedefinition apiextensions io fluentds config istio io created customresourcedefinition apiextensions io kubernetesenvs config istio io created customresourcedefinition apiextensions io listcheckers config istio io created customresourcedefinition apiextensions io memquotas config istio io created customresourcedefinition apiextensions io noops config istio io created customresourcedefinition apiextensions io opas config istio io created customresourcedefinition apiextensions io prometheuses config istio io created customresourcedefinition apiextensions io rbacs config istio io created customresourcedefinition apiextensions io redisquotas config istio io created customresourcedefinition apiextensions io signalfxs config istio io created customresourcedefinition apiextensions io solarwindses config istio io created customresourcedefinition apiextensions io stackdrivers config istio io created customresourcedefinition apiextensions io statsds config istio io created customresourcedefinition apiextensions io stdios config istio io created customresourcedefinition apiextensions io apikeys config istio io created customresourcedefinition apiextensions io authorizations config istio io created customresourcedefinition apiextensions io checknothings config istio io created customresourcedefinition apiextensions io kuberneteses config istio io created customresourcedefinition apiextensions io listentries config istio io created customresourcedefinition apiextensions io logentries config istio io created customresourcedefinition apiextensions io edges config istio io created customresourcedefinition apiextensions io metrics config istio io created customresourcedefinition apiextensions io quotas config istio io created customresourcedefinition apiextensions io reportnothings config istio io created customresourcedefinition apiextensions io tracespans config istio io created customresourcedefinition apiextensions io rbacconfigs rbac istio io created customresourcedefinition apiextensions io serviceroles rbac istio io created customresourcedefinition apiextensions io servicerolebindings rbac istio io created customresourcedefinition apiextensions io adapters config istio io created customresourcedefinition apiextensions io instances config istio io created customresourcedefinition apiextensions io templates config istio io created customresourcedefinition apiextensions io handlers config istio io created customresourcedefinition apiextensions io cloudwatches config istio io created customresourcedefinition apiextensions io dogstatsds config istio io created customresourcedefinition apiextensions io sidecars networking istio io created customresourcedefinition apiextensions io zipkins config istio io created customresourcedefinition apiextensions io clusterissuers certmanager io created customresourcedefinition apiextensions io issuers certmanager io created customresourcedefinition apiextensions io certificates certmanager io created customresourcedefinition apiextensions io orders certmanager io created customresourcedefinition apiextensions io challenges certmanager io created namespace istio system created customresourcedefinition apiextensions io virtualservices networking istio io unchanged customresourcedefinition apiextensions io destinationrules networking istio io unchanged customresourcedefinition apiextensions io serviceentries networking istio io unchanged customresourcedefinition apiextensions io gateways networking istio io unchanged customresourcedefinition apiextensions io envoyfilters networking istio io unchanged customresourcedefinition apiextensions io clusterrbacconfigs rbac istio io unchanged customresourcedefinition apiextensions io policies authentication istio io unchanged customresourcedefinition apiextensions io meshpolicies authentication istio io unchanged customresourcedefinition apiextensions io httpapispecbindings config istio io unchanged customresourcedefinition apiextensions io httpapispecs config istio io unchanged customresourcedefinition apiextensions io quotaspecbindings config istio io unchanged customresourcedefinition apiextensions io quotaspecs config istio io unchanged customresourcedefinition apiextensions io rules config istio io unchanged customresourcedefinition apiextensions io attributemanifests config istio io unchanged customresourcedefinition apiextensions io bypasses config istio io unchanged customresourcedefinition apiextensions io circonuses config istio io unchanged customresourcedefinition apiextensions io deniers config istio io unchanged customresourcedefinition apiextensions io fluentds config istio io unchanged customresourcedefinition apiextensions io kubernetesenvs config istio io unchanged customresourcedefinition apiextensions io listcheckers config istio io unchanged customresourcedefinition apiextensions io memquotas config istio io unchanged customresourcedefinition apiextensions io noops config istio io unchanged customresourcedefinition apiextensions io opas config istio io unchanged customresourcedefinition apiextensions io prometheuses config istio io unchanged customresourcedefinition apiextensions io rbacs config istio io unchanged customresourcedefinition apiextensions io redisquotas config istio io unchanged customresourcedefinition apiextensions io signalfxs config istio io unchanged customresourcedefinition apiextensions io solarwindses config istio io unchanged customresourcedefinition apiextensions io stackdrivers config istio io unchanged customresourcedefinition apiextensions io statsds config istio io unchanged customresourcedefinition apiextensions io stdios config istio io unchanged customresourcedefinition apiextensions io apikeys config istio io unchanged customresourcedefinition apiextensions io authorizations config istio io unchanged customresourcedefinition apiextensions io checknothings config istio io unchanged customresourcedefinition apiextensions io kuberneteses config istio io unchanged customresourcedefinition apiextensions io listentries config istio io unchanged customresourcedefinition apiextensions io logentries config istio io unchanged customresourcedefinition apiextensions io edges config istio io unchanged customresourcedefinition apiextensions io metrics config istio io unchanged customresourcedefinition apiextensions io quotas config istio io unchanged customresourcedefinition apiextensions io reportnothings config istio io unchanged customresourcedefinition apiextensions io tracespans config istio io unchanged customresourcedefinition apiextensions io rbacconfigs rbac istio io unchanged customresourcedefinition apiextensions io serviceroles rbac istio io unchanged customresourcedefinition apiextensions io servicerolebindings rbac istio io unchanged customresourcedefinition apiextensions io adapters config istio io unchanged customresourcedefinition apiextensions io instances config istio io unchanged customresourcedefinition apiextensions io templates config istio io unchanged customresourcedefinition apiextensions io handlers config istio io unchanged customresourcedefinition apiextensions io cloudwatches config istio io unchanged customresourcedefinition apiextensions io dogstatsds config istio io unchanged customresourcedefinition apiextensions io sidecars networking istio io unchanged customresourcedefinition apiextensions io zipkins config istio io unchanged customresourcedefinition apiextensions io clusterissuers certmanager io unchanged customresourcedefinition apiextensions io issuers certmanager io unchanged customresourcedefinition apiextensions io orders certmanager io unchanged customresourcedefinition apiextensions io challenges certmanager io unchanged secret kiali created configmap istio galley configuration created configmap istio grafana custom resources created configmap istio grafana configuration dashboards galley dashboard created configmap istio grafana configuration dashboards istio mesh dashboard created configmap istio grafana configuration dashboards istio performance dashboard created configmap istio grafana configuration dashboards istio service dashboard created configmap istio grafana configuration dashboards istio workload dashboard created configmap istio grafana configuration dashboards mixer dashboard created configmap istio grafana configuration dashboards pilot dashboard created configmap istio grafana created configmap kiali created configmap prometheus created configmap istio security custom resources created configmap istio created configmap istio sidecar injector created serviceaccount istio galley service account created serviceaccount istio egressgateway service account created serviceaccount istio ingressgateway service account created serviceaccount istio grafana post install account created clusterrole rbac authorization io istio grafana post install istio system created clusterrolebinding rbac authorization io istio grafana post install role binding istio system created job batch istio grafana post install created serviceaccount kiali service account created serviceaccount istio mixer service account created serviceaccount istio pilot service account created serviceaccount prometheus created serviceaccount istio cleanup secrets service account created clusterrole rbac authorization io istio cleanup secrets istio system created clusterrolebinding rbac authorization io istio cleanup secrets istio system created job batch istio cleanup secrets created serviceaccount istio security post install account created clusterrole rbac authorization io istio security post install istio system created clusterrolebinding rbac authorization io istio security post install role binding istio system created job batch istio security post install created serviceaccount istio citadel service account created serviceaccount istio sidecar injector service account created serviceaccount istio multi created clusterrole rbac authorization io istio galley istio system created clusterrole rbac authorization io istio egressgateway istio system created clusterrole rbac authorization io istio ingressgateway istio system created clusterrole rbac authorization io kiali created clusterrole rbac authorization io kiali viewer created clusterrole rbac authorization io istio mixer istio system created clusterrole rbac authorization io istio pilot istio system created clusterrole rbac authorization io prometheus istio system created clusterrole rbac authorization io istio citadel istio system created clusterrole rbac authorization io istio sidecar injector istio system created clusterrole rbac authorization io istio reader created clusterrolebinding rbac authorization io istio galley admin role binding istio system created clusterrolebinding rbac authorization io istio egressgateway istio system created clusterrolebinding rbac authorization io istio ingressgateway istio system created clusterrolebinding rbac authorization io istio kiali admin role binding istio system created clusterrolebinding rbac authorization io istio mixer admin role binding istio system created clusterrolebinding rbac authorization io istio pilot istio system created clusterrolebinding rbac authorization io prometheus istio system created clusterrolebinding rbac authorization io istio citadel istio system created clusterrolebinding rbac authorization io istio sidecar injector admin role binding istio system created clusterrolebinding rbac authorization io istio multi created role rbac authorization io istio ingressgateway sds created rolebinding rbac authorization io istio ingressgateway sds created service istio galley created service istio egressgateway created service istio ingressgateway created service grafana created service kiali created service istio policy created service istio telemetry created service istio pilot created service prometheus created service istio citadel created service istio sidecar injector created deployment extensions istio galley created deployment extensions istio egressgateway created deployment extensions istio ingressgateway created deployment extensions grafana created deployment extensions kiali created deployment extensions istio policy created deployment extensions istio telemetry created deployment extensions istio pilot created deployment extensions prometheus created deployment extensions istio citadel created deployment extensions istio sidecar injector created deployment extensions istio tracing created horizontalpodautoscaler autoscaling istio egressgateway created horizontalpodautoscaler autoscaling istio ingressgateway created horizontalpodautoscaler autoscaling istio policy created horizontalpodautoscaler autoscaling istio telemetry created horizontalpodautoscaler autoscaling istio pilot created service jaeger query created service jaeger collector created service jaeger agent created service zipkin created service tracing created mutatingwebhookconfiguration admissionregistration io istio sidecar injector created poddisruptionbudget policy istio galley created poddisruptionbudget policy istio egressgateway created poddisruptionbudget policy istio ingressgateway created poddisruptionbudget policy istio policy created poddisruptionbudget policy istio telemetry created poddisruptionbudget policy istio pilot created attributemanifest config istio io istioproxy created attributemanifest config istio io kubernetes created handler config istio io stdio created logentry config istio io accesslog created logentry config istio io tcpaccesslog created rule config istio io stdio created rule config istio io stdiotcp created metric config istio io requestcount created metric config istio io requestduration created metric config istio io requestsize created metric config istio io responsesize created metric config istio io tcpbytesent created metric config istio io tcpbytereceived created metric config istio io tcpconnectionsopened created metric config istio io tcpconnectionsclosed created handler config istio io prometheus created rule config istio io promhttp created rule config istio io promtcp created rule config istio io promtcpconnectionopen created rule config istio io promtcpconnectionclosed created handler config istio io kubernetesenv created rule config istio io kubeattrgenrulerule created rule config istio io tcpkubeattrgenrulerule created kubernetes config istio io attributes created destinationrule networking istio io istio policy created destinationrule networking istio io istio telemetry created now using project kabanero on server you can add applications to this project with the new app command for example try oc new app django psql example to build a new example application in python or use kubectl to deploy a simple kubernetes application kubectl create deployment hello node image gcr io hello minikube zero install hello node namespace kabanero configured deployment apps knative eventing operator created role rbac authorization io knative eventing operator created clusterrole rbac authorization io knative eventing operator created rolebinding rbac authorization io knative eventing operator created clusterrolebinding rbac authorization io knative eventing operator created serviceaccount knative eventing operator created customresourcedefinition apiextensions io knativeeventings eventing knative dev created deployment apps knative serving operator created role rbac authorization io knative serving operator created clusterrole rbac authorization io knative serving operator created rolebinding rbac authorization io knative serving operator created clusterrolebinding rbac authorization io knative serving operator created serviceaccount knative serving operator created customresourcedefinition apiextensions io knativeservings serving knative dev created namespace knative sources created serviceaccount controller manager created clusterrole rbac authorization io eventing sources controller created clusterrolebinding rbac authorization io eventing sources controller created clusterrolebinding rbac authorization io eventing sources addressable resolver created customresourcedefinition apiextensions io githubsources sources eventing knative dev created service controller created statefulset apps controller manager created deployment apps openshift pipelines operator created clusterrole rbac authorization io openshift pipelines operator created clusterrolebinding rbac authorization io openshift pipelines operator created serviceaccount openshift pipelines operator created namespace openshift pipelines created customresourcedefinition apiextensions io config operator tekton dev created clusterrole rbac authorization io knative serving operator configured clusterrole rbac authorization io knative eventing operator configured customresourcedefinition apiextensions io kabaneros kabanero io created customresourcedefinition apiextensions io collections kabanero io created serviceaccount kabanero operator created deployment apps kabanero operator created clusterrole rbac authorization io kabanero operator created clusterrolebinding rbac authorization io kabanero operator created namespace kabanero unchanged deployment apps knative eventing operator configured role rbac authorization io knative eventing operator configured clusterrole rbac authorization io knative eventing operator configured rolebinding rbac authorization io knative eventing operator unchanged clusterrolebinding rbac authorization io knative eventing operator unchanged serviceaccount knative eventing operator unchanged customresourcedefinition apiextensions io knativeeventings eventing knative dev unchanged deployment apps knative serving operator configured role rbac authorization io knative serving operator configured clusterrole rbac authorization io knative serving operator configured rolebinding rbac authorization io knative serving operator unchanged clusterrolebinding rbac authorization io knative serving operator unchanged serviceaccount knative serving operator unchanged customresourcedefinition apiextensions io knativeservings serving knative dev unchanged namespace knative sources unchanged serviceaccount controller manager unchanged clusterrole rbac authorization io eventing sources controller unchanged clusterrolebinding rbac authorization io eventing sources controller unchanged clusterrolebinding rbac authorization io eventing sources addressable resolver unchanged customresourcedefinition apiextensions io githubsources sources eventing knative dev unchanged service controller unchanged statefulset apps controller manager configured deployment apps openshift pipelines operator configured clusterrole rbac authorization io openshift pipelines operator configured clusterrolebinding rbac authorization io openshift pipelines operator unchanged serviceaccount openshift pipelines operator unchanged namespace openshift pipelines unchanged customresourcedefinition apiextensions io config operator tekton dev unchanged config operator tekton dev cluster created clusterrole rbac authorization io knative serving operator configured clusterrole rbac authorization io knative eventing operator configured customresourcedefinition apiextensions io kabaneros kabanero io configured customresourcedefinition apiextensions io collections kabanero io configured serviceaccount kabanero operator unchanged deployment apps kabanero operator configured clusterrole rbac authorization io kabanero operator configured clusterrolebinding rbac authorization io kabanero operator unchanged clusterrole rbac authorization io cluster admin added kabanero operator name created at services serving knative dev name created at clustertasks tekton dev config operator tekton dev pipelineresources tekton dev pipelineruns tekton dev pipelines tekton dev taskruns tekton dev tasks tekton dev serviceaccount tekton webhooks extension created clusterrole rbac authorization io tekton webhooks extension minimal created clusterrolebinding rbac authorization io tekton webhooks extension minimal created deployment apps webhooks extension created service webhooks extension created service serving knative dev webhooks extension sink created task tekton dev monitor result task created serviceaccount tekton dashboard created customresourcedefinition apiextensions io extensions dashboard tekton dev created clusterrole rbac authorization io tekton dashboard minimal created clusterrolebinding rbac authorization io tekton dashboard minimal created deployment apps tekton dashboard created route route openshift io tekton dashboard created service tekton dashboard created pipeline tekton dev created task tekton dev task created name age tekton dashboard minimal clusterrole rbac authorization io tekton dashboard minimal patched deployment extensions tekton dashboard scaled deployment extensions tekton dashboard scaled configmap config domain patched name created at extensions dashboard tekton dev the installation script is complete you can now create an instance of the kabanero cr if you have cloned and curated a collection set apply the kabanero cr that you created or to create the default instance oc apply n kabanero f task thursday october ok msg task thursday october skipping task thursday october changed task thursday october ok msg statefulset apps alertmanager main scaled statefulset apps prometheus scaled deployment extensions cluster monitoring operator scaled deployment extensions grafana scaled deployment extensions kube state metrics scaled deployment extensions openshift state metrics scaled deployment extensions prometheus adapter scaled deployment extensions prometheus operator scaled deployment extensions telemeter client scaled task thursday october changed task thursday october ok msg client version server version kubernetes version play recap svtcrc fyre ibm com ok changed unreachable failed skipped rescued ignored thursday october install kabanero foundation crc start install networkmanager firefox crc setup retrieve kabanero foundation master zip gathering facts prereq directories oc login enable monitoring alerting and telemetry services enable and start networkmanager crc pull secret unzip kabanerio foundation master zip oc version check if crc is installed whoami oc in user path password masking process crc creds check if crc is created link oc kubectl set admin passwd install kabanero foundation stdout set fact crc setup stdout crc start stdout crc creds stdout set fact oc login stdout enable monitoring alerting and telemetry services stdout set fact fail crc delete crcstop stdout install kabanero foundation errors install kabanero foundation return code crc stop set fact set fact set fact crc delete stdout whoami stdout oc version stdout kabanero crc complete runtime minutes setup complete overall runtime minutes
1
270,534
23,516,313,901
IssuesEvent
2022-08-18 21:50:01
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
MonoTests.System.Runtime.Caching.MemoryCacheTestExpires4.TestCacheShrink test failure
area-System.Runtime.Caching test-bug
Configuration: `net6.0-windows-Release-arm64-CoreCLR_release-Windows.10.Arm64.Open` ``` MonoTests.System.Runtime.Caching.MemoryCacheTestExpires4.TestCacheShrink [FAIL] Assert.Null() Failure Expected: (null) Actual: 0 Stack Trace: /_/src/libraries/System.Runtime.Caching/tests/System.Runtime.Caching/MemoryCacheTest.cs(0,0): at MonoTests.System.Runtime.Caching.MemoryCacheTestExpires4.TestCacheShrink() --- End of stack trace from previous location --- ``` Console: https://helixre8s23ayyeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-f2e3e082ea3b44a49c/System.Runtime.Caching.Tests/1/console.f9d74dab.log?sv=2019-07-07&se=2021-09-10T12%3A50%3A10Z&sr=c&sp=rl&sig=yohDg5MjQ17E738GWrsKtd5oBkc5WxKT8vvsgpAsogE%3D Build: https://dev.azure.com/dnceng/public/_build/results?buildId=1309902&view=ms.vss-test-web.build-test-results-tab&runId=38562306&paneView=debug
1.0
MonoTests.System.Runtime.Caching.MemoryCacheTestExpires4.TestCacheShrink test failure - Configuration: `net6.0-windows-Release-arm64-CoreCLR_release-Windows.10.Arm64.Open` ``` MonoTests.System.Runtime.Caching.MemoryCacheTestExpires4.TestCacheShrink [FAIL] Assert.Null() Failure Expected: (null) Actual: 0 Stack Trace: /_/src/libraries/System.Runtime.Caching/tests/System.Runtime.Caching/MemoryCacheTest.cs(0,0): at MonoTests.System.Runtime.Caching.MemoryCacheTestExpires4.TestCacheShrink() --- End of stack trace from previous location --- ``` Console: https://helixre8s23ayyeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-f2e3e082ea3b44a49c/System.Runtime.Caching.Tests/1/console.f9d74dab.log?sv=2019-07-07&se=2021-09-10T12%3A50%3A10Z&sr=c&sp=rl&sig=yohDg5MjQ17E738GWrsKtd5oBkc5WxKT8vvsgpAsogE%3D Build: https://dev.azure.com/dnceng/public/_build/results?buildId=1309902&view=ms.vss-test-web.build-test-results-tab&runId=38562306&paneView=debug
test
monotests system runtime caching testcacheshrink test failure configuration windows release coreclr release windows open monotests system runtime caching testcacheshrink assert null failure expected null actual stack trace src libraries system runtime caching tests system runtime caching memorycachetest cs at monotests system runtime caching testcacheshrink end of stack trace from previous location console build
1
809,037
30,121,835,043
IssuesEvent
2023-06-30 15:46:56
conaticus/FileExplorer
https://api.github.com/repos/conaticus/FileExplorer
opened
Recaching interval not long enough on slower systems
bug medium priority
The interval works well on most systems, but slower systems or systems with a bigger file system have the chance of not being able to recache in time for this 30s interval. The best solution is to have a check to ensure that the previous recache is completed, and skip if it is not. We could implement an algorithm to check for file changes first, but these usually happen so often that it would be redundant.
1.0
Recaching interval not long enough on slower systems - The interval works well on most systems, but slower systems or systems with a bigger file system have the chance of not being able to recache in time for this 30s interval. The best solution is to have a check to ensure that the previous recache is completed, and skip if it is not. We could implement an algorithm to check for file changes first, but these usually happen so often that it would be redundant.
non_test
recaching interval not long enough on slower systems the interval works well on most systems but slower systems or systems with a bigger file system have the chance of not being able to recache in time for this interval the best solution is to have a check to ensure that the previous recache is completed and skip if it is not we could implement an algorithm to check for file changes first but these usually happen so often that it would be redundant
0
119,786
10,064,413,621
IssuesEvent
2019-07-23 08:37:02
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
The created new table doesn't show on the tree view for one Table Account
:gear: cosmosdb 🧪 testing
**Storage Explorer Version:** 1.9.0_**master**_20190709.5 **Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10 **Architecture:** ia32/x64 **Commit:** 8a04af7f **Regression From:** Not a regression **Steps to reproduce:** 1. Create a new **Table** account then navigate to it from Storage Explorer. 2. Right click the create Table account then select 'Create Table...'. 3. Type a valid table id then click 'OK'. 4. Check whether the created table shows well or not on the tree view. **Expect Experience:** The created table shows well on the tree view. **Actual Experience:** The created table doesn't show on the tree view. **More info:** The created table shows well after refreshing the Table account node.
1.0
The created new table doesn't show on the tree view for one Table Account - **Storage Explorer Version:** 1.9.0_**master**_20190709.5 **Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10 **Architecture:** ia32/x64 **Commit:** 8a04af7f **Regression From:** Not a regression **Steps to reproduce:** 1. Create a new **Table** account then navigate to it from Storage Explorer. 2. Right click the create Table account then select 'Create Table...'. 3. Type a valid table id then click 'OK'. 4. Check whether the created table shows well or not on the tree view. **Expect Experience:** The created table shows well on the tree view. **Actual Experience:** The created table doesn't show on the tree view. **More info:** The created table shows well after refreshing the Table account node.
test
the created new table doesn t show on the tree view for one table account storage explorer version master platform os linux ubuntu macos high sierra windows architecture commit regression from not a regression steps to reproduce create a new table account then navigate to it from storage explorer right click the create table account then select create table type a valid table id then click ok check whether the created table shows well or not on the tree view expect experience the created table shows well on the tree view actual experience the created table doesn t show on the tree view more info the created table shows well after refreshing the table account node
1
49,003
5,996,795,874
IssuesEvent
2017-06-03 17:33:57
ppazos/cabolabs-ehrserver
https://api.github.com/repos/ppazos/cabolabs-ehrserver
opened
Test spring security protection against brute force attacks on the login
mid_prio test
1. check if there is any built-in solution, like IP blocking. 2. check if invalid logins for a valid username generate DOS for that user. 3. check the two step login like the new google login to avoid attacks (analyze pros/cons and if that applies for this context).
1.0
Test spring security protection against brute force attacks on the login - 1. check if there is any built-in solution, like IP blocking. 2. check if invalid logins for a valid username generate DOS for that user. 3. check the two step login like the new google login to avoid attacks (analyze pros/cons and if that applies for this context).
test
test spring security protection against brute force attacks on the login check if there is any built in solution like ip blocking check if invalid logins for a valid username generate dos for that user check the two step login like the new google login to avoid attacks analyze pros cons and if that applies for this context
1
389
2,535,799,006
IssuesEvent
2015-01-26 07:46:45
localwiki/localwiki
https://api.github.com/repos/localwiki/localwiki
closed
First-time usability confusion points
defect general
Double click to end path when drawing -- was confusing. Wasn't obvious. ----- Deleting things in the map -- delete button on kb should delete an area or line or point. deleting is just kinda weird. ------ Undo button in editor?
1.0
First-time usability confusion points - Double click to end path when drawing -- was confusing. Wasn't obvious. ----- Deleting things in the map -- delete button on kb should delete an area or line or point. deleting is just kinda weird. ------ Undo button in editor?
non_test
first time usability confusion points double click to end path when drawing was confusing wasn t obvious deleting things in the map delete button on kb should delete an area or line or point deleting is just kinda weird undo button in editor
0
319,222
9,740,469,361
IssuesEvent
2019-06-01 20:48:54
Terrastories/terrastories
https://api.github.com/repos/Terrastories/terrastories
closed
[Video] Opening a video in fullscreen crashes the display
difficulty: intermediate priority: medium status: help wanted type: bug type: video
To reproduce: 1. Find a story with a video 2. Play the video 3. Click the fullscreen icon on the video player 4. Notice that the fullscreen display crashes after ~half a second A good amount of discussion took place on this subject at Ruby by the Bay 2019: https://github.com/rubyforgood/terrastories/issues/24
1.0
[Video] Opening a video in fullscreen crashes the display - To reproduce: 1. Find a story with a video 2. Play the video 3. Click the fullscreen icon on the video player 4. Notice that the fullscreen display crashes after ~half a second A good amount of discussion took place on this subject at Ruby by the Bay 2019: https://github.com/rubyforgood/terrastories/issues/24
non_test
opening a video in fullscreen crashes the display to reproduce find a story with a video play the video click the fullscreen icon on the video player notice that the fullscreen display crashes after half a second a good amount of discussion took place on this subject at ruby by the bay
0
156,210
12,301,091,581
IssuesEvent
2020-05-11 14:55:13
jorandmol/DP2-G7
https://api.github.com/repos/jorandmol/DP2-G7
closed
T18 - Actualizar Medicine
develop unit testing
Como auxiliar de clínica quiero poder editar los datos de un medicamento para solucionar posibles errores o actualizar dicho objeto.
1.0
T18 - Actualizar Medicine - Como auxiliar de clínica quiero poder editar los datos de un medicamento para solucionar posibles errores o actualizar dicho objeto.
test
actualizar medicine como auxiliar de clínica quiero poder editar los datos de un medicamento para solucionar posibles errores o actualizar dicho objeto
1
275,261
23,901,232,792
IssuesEvent
2022-09-08 18:59:01
PharmaLedger-IMI/eco-iot-pmed-workspace
https://api.github.com/repos/PharmaLedger-IMI/eco-iot-pmed-workspace
closed
[Clinical Site] Check the Notification "view questions"
business-testing
Questionnaire Notification LINK “view questions” to HCP does not show the results of the questions
1.0
[Clinical Site] Check the Notification "view questions" - Questionnaire Notification LINK “view questions” to HCP does not show the results of the questions
test
check the notification view questions questionnaire notification link “view questions” to hcp does not show the results of the questions
1
106,207
23,193,281,440
IssuesEvent
2022-08-01 14:20:06
open-contracting/standard
https://api.github.com/repos/open-contracting/standard
closed
Fix markup to avoid unwanted links
Schema Codelist: Open quick
linkify automatically hyperlinks URLs. We need to: * Add backticks around release.id in the schema * Remove the sentence about buyandsell.gc.ca from itemClassificationScheme.csv (no other code has a similar sentence) Once that's done, we can remove the lines in `conf.py` that ignore the auto-created links.
1.0
Fix markup to avoid unwanted links - linkify automatically hyperlinks URLs. We need to: * Add backticks around release.id in the schema * Remove the sentence about buyandsell.gc.ca from itemClassificationScheme.csv (no other code has a similar sentence) Once that's done, we can remove the lines in `conf.py` that ignore the auto-created links.
non_test
fix markup to avoid unwanted links linkify automatically hyperlinks urls we need to add backticks around release id in the schema remove the sentence about buyandsell gc ca from itemclassificationscheme csv no other code has a similar sentence once that s done we can remove the lines in conf py that ignore the auto created links
0
405,773
27,533,225,863
IssuesEvent
2023-03-07 00:18:09
sveltejs/kit
https://api.github.com/repos/sveltejs/kit
closed
Prerender tries to render outgoing link
documentation p3-edge-case
### Describe the bug First off, sorry if this is the wrong place, I am fairly new to all the things going on here. I have been trying to create a blog which uses SvelteKit, mdsvex and a static adapter. Everything works smooth until I start adding links inside the markdown files. If any .md file has any outgoing link, it seems the prerender tries to find those sites on my site. This causes it to fail ofcourse. The issue can be resolved by setting `onerror` for the prerenderer into `continue` but I would like to avoid that since now any missing posts will also be continuned. ### Reproduction Clone the repository and try running the build command ```bash git clone git@github.com:percybolmer/pp-blog-svelte.git cd pp-blog-svelte git checkout bug/prerender-outgoinglink npm install npm run build ``` ### Logs ```Shell percy@pc038:~/private/blog/pp-blog-svelte$ npm run build > build > vite build vite v3.1.8 building for production... ✓ 75 modules transformed. vite v3.1.8 building SSR bundle for production... ✓ 81 modules transformed. Generated an empty chunk: "hooks" .svelte-kit/output/server/vite-manifest.json 4.99 KiB .svelte-kit/output/server/index.js 69.60 KiB .svelte-kit/output/server/entries/endpoints/rss/_server.js 1.42 KiB .svelte-kit/output/server/entries/endpoints/api/posts/_server.js 0.43 KiB .svelte-kit/output/server/entries/pages/_layout.svelte.js 1.58 KiB .svelte-kit/output/server/entries/pages/_layout.js 0.15 KiB .svelte-kit/output/server/entries/fallbacks/error.svelte.js 1.50 KiB .svelte-kit/output/server/entries/pages/_page.svelte.js 0.65 KiB .svelte-kit/output/server/entries/pages/about/_page.svelte.js 0.38 KiB .svelte-kit/output/server/entries/pages/blog/_page.svelte.js 2.57 KiB .svelte-kit/output/server/entries/pages/blog/_page.js 0.16 KiB .svelte-kit/output/server/entries/pages/blog/_slug_/_page.svelte.js 0.81 KiB .svelte-kit/output/server/entries/pages/blog/_slug_/_page.js 0.87 KiB .svelte-kit/output/server/entries/pages/contact/_page.svelte.js 0.38 KiB .svelte-kit/output/server/entries/pages/uses/_page.md.js 0.45 KiB .svelte-kit/output/server/chunks/paths.js 0.17 KiB .svelte-kit/output/server/chunks/index.js 3.53 KiB .svelte-kit/output/server/chunks/index2.js 0.97 KiB .svelte-kit/output/server/chunks/index3.js 0.79 KiB .svelte-kit/output/server/chunks/hooks.js 0.00 KiB .svelte-kit/output/server/chunks/2.js 0.91 KiB .svelte-kit/output/server/chunks/thumbnail.js 0.12 KiB .svelte-kit/output/server/chunks/preview-ide-of-the-future.js 23.67 KiB .svelte-kit/output/server/chunks/test.js 0.54 KiB Error: Unknown variable dynamic import: ../posts/www.jetbrains.com.md at file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js:7:92 at new Promise (<anonymous>) at __variableDynamicImportRuntimeHelper (file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js:6:10) at load (file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js:11:22) at load_data (file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/index.js:626:39) at async file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/index.js:1416:18 500 /pp-blog-svelte/blog/www.jetbrains.com (linked from /pp-blog-svelte/blog/preview-ide-of-the-future) ``` ### System Info ```Shell System: OS: Linux 5.15 Ubuntu 20.04.5 LTS (Focal Fossa) CPU: (16) x64 11th Gen Intel(R) Core(TM) i7-11800H @ 2.30GHz Memory: 1.99 GB / 15.36 GB Container: Yes Shell: 5.0.17 - /bin/bash Binaries: Node: 17.9.1 - ~/.nvm/versions/node/v17.9.1/bin/node Yarn: 1.22.19 - ~/.nvm/versions/node/v14.17.6/bin/yarn npm: 8.11.0 - ~/.nvm/versions/node/v17.9.1/bin/npm Browsers: Brave Browser: 106.1.44.101 Chrome: 106.0.5249.61 Firefox: 105.0 npmPackages: @sveltejs/adapter-static: ^1.0.0-next.46 => 1.0.0-next.46 @sveltejs/kit: 1.0.0-next.516 => 1.0.0-next.516 svelte: ^3.52.0 => 3.52.0 vite: ^3.1.8 => 3.1.8 ``` ### Severity annoyance ### Additional Information Hopefully it is not just me being dumb!
1.0
Prerender tries to render outgoing link - ### Describe the bug First off, sorry if this is the wrong place, I am fairly new to all the things going on here. I have been trying to create a blog which uses SvelteKit, mdsvex and a static adapter. Everything works smooth until I start adding links inside the markdown files. If any .md file has any outgoing link, it seems the prerender tries to find those sites on my site. This causes it to fail ofcourse. The issue can be resolved by setting `onerror` for the prerenderer into `continue` but I would like to avoid that since now any missing posts will also be continuned. ### Reproduction Clone the repository and try running the build command ```bash git clone git@github.com:percybolmer/pp-blog-svelte.git cd pp-blog-svelte git checkout bug/prerender-outgoinglink npm install npm run build ``` ### Logs ```Shell percy@pc038:~/private/blog/pp-blog-svelte$ npm run build > build > vite build vite v3.1.8 building for production... ✓ 75 modules transformed. vite v3.1.8 building SSR bundle for production... ✓ 81 modules transformed. Generated an empty chunk: "hooks" .svelte-kit/output/server/vite-manifest.json 4.99 KiB .svelte-kit/output/server/index.js 69.60 KiB .svelte-kit/output/server/entries/endpoints/rss/_server.js 1.42 KiB .svelte-kit/output/server/entries/endpoints/api/posts/_server.js 0.43 KiB .svelte-kit/output/server/entries/pages/_layout.svelte.js 1.58 KiB .svelte-kit/output/server/entries/pages/_layout.js 0.15 KiB .svelte-kit/output/server/entries/fallbacks/error.svelte.js 1.50 KiB .svelte-kit/output/server/entries/pages/_page.svelte.js 0.65 KiB .svelte-kit/output/server/entries/pages/about/_page.svelte.js 0.38 KiB .svelte-kit/output/server/entries/pages/blog/_page.svelte.js 2.57 KiB .svelte-kit/output/server/entries/pages/blog/_page.js 0.16 KiB .svelte-kit/output/server/entries/pages/blog/_slug_/_page.svelte.js 0.81 KiB .svelte-kit/output/server/entries/pages/blog/_slug_/_page.js 0.87 KiB .svelte-kit/output/server/entries/pages/contact/_page.svelte.js 0.38 KiB .svelte-kit/output/server/entries/pages/uses/_page.md.js 0.45 KiB .svelte-kit/output/server/chunks/paths.js 0.17 KiB .svelte-kit/output/server/chunks/index.js 3.53 KiB .svelte-kit/output/server/chunks/index2.js 0.97 KiB .svelte-kit/output/server/chunks/index3.js 0.79 KiB .svelte-kit/output/server/chunks/hooks.js 0.00 KiB .svelte-kit/output/server/chunks/2.js 0.91 KiB .svelte-kit/output/server/chunks/thumbnail.js 0.12 KiB .svelte-kit/output/server/chunks/preview-ide-of-the-future.js 23.67 KiB .svelte-kit/output/server/chunks/test.js 0.54 KiB Error: Unknown variable dynamic import: ../posts/www.jetbrains.com.md at file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js:7:92 at new Promise (<anonymous>) at __variableDynamicImportRuntimeHelper (file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js:6:10) at load (file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js:11:22) at load_data (file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/index.js:626:39) at async file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/index.js:1416:18 500 /pp-blog-svelte/blog/www.jetbrains.com (linked from /pp-blog-svelte/blog/preview-ide-of-the-future) ``` ### System Info ```Shell System: OS: Linux 5.15 Ubuntu 20.04.5 LTS (Focal Fossa) CPU: (16) x64 11th Gen Intel(R) Core(TM) i7-11800H @ 2.30GHz Memory: 1.99 GB / 15.36 GB Container: Yes Shell: 5.0.17 - /bin/bash Binaries: Node: 17.9.1 - ~/.nvm/versions/node/v17.9.1/bin/node Yarn: 1.22.19 - ~/.nvm/versions/node/v14.17.6/bin/yarn npm: 8.11.0 - ~/.nvm/versions/node/v17.9.1/bin/npm Browsers: Brave Browser: 106.1.44.101 Chrome: 106.0.5249.61 Firefox: 105.0 npmPackages: @sveltejs/adapter-static: ^1.0.0-next.46 => 1.0.0-next.46 @sveltejs/kit: 1.0.0-next.516 => 1.0.0-next.516 svelte: ^3.52.0 => 3.52.0 vite: ^3.1.8 => 3.1.8 ``` ### Severity annoyance ### Additional Information Hopefully it is not just me being dumb!
non_test
prerender tries to render outgoing link describe the bug first off sorry if this is the wrong place i am fairly new to all the things going on here i have been trying to create a blog which uses sveltekit mdsvex and a static adapter everything works smooth until i start adding links inside the markdown files if any md file has any outgoing link it seems the prerender tries to find those sites on my site this causes it to fail ofcourse the issue can be resolved by setting onerror for the prerenderer into continue but i would like to avoid that since now any missing posts will also be continuned reproduction clone the repository and try running the build command bash git clone git github com percybolmer pp blog svelte git cd pp blog svelte git checkout bug prerender outgoinglink npm install npm run build logs shell percy private blog pp blog svelte npm run build build vite build vite building for production ✓ modules transformed vite building ssr bundle for production ✓ modules transformed generated an empty chunk hooks svelte kit output server vite manifest json kib svelte kit output server index js kib svelte kit output server entries endpoints rss server js kib svelte kit output server entries endpoints api posts server js kib svelte kit output server entries pages layout svelte js kib svelte kit output server entries pages layout js kib svelte kit output server entries fallbacks error svelte js kib svelte kit output server entries pages page svelte js kib svelte kit output server entries pages about page svelte js kib svelte kit output server entries pages blog page svelte js kib svelte kit output server entries pages blog page js kib svelte kit output server entries pages blog slug page svelte js kib svelte kit output server entries pages blog slug page js kib svelte kit output server entries pages contact page svelte js kib svelte kit output server entries pages uses page md js kib svelte kit output server chunks paths js kib svelte kit output server chunks index js kib svelte kit output server chunks js kib svelte kit output server chunks js kib svelte kit output server chunks hooks js kib svelte kit output server chunks js kib svelte kit output server chunks thumbnail js kib svelte kit output server chunks preview ide of the future js kib svelte kit output server chunks test js kib error unknown variable dynamic import posts at file home percy private blog pp blog svelte svelte kit output server entries pages blog slug page js at new promise at variabledynamicimportruntimehelper file home percy private blog pp blog svelte svelte kit output server entries pages blog slug page js at load file home percy private blog pp blog svelte svelte kit output server entries pages blog slug page js at load data file home percy private blog pp blog svelte svelte kit output server index js at async file home percy private blog pp blog svelte svelte kit output server index js pp blog svelte blog linked from pp blog svelte blog preview ide of the future system info shell system os linux ubuntu lts focal fossa cpu gen intel r core tm memory gb gb container yes shell bin bash binaries node nvm versions node bin node yarn nvm versions node bin yarn npm nvm versions node bin npm browsers brave browser chrome firefox npmpackages sveltejs adapter static next next sveltejs kit next next svelte vite severity annoyance additional information hopefully it is not just me being dumb
0
59,032
14,524,957,564
IssuesEvent
2020-12-14 12:12:41
neovim/neovim
https://api.github.com/repos/neovim/neovim
closed
`brew install --HEAD` failing with errors
bug build duplicate platform:macos
<!-- Before reporting: search existing issues and check the FAQ. --> - `nvim --version`: NVIM v0.5.0-dev+944-g82100a6bd - `vim -u DEFAULTS` (version: ) behaves differently? no - Operating system/version: macOS 11.0.1 (on MacBook Air with Apple Silicon M1 SoC) - Terminal name/version: iTerm2 - `$TERM`: xterm-256color ### Actual behaviour Note: I've already found #8809 and followed the build-from-source prerequisite instructions. ``` ❯ brew install --HEAD neovim Updating Homebrew... ==> Downloading https://github.com/libmpack/libmpack-lua/releases/download/1.0.8 Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/4b91349796a66e38e93acd77e6ca14212c869090429fd88271def2be5293d83f--libmpack-lua-1.0.8.tar.gz ==> Downloading https://luarocks.org/manifests/gvvaughan/lpeg-1.0.2-1.src.rock Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/0bc17975f0cf3981083172aa7da329b874e5780fa81494f5dd07d0665fcec23b--lpeg-1.0.2-1.src.rock ==> Downloading https://luarocks.org/manifests/kikito/inspect-3.1.1-0.src.rock Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/ecc1fa84ce4c83c9406f6d1202d4a88c80ede4a0a2bfd8d8d2bd54fd042226bb--inspect-3.1.1-0.src.rock ==> Downloading https://github.com/keplerproject/lua-compat-5.3/archive/v0.7.tar Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/5f4e552e95b9a6223a5e4fabb7f04bf4b886057024507c76b1d5b803218c9363--lua-compat-5.3-0.7.tar.gz ==> Downloading https://github.com/luvit/luv/releases/download/1.30.0-0/luv-1.30 Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/a9ce5a949370965d8f0fe70a47e7d102a3b4b6b831dbd4d088a83bbe1c00a3a2--luv-1.30.0-0.tar.gz ==> Cloning https://github.com/neovim/neovim.git Updating /Users/mpcsh/Library/Caches/Homebrew/neovim--git ==> Checking out branch master Already on 'master' Your branch is up to date with 'origin/master'. HEAD is now at 82100a6bd healthcheck: fix health check issue with shada file (#13291) ==> luarocks make --lua-dir=/usr/local/opt/luajit --tree=/private/tmp/neovim-202 ==> luarocks make --lua-dir=/usr/local/opt/luajit --tree=/private/tmp/neovim-202 ==> luarocks make --lua-dir=/usr/local/opt/luajit --tree=/private/tmp/neovim-202 ==> cmake . -DCMAKE_C_COMPILER=/usr/bin/clang -DCMAKE_CXX_COMPILER=/usr/bin/clan ==> make install ==> cmake .. -DCMAKE_C_COMPILER=/usr/bin/clang -DCMAKE_CXX_COMPILER=/usr/bin/cla ==> make install Last 15 lines from /Users/mpcsh/Library/Logs/Homebrew/neovim/07.make: ^ /tmp/neovim-20201214-42949-1h90lp3/src/nvim/os/lang.c:63:15: error: use of undeclared identifier 'LC_ALL'; did you mean 'P_ALL'? setlocale(LC_ALL, ""); ^~~~~~ P_ALL /Library/Developer/CommandLineTools/SDKs/MacOSX11.0.sdk/usr/include/sys/wait.h:80:2: note: 'P_ALL' declared here P_ALL, ^ 2 errors generated. [ 50%] Building C object src/nvim/CMakeFiles/nvim.dir/os/time.c.o cd /tmp/neovim-20201214-42949-1h90lp3/build/src/nvim && /usr/bin/clang -DINCLUDE_GENERATED_DECLARATIONS -DMIN_LOG_LEVEL=3 -DNVIM_MSGPACK_HAS_FLOAT32 -D_GNU_SOURCE -Dnvim_EXPORTS -I/tmp/neovim-20201214-42949-1h90lp3/build/config -I/tmp/neovim-20201214-42949-1h90lp3/src -I/tmp/neovim-20201214-42949-1h90lp3/build/src/nvim/auto -I/tmp/neovim-20201214-42949-1h90lp3/build/include -I/usr/local/include/luajit-2.0 -isystem /usr/local/include -isystem /tmp/neovim-20201214-42949-1h90lp3/deps-build/include -isystem /usr/local/opt/gettext/include -U_FORTIFY_SOURCE -D_FORTIFY_SOURCE=1 -DNDEBUG -flto=thin -arch arm64 -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX11.0.sdk -Wall -Wextra -pedantic -Wno-unused-parameter -Wstrict-prototypes -std=gnu99 -Wshadow -Wconversion -Wmissing-prototypes -Wimplicit-fallthrough -Wvla -fstack-protector-strong -fno-common -fdiagnostics-color=auto -o CMakeFiles/nvim.dir/os/time.c.o -c /tmp/neovim-20201214-42949-1h90lp3/src/nvim/os/time.c make[2]: *** [src/nvim/CMakeFiles/nvim.dir/os/lang.c.o] Error 1 make[2]: *** Waiting for unfinished jobs.... make[1]: *** [src/nvim/CMakeFiles/nvim.dir/all] Error 2 make: *** [all] Error 2 READ THIS: https://docs.brew.sh/Troubleshooting Please create pull requests instead of asking for help on Homebrew's GitHub, Twitter or any other official channels. ``` ### Expected behaviour `brew install --HEAD neovim` works
1.0
`brew install --HEAD` failing with errors - <!-- Before reporting: search existing issues and check the FAQ. --> - `nvim --version`: NVIM v0.5.0-dev+944-g82100a6bd - `vim -u DEFAULTS` (version: ) behaves differently? no - Operating system/version: macOS 11.0.1 (on MacBook Air with Apple Silicon M1 SoC) - Terminal name/version: iTerm2 - `$TERM`: xterm-256color ### Actual behaviour Note: I've already found #8809 and followed the build-from-source prerequisite instructions. ``` ❯ brew install --HEAD neovim Updating Homebrew... ==> Downloading https://github.com/libmpack/libmpack-lua/releases/download/1.0.8 Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/4b91349796a66e38e93acd77e6ca14212c869090429fd88271def2be5293d83f--libmpack-lua-1.0.8.tar.gz ==> Downloading https://luarocks.org/manifests/gvvaughan/lpeg-1.0.2-1.src.rock Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/0bc17975f0cf3981083172aa7da329b874e5780fa81494f5dd07d0665fcec23b--lpeg-1.0.2-1.src.rock ==> Downloading https://luarocks.org/manifests/kikito/inspect-3.1.1-0.src.rock Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/ecc1fa84ce4c83c9406f6d1202d4a88c80ede4a0a2bfd8d8d2bd54fd042226bb--inspect-3.1.1-0.src.rock ==> Downloading https://github.com/keplerproject/lua-compat-5.3/archive/v0.7.tar Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/5f4e552e95b9a6223a5e4fabb7f04bf4b886057024507c76b1d5b803218c9363--lua-compat-5.3-0.7.tar.gz ==> Downloading https://github.com/luvit/luv/releases/download/1.30.0-0/luv-1.30 Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/a9ce5a949370965d8f0fe70a47e7d102a3b4b6b831dbd4d088a83bbe1c00a3a2--luv-1.30.0-0.tar.gz ==> Cloning https://github.com/neovim/neovim.git Updating /Users/mpcsh/Library/Caches/Homebrew/neovim--git ==> Checking out branch master Already on 'master' Your branch is up to date with 'origin/master'. HEAD is now at 82100a6bd healthcheck: fix health check issue with shada file (#13291) ==> luarocks make --lua-dir=/usr/local/opt/luajit --tree=/private/tmp/neovim-202 ==> luarocks make --lua-dir=/usr/local/opt/luajit --tree=/private/tmp/neovim-202 ==> luarocks make --lua-dir=/usr/local/opt/luajit --tree=/private/tmp/neovim-202 ==> cmake . -DCMAKE_C_COMPILER=/usr/bin/clang -DCMAKE_CXX_COMPILER=/usr/bin/clan ==> make install ==> cmake .. -DCMAKE_C_COMPILER=/usr/bin/clang -DCMAKE_CXX_COMPILER=/usr/bin/cla ==> make install Last 15 lines from /Users/mpcsh/Library/Logs/Homebrew/neovim/07.make: ^ /tmp/neovim-20201214-42949-1h90lp3/src/nvim/os/lang.c:63:15: error: use of undeclared identifier 'LC_ALL'; did you mean 'P_ALL'? setlocale(LC_ALL, ""); ^~~~~~ P_ALL /Library/Developer/CommandLineTools/SDKs/MacOSX11.0.sdk/usr/include/sys/wait.h:80:2: note: 'P_ALL' declared here P_ALL, ^ 2 errors generated. [ 50%] Building C object src/nvim/CMakeFiles/nvim.dir/os/time.c.o cd /tmp/neovim-20201214-42949-1h90lp3/build/src/nvim && /usr/bin/clang -DINCLUDE_GENERATED_DECLARATIONS -DMIN_LOG_LEVEL=3 -DNVIM_MSGPACK_HAS_FLOAT32 -D_GNU_SOURCE -Dnvim_EXPORTS -I/tmp/neovim-20201214-42949-1h90lp3/build/config -I/tmp/neovim-20201214-42949-1h90lp3/src -I/tmp/neovim-20201214-42949-1h90lp3/build/src/nvim/auto -I/tmp/neovim-20201214-42949-1h90lp3/build/include -I/usr/local/include/luajit-2.0 -isystem /usr/local/include -isystem /tmp/neovim-20201214-42949-1h90lp3/deps-build/include -isystem /usr/local/opt/gettext/include -U_FORTIFY_SOURCE -D_FORTIFY_SOURCE=1 -DNDEBUG -flto=thin -arch arm64 -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX11.0.sdk -Wall -Wextra -pedantic -Wno-unused-parameter -Wstrict-prototypes -std=gnu99 -Wshadow -Wconversion -Wmissing-prototypes -Wimplicit-fallthrough -Wvla -fstack-protector-strong -fno-common -fdiagnostics-color=auto -o CMakeFiles/nvim.dir/os/time.c.o -c /tmp/neovim-20201214-42949-1h90lp3/src/nvim/os/time.c make[2]: *** [src/nvim/CMakeFiles/nvim.dir/os/lang.c.o] Error 1 make[2]: *** Waiting for unfinished jobs.... make[1]: *** [src/nvim/CMakeFiles/nvim.dir/all] Error 2 make: *** [all] Error 2 READ THIS: https://docs.brew.sh/Troubleshooting Please create pull requests instead of asking for help on Homebrew's GitHub, Twitter or any other official channels. ``` ### Expected behaviour `brew install --HEAD neovim` works
non_test
brew install head failing with errors nvim version nvim dev vim u defaults version behaves differently no operating system version macos on macbook air with apple silicon soc terminal name version term xterm actual behaviour note i ve already found and followed the build from source prerequisite instructions ❯ brew install head neovim updating homebrew downloading already downloaded users mpcsh library caches homebrew downloads libmpack lua tar gz downloading already downloaded users mpcsh library caches homebrew downloads lpeg src rock downloading already downloaded users mpcsh library caches homebrew downloads inspect src rock downloading already downloaded users mpcsh library caches homebrew downloads lua compat tar gz downloading already downloaded users mpcsh library caches homebrew downloads luv tar gz cloning updating users mpcsh library caches homebrew neovim git checking out branch master already on master your branch is up to date with origin master head is now at healthcheck fix health check issue with shada file luarocks make lua dir usr local opt luajit tree private tmp neovim luarocks make lua dir usr local opt luajit tree private tmp neovim luarocks make lua dir usr local opt luajit tree private tmp neovim cmake dcmake c compiler usr bin clang dcmake cxx compiler usr bin clan make install cmake dcmake c compiler usr bin clang dcmake cxx compiler usr bin cla make install last lines from users mpcsh library logs homebrew neovim make tmp neovim src nvim os lang c error use of undeclared identifier lc all did you mean p all setlocale lc all p all library developer commandlinetools sdks sdk usr include sys wait h note p all declared here p all errors generated building c object src nvim cmakefiles nvim dir os time c o cd tmp neovim build src nvim usr bin clang dinclude generated declarations dmin log level dnvim msgpack has d gnu source dnvim exports i tmp neovim build config i tmp neovim src i tmp neovim build src nvim auto i tmp neovim build include i usr local include luajit isystem usr local include isystem tmp neovim deps build include isystem usr local opt gettext include u fortify source d fortify source dndebug flto thin arch isysroot library developer commandlinetools sdks sdk wall wextra pedantic wno unused parameter wstrict prototypes std wshadow wconversion wmissing prototypes wimplicit fallthrough wvla fstack protector strong fno common fdiagnostics color auto o cmakefiles nvim dir os time c o c tmp neovim src nvim os time c make error make waiting for unfinished jobs make error make error read this please create pull requests instead of asking for help on homebrew s github twitter or any other official channels expected behaviour brew install head neovim works
0
781,312
27,432,310,653
IssuesEvent
2023-03-02 03:01:14
AY2223S2-CS2103-F11-3/tp
https://api.github.com/repos/AY2223S2-CS2103-F11-3/tp
opened
Update command syntax - argument parsing
priority.High type.Task
Argument flags are delimited by a common `--` delimiter.
1.0
Update command syntax - argument parsing - Argument flags are delimited by a common `--` delimiter.
non_test
update command syntax argument parsing argument flags are delimited by a common delimiter
0
232,088
17,771,421,977
IssuesEvent
2021-08-30 14:05:18
stakwork/sphinx-kotlin
https://api.github.com/repos/stakwork/sphinx-kotlin
closed
Contact detail screen
documentation
When tapping on chat header of a chat of type Conversation, instead of taking user to TribeDetail screen, we should show ContactDetail screen. ![IMG_9B6A42E2F4B2-1](https://user-images.githubusercontent.com/2242281/129960206-95997c71-18d2-49fa-8a6e-8ba1fba0e40c.jpeg) The only 2 fields editable are: - Nickname - Route Hint Address can't be edited, and the QR code icon on the right side should take user to QR code details view so it can be scanned or copied to clipboard.
1.0
Contact detail screen - When tapping on chat header of a chat of type Conversation, instead of taking user to TribeDetail screen, we should show ContactDetail screen. ![IMG_9B6A42E2F4B2-1](https://user-images.githubusercontent.com/2242281/129960206-95997c71-18d2-49fa-8a6e-8ba1fba0e40c.jpeg) The only 2 fields editable are: - Nickname - Route Hint Address can't be edited, and the QR code icon on the right side should take user to QR code details view so it can be scanned or copied to clipboard.
non_test
contact detail screen when tapping on chat header of a chat of type conversation instead of taking user to tribedetail screen we should show contactdetail screen the only fields editable are nickname route hint address can t be edited and the qr code icon on the right side should take user to qr code details view so it can be scanned or copied to clipboard
0
72,613
7,302,942,093
IssuesEvent
2018-02-27 11:21:25
Microsoft/vscode
https://api.github.com/repos/Microsoft/vscode
closed
Test: Language Pack cache cleanup
testplan-item
Tests cleaning up language pack caches - [x] anyOS: @sandy081 - tested on Windows - [x] anyOS: @isidorn - tested on OS X Complexity: 2 ## Steps - install a Language pack from the market place (e.g Bulgarian) - close VS Code. - restart VS Code with --locale=`the language pack you downloaded` (e.g --locale=bg) - go to the user data directory on your file system. Under Windows that is `C:\Users\${user}\AppData\Roaming\Code - Insiders` - ensure that there is a `languagepacks.json` file and a directorry `clp` - navigate into `clp`and ensure that there is a directory with a hash and the locale of the language pack you installed - close VS Code - restart again with the locale switch - open the log for the shared process. After ~40 seconds you should see a message that the language pack cleanup starts and that the installed language pack is skipped since it is still in use. - uninstall the language pack - close VS Code and restart it without a locale switch - open the log for the shared process. Again after ~40 seconds you should see a message that the language pack cleanup starts . This time a language pack cache entry should be deleted. - go to the file system again. Ensure that the directory entry below `clp` got deleted.
1.0
Test: Language Pack cache cleanup - Tests cleaning up language pack caches - [x] anyOS: @sandy081 - tested on Windows - [x] anyOS: @isidorn - tested on OS X Complexity: 2 ## Steps - install a Language pack from the market place (e.g Bulgarian) - close VS Code. - restart VS Code with --locale=`the language pack you downloaded` (e.g --locale=bg) - go to the user data directory on your file system. Under Windows that is `C:\Users\${user}\AppData\Roaming\Code - Insiders` - ensure that there is a `languagepacks.json` file and a directorry `clp` - navigate into `clp`and ensure that there is a directory with a hash and the locale of the language pack you installed - close VS Code - restart again with the locale switch - open the log for the shared process. After ~40 seconds you should see a message that the language pack cleanup starts and that the installed language pack is skipped since it is still in use. - uninstall the language pack - close VS Code and restart it without a locale switch - open the log for the shared process. Again after ~40 seconds you should see a message that the language pack cleanup starts . This time a language pack cache entry should be deleted. - go to the file system again. Ensure that the directory entry below `clp` got deleted.
test
test language pack cache cleanup tests cleaning up language pack caches anyos tested on windows anyos isidorn tested on os x complexity steps install a language pack from the market place e g bulgarian close vs code restart vs code with locale the language pack you downloaded e g locale bg go to the user data directory on your file system under windows that is c users user appdata roaming code insiders ensure that there is a languagepacks json file and a directorry clp navigate into clp and ensure that there is a directory with a hash and the locale of the language pack you installed close vs code restart again with the locale switch open the log for the shared process after seconds you should see a message that the language pack cleanup starts and that the installed language pack is skipped since it is still in use uninstall the language pack close vs code and restart it without a locale switch open the log for the shared process again after seconds you should see a message that the language pack cleanup starts this time a language pack cache entry should be deleted go to the file system again ensure that the directory entry below clp got deleted
1
61,202
14,619,099,206
IssuesEvent
2020-12-22 17:16:21
kenferrara/calcite
https://api.github.com/repos/kenferrara/calcite
opened
CVE-2020-13956 (Medium) detected in httpclient-4.5.9.jar
security vulnerability
## CVE-2020-13956 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>httpclient-4.5.9.jar</b></p></summary> <p>Apache HttpComponents Client</p> <p>Library home page: <a href="http://hc.apache.org/">http://hc.apache.org/</a></p> <p>Path to dependency file: calcite/server/build.gradle.kts</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.9/a25c1be5ce99d0ce99aa43eb982868c796dd0775/httpclient-4.5.9.jar,canner/.gradle/caches/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.9/a25c1be5ce99d0ce99aa43eb982868c796dd0775/httpclient-4.5.9.jar</p> <p> Dependency Hierarchy: - pig-0.16.0.jar (Root Library) - jets3t-0.9.0.jar - :x: **httpclient-4.5.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kenferrara/calcite/commit/5e2bb91de24b843d35df04b29fe7158fad414ee0">5e2bb91de24b843d35df04b29fe7158fad414ee0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution. <p>Publish Date: 2020-12-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956>CVE-2020-13956</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.httpcomponents","packageName":"httpclient","packageVersion":"4.5.9","isTransitiveDependency":true,"dependencyTree":"org.apache.pig:pig:0.16.0;net.java.dev.jets3t:jets3t:0.9.0;org.apache.httpcomponents:httpclient:4.5.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3"}],"vulnerabilityIdentifier":"CVE-2020-13956","vulnerabilityDetails":"Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-13956 (Medium) detected in httpclient-4.5.9.jar - ## CVE-2020-13956 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>httpclient-4.5.9.jar</b></p></summary> <p>Apache HttpComponents Client</p> <p>Library home page: <a href="http://hc.apache.org/">http://hc.apache.org/</a></p> <p>Path to dependency file: calcite/server/build.gradle.kts</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.9/a25c1be5ce99d0ce99aa43eb982868c796dd0775/httpclient-4.5.9.jar,canner/.gradle/caches/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.9/a25c1be5ce99d0ce99aa43eb982868c796dd0775/httpclient-4.5.9.jar</p> <p> Dependency Hierarchy: - pig-0.16.0.jar (Root Library) - jets3t-0.9.0.jar - :x: **httpclient-4.5.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kenferrara/calcite/commit/5e2bb91de24b843d35df04b29fe7158fad414ee0">5e2bb91de24b843d35df04b29fe7158fad414ee0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution. <p>Publish Date: 2020-12-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956>CVE-2020-13956</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.httpcomponents","packageName":"httpclient","packageVersion":"4.5.9","isTransitiveDependency":true,"dependencyTree":"org.apache.pig:pig:0.16.0;net.java.dev.jets3t:jets3t:0.9.0;org.apache.httpcomponents:httpclient:4.5.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3"}],"vulnerabilityIdentifier":"CVE-2020-13956","vulnerabilityDetails":"Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in httpclient jar cve medium severity vulnerability vulnerable library httpclient jar apache httpcomponents client library home page a href path to dependency file calcite server build gradle kts path to vulnerable library home wss scanner gradle caches modules files org apache httpcomponents httpclient httpclient jar canner gradle caches modules files org apache httpcomponents httpclient httpclient jar dependency hierarchy pig jar root library jar x httpclient jar vulnerable library found in head commit a href found in base branch master vulnerability details apache httpclient versions prior to version and can misinterpret malformed authority component in request uris passed to the library as java net uri object and pick the wrong target host for request execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache httpcomponents httpclient org apache httpcomponents httpclient osgi org apache httpcomponents org apache httpcomponents osgi isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails apache httpclient versions prior to version and can misinterpret malformed authority component in request uris passed to the library as java net uri object and pick the wrong target host for request execution vulnerabilityurl
0
18,234
5,612,817,055
IssuesEvent
2017-04-03 07:11:06
se-edu/addressbook-level4
https://api.github.com/repos/se-edu/addressbook-level4
opened
Remove magic string
a-CodeQuality
we should remove magic string like the command word "add" and prefix "p/" from where they are defined, as discussed [here](https://github.com/se-edu/addressbook-level4/pull/335#discussion_r107157865) and [here](https://github.com/se-edu/addressbook-level4/pull/373#pullrequestreview-30395535). Places to look for these magic strings include various tests (including testutil) and some commands' MESSAGE_USAGE.
1.0
Remove magic string - we should remove magic string like the command word "add" and prefix "p/" from where they are defined, as discussed [here](https://github.com/se-edu/addressbook-level4/pull/335#discussion_r107157865) and [here](https://github.com/se-edu/addressbook-level4/pull/373#pullrequestreview-30395535). Places to look for these magic strings include various tests (including testutil) and some commands' MESSAGE_USAGE.
non_test
remove magic string we should remove magic string like the command word add and prefix p from where they are defined as discussed and places to look for these magic strings include various tests including testutil and some commands message usage
0
128,328
10,525,265,020
IssuesEvent
2019-09-30 14:47:19
sbt/sbt
https://api.github.com/repos/sbt/sbt
closed
JUnitXmlReportPlugin generates invalid testcase name
area/test_framework uncategorized
It seems that in certain cases an invalid test-case name is generated in the JUnit XML output. ## steps This is one arbitrary example demonstrating the issue: ```scala "This example" should { "given <VER.123 FOOBAR> yield <ver.123 foobar>" in { ok } } ``` ## problem With this example (using specs2 `3.8.6`) the generated JUnit XML contains the following test-case: ```xml <testcase classname="org.example.ExampleSpec" name="123 foobar&gt;" time="0.006"> </testcase> ``` ## expectation Now, obviously the correct XML should be as follows: ```xml <testcase classname="org.example.ExampleSpec" name="This example should::given &lt;VER.123 FOOBAR&gt; yield &lt;ver.123 foobar&gt;" time="0.006"> </testcase> ``` ## notes * See https://github.com/Pyppe/specs2-junitxml-testcase-name as a simple project demonstrating the issue. * I initially reported this in https://github.com/etorreborre/specs2/issues/537, but @etorreborre graciously debugged it to be part of sbt's `JUnitXmlReportPlugin`.
1.0
JUnitXmlReportPlugin generates invalid testcase name - It seems that in certain cases an invalid test-case name is generated in the JUnit XML output. ## steps This is one arbitrary example demonstrating the issue: ```scala "This example" should { "given <VER.123 FOOBAR> yield <ver.123 foobar>" in { ok } } ``` ## problem With this example (using specs2 `3.8.6`) the generated JUnit XML contains the following test-case: ```xml <testcase classname="org.example.ExampleSpec" name="123 foobar&gt;" time="0.006"> </testcase> ``` ## expectation Now, obviously the correct XML should be as follows: ```xml <testcase classname="org.example.ExampleSpec" name="This example should::given &lt;VER.123 FOOBAR&gt; yield &lt;ver.123 foobar&gt;" time="0.006"> </testcase> ``` ## notes * See https://github.com/Pyppe/specs2-junitxml-testcase-name as a simple project demonstrating the issue. * I initially reported this in https://github.com/etorreborre/specs2/issues/537, but @etorreborre graciously debugged it to be part of sbt's `JUnitXmlReportPlugin`.
test
junitxmlreportplugin generates invalid testcase name it seems that in certain cases an invalid test case name is generated in the junit xml output steps this is one arbitrary example demonstrating the issue scala this example should given yield in ok problem with this example using the generated junit xml contains the following test case xml expectation now obviously the correct xml should be as follows xml notes see as a simple project demonstrating the issue i initially reported this in but etorreborre graciously debugged it to be part of sbt s junitxmlreportplugin
1
85,353
10,606,331,140
IssuesEvent
2019-10-10 22:58:07
EdgeFortress/OSS-13
https://api.github.com/repos/EdgeFortress/OSS-13
closed
UI Design Layout
design ui
Original game's interface looks a bit outdated: ![image](https://user-images.githubusercontent.com/22466026/57739873-e38bd780-76bd-11e9-81f9-10b0a9d4685b.png) We need new UI Design Layout with next prerequisites: 1. Optimum TileGrid size is 960 x 960 pixels (64 pixels one tile * 15 tiles); 2. The chat should be large enough to accommodate about as much text as the original one. 3. Least possible number of elements should placed on the TileGrid. 4. Mostly all GUI elements above chat can be moved to hotkeys and sub-windows. 5. RMB menu can be implemented as pop-up bubbles with objects/actions around clicked point. 6. You can use as much subwindows and hotkeys as you needed. 7. Console at the bottom of window is obsolete and will not be implemented. Key gameplay UI elements need to be designed: - [ ] Chat - [ ] Inventory - [ ] Right Mouse Button menu (now it's simple Desktop-like drop-down menu). - [ ] Health, Body Temperature, Gas indicators.
1.0
UI Design Layout - Original game's interface looks a bit outdated: ![image](https://user-images.githubusercontent.com/22466026/57739873-e38bd780-76bd-11e9-81f9-10b0a9d4685b.png) We need new UI Design Layout with next prerequisites: 1. Optimum TileGrid size is 960 x 960 pixels (64 pixels one tile * 15 tiles); 2. The chat should be large enough to accommodate about as much text as the original one. 3. Least possible number of elements should placed on the TileGrid. 4. Mostly all GUI elements above chat can be moved to hotkeys and sub-windows. 5. RMB menu can be implemented as pop-up bubbles with objects/actions around clicked point. 6. You can use as much subwindows and hotkeys as you needed. 7. Console at the bottom of window is obsolete and will not be implemented. Key gameplay UI elements need to be designed: - [ ] Chat - [ ] Inventory - [ ] Right Mouse Button menu (now it's simple Desktop-like drop-down menu). - [ ] Health, Body Temperature, Gas indicators.
non_test
ui design layout original game s interface looks a bit outdated we need new ui design layout with next prerequisites optimum tilegrid size is x pixels pixels one tile tiles the chat should be large enough to accommodate about as much text as the original one least possible number of elements should placed on the tilegrid mostly all gui elements above chat can be moved to hotkeys and sub windows rmb menu can be implemented as pop up bubbles with objects actions around clicked point you can use as much subwindows and hotkeys as you needed console at the bottom of window is obsolete and will not be implemented key gameplay ui elements need to be designed chat inventory right mouse button menu now it s simple desktop like drop down menu health body temperature gas indicators
0
88,020
25,281,481,510
IssuesEvent
2022-11-16 16:04:38
intel/media-driver
https://api.github.com/repos/intel/media-driver
closed
[Bug]: Cannot build on musl
Build Common
### Which component impacted? Build ### Is it regression? Good in old configuration? _No response_ ### What happened? Using musl as libc, I am not able to build because execinfo.h is not found. ### What's the usage scenario when you are seeing the problem? Others ### What impacted? _No response_ ### Debug Information Here is the output of the build : media_softlet/linux/common/os/osservice/mos_utilities_specific.cpp:37:10: fatal error: execinfo.h: No such file or directory ### Do you want to contribute a patch to fix the issue? [Here](https://raw.githubusercontent.com/void-linux/void-packages/master/srcpkgs/intel-media-driver/patches/execinfo.patch) is a patch from void linux. It doesn't work on master, but I'll try to edit it.
1.0
[Bug]: Cannot build on musl - ### Which component impacted? Build ### Is it regression? Good in old configuration? _No response_ ### What happened? Using musl as libc, I am not able to build because execinfo.h is not found. ### What's the usage scenario when you are seeing the problem? Others ### What impacted? _No response_ ### Debug Information Here is the output of the build : media_softlet/linux/common/os/osservice/mos_utilities_specific.cpp:37:10: fatal error: execinfo.h: No such file or directory ### Do you want to contribute a patch to fix the issue? [Here](https://raw.githubusercontent.com/void-linux/void-packages/master/srcpkgs/intel-media-driver/patches/execinfo.patch) is a patch from void linux. It doesn't work on master, but I'll try to edit it.
non_test
cannot build on musl which component impacted build is it regression good in old configuration no response what happened using musl as libc i am not able to build because execinfo h is not found what s the usage scenario when you are seeing the problem others what impacted no response debug information here is the output of the build media softlet linux common os osservice mos utilities specific cpp fatal error execinfo h no such file or directory do you want to contribute a patch to fix the issue is a patch from void linux it doesn t work on master but i ll try to edit it
0
288,813
31,930,913,635
IssuesEvent
2023-09-19 07:21:32
Trinadh465/linux-4.1.15_CVE-2023-4128
https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2023-4128
opened
CVE-2021-26931 (Medium) detected in linux-stable-rtv4.1.33, linux-stable-rtv4.1.33
Mend: dependency security vulnerability
## CVE-2021-26931 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv4.1.33</b>, <b>linux-stable-rtv4.1.33</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel 2.6.39 through 5.10.16, as used in Xen. Block, net, and SCSI backends consider certain errors a plain bug, deliberately causing a kernel crash. For errors potentially being at least under the influence of guests (such as out of memory conditions), it isn't correct to assume a plain bug. Memory allocations potentially causing such crashes occur only when Linux is running in PV mode, though. This affects drivers/block/xen-blkback/blkback.c and drivers/xen/xen-scsiback.c. <p>Publish Date: 2021-02-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-26931>CVE-2021-26931</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-26931">https://nvd.nist.gov/vuln/detail/CVE-2021-26931</a></p> <p>Release Date: 2021-02-17</p> <p>Fix Resolution: linux-libc-headers - 5.13;linux-yocto - 5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1,4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-26931 (Medium) detected in linux-stable-rtv4.1.33, linux-stable-rtv4.1.33 - ## CVE-2021-26931 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv4.1.33</b>, <b>linux-stable-rtv4.1.33</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel 2.6.39 through 5.10.16, as used in Xen. Block, net, and SCSI backends consider certain errors a plain bug, deliberately causing a kernel crash. For errors potentially being at least under the influence of guests (such as out of memory conditions), it isn't correct to assume a plain bug. Memory allocations potentially causing such crashes occur only when Linux is running in PV mode, though. This affects drivers/block/xen-blkback/blkback.c and drivers/xen/xen-scsiback.c. <p>Publish Date: 2021-02-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-26931>CVE-2021-26931</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-26931">https://nvd.nist.gov/vuln/detail/CVE-2021-26931</a></p> <p>Release Date: 2021-02-17</p> <p>Fix Resolution: linux-libc-headers - 5.13;linux-yocto - 5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1,4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in linux stable linux stable cve medium severity vulnerability vulnerable libraries linux stable linux stable vulnerability details an issue was discovered in the linux kernel through as used in xen block net and scsi backends consider certain errors a plain bug deliberately causing a kernel crash for errors potentially being at least under the influence of guests such as out of memory conditions it isn t correct to assume a plain bug memory allocations potentially causing such crashes occur only when linux is running in pv mode though this affects drivers block xen blkback blkback c and drivers xen xen scsiback c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux libc headers linux yocto gitautoinc gitautoinc step up your open source security game with mend
0
107,784
9,223,256,978
IssuesEvent
2019-03-12 02:37:54
brave/browser-android-tabs
https://api.github.com/repos/brave/browser-android-tabs
closed
twitch.tv crashes when trying to view Desktop Site
QA/Test-plan-specified QA/Yes crash
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> When I go to twitch.tv and try to select 'Desktop Site' from hamburger menu, Brave crashes. ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Install 1.0.86 (BR_6) 2. Navigate to twitch.tv, page loads 3. Tap on hamburger menu 4. Select Desktop site. ## Actual result: <!--Please add screenshots if needed--> Get message which says 'Unfortunately, Brave has stopped.' ## Expected result: No crash, desktop site should open. ## Issue reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> easily ## Issue happens on: <!-- Mention yes or no --> - Current Playstore version? no, does not reproduce on 1.0.84 - Beta build? Yes, Rewards Beta ## Device Details: - Install Type(ARM, x86): all - Device(Phone, Tablet, Phablet): all - Android Version: all ## Brave version: 1.0.86 (BR_6) ### Website problems only: - Does the issue resolve itself when disabling Brave Shields? no - Is the issue reproducible on the latest version of Chrome? no ### Additional Information Have only experienced on the Rewards build so far, but do not need to enable Rewards to experience the crash.
1.0
twitch.tv crashes when trying to view Desktop Site - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> When I go to twitch.tv and try to select 'Desktop Site' from hamburger menu, Brave crashes. ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Install 1.0.86 (BR_6) 2. Navigate to twitch.tv, page loads 3. Tap on hamburger menu 4. Select Desktop site. ## Actual result: <!--Please add screenshots if needed--> Get message which says 'Unfortunately, Brave has stopped.' ## Expected result: No crash, desktop site should open. ## Issue reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> easily ## Issue happens on: <!-- Mention yes or no --> - Current Playstore version? no, does not reproduce on 1.0.84 - Beta build? Yes, Rewards Beta ## Device Details: - Install Type(ARM, x86): all - Device(Phone, Tablet, Phablet): all - Android Version: all ## Brave version: 1.0.86 (BR_6) ### Website problems only: - Does the issue resolve itself when disabling Brave Shields? no - Is the issue reproducible on the latest version of Chrome? no ### Additional Information Have only experienced on the Rewards build so far, but do not need to enable Rewards to experience the crash.
test
twitch tv crashes when trying to view desktop site have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description when i go to twitch tv and try to select desktop site from hamburger menu brave crashes steps to reproduce install br navigate to twitch tv page loads tap on hamburger menu select desktop site actual result get message which says unfortunately brave has stopped expected result no crash desktop site should open issue reproduces how often easily issue happens on current playstore version no does not reproduce on beta build yes rewards beta device details install type arm all device phone tablet phablet all android version all brave version br website problems only does the issue resolve itself when disabling brave shields no is the issue reproducible on the latest version of chrome no additional information have only experienced on the rewards build so far but do not need to enable rewards to experience the crash
1
61,970
14,648,070,031
IssuesEvent
2020-12-27 00:02:22
wallanpsantos/cursoAngularNove
https://api.github.com/repos/wallanpsantos/cursoAngularNove
opened
CVE-2020-7774 (High) detected in y18n-4.0.0.tgz
security vulnerability
## CVE-2020-7774 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>y18n-4.0.0.tgz</b></p></summary> <p>the bare-bones internationalization library used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz">https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz</a></p> <p>Path to dependency file: cursoAngularNove/backend/package.json</p> <p>Path to vulnerable library: cursoAngularNove/backend/node_modules/y18n/package.json,cursoAngularNove/frontend/node_modules/y18n/package.json</p> <p> Dependency Hierarchy: - json-server-0.16.1.tgz (Root Library) - yargs-15.4.1.tgz - :x: **y18n-4.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/wallanpsantos/cursoAngularNove/commit/c6b9eaff3e9f65596ffd12cd6b64751337db2942">c6b9eaff3e9f65596ffd12cd6b64751337db2942</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package y18n before 4.0.1 and 5.0.5. PoC by po6ix: const y18n = require('y18n')(); y18n.setLocale('__proto__'); y18n.updateLocale({polluted: true}); console.log(polluted); // true <p>Publish Date: 2020-11-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7774>CVE-2020-7774</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7774">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7774</a></p> <p>Release Date: 2020-11-17</p> <p>Fix Resolution: 5.0.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7774 (High) detected in y18n-4.0.0.tgz - ## CVE-2020-7774 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>y18n-4.0.0.tgz</b></p></summary> <p>the bare-bones internationalization library used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz">https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz</a></p> <p>Path to dependency file: cursoAngularNove/backend/package.json</p> <p>Path to vulnerable library: cursoAngularNove/backend/node_modules/y18n/package.json,cursoAngularNove/frontend/node_modules/y18n/package.json</p> <p> Dependency Hierarchy: - json-server-0.16.1.tgz (Root Library) - yargs-15.4.1.tgz - :x: **y18n-4.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/wallanpsantos/cursoAngularNove/commit/c6b9eaff3e9f65596ffd12cd6b64751337db2942">c6b9eaff3e9f65596ffd12cd6b64751337db2942</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package y18n before 4.0.1 and 5.0.5. PoC by po6ix: const y18n = require('y18n')(); y18n.setLocale('__proto__'); y18n.updateLocale({polluted: true}); console.log(polluted); // true <p>Publish Date: 2020-11-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7774>CVE-2020-7774</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7774">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7774</a></p> <p>Release Date: 2020-11-17</p> <p>Fix Resolution: 5.0.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in tgz cve high severity vulnerability vulnerable library tgz the bare bones internationalization library used by yargs library home page a href path to dependency file cursoangularnove backend package json path to vulnerable library cursoangularnove backend node modules package json cursoangularnove frontend node modules package json dependency hierarchy json server tgz root library yargs tgz x tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects the package before and poc by const require setlocale proto updatelocale polluted true console log polluted true publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
356,411
25,176,183,737
IssuesEvent
2022-11-11 09:27:50
Santosh3007/pe
https://api.github.com/repos/Santosh3007/pe
opened
Uneven Text size in table of contents
severity.VeryLow type.DocumentationBug
## Issue - Uneven text size for `Appendix: Instructions for manual testing` ## Screenshot ![image.png](https://raw.githubusercontent.com/Santosh3007/pe/main/files/5e9c3280-9479-48fd-ac5f-71e72edec01d.png) <!--session: 1668151550995-b01faa43-82fc-43e2-93df-11e1d10a284a--> <!--Version: Web v3.4.4-->
1.0
Uneven Text size in table of contents - ## Issue - Uneven text size for `Appendix: Instructions for manual testing` ## Screenshot ![image.png](https://raw.githubusercontent.com/Santosh3007/pe/main/files/5e9c3280-9479-48fd-ac5f-71e72edec01d.png) <!--session: 1668151550995-b01faa43-82fc-43e2-93df-11e1d10a284a--> <!--Version: Web v3.4.4-->
non_test
uneven text size in table of contents issue uneven text size for appendix instructions for manual testing screenshot
0
294,571
25,383,735,207
IssuesEvent
2022-11-21 19:53:33
chocolatey/choco
https://api.github.com/repos/chocolatey/choco
opened
PushCommand tests should push to a mocked endpoint and not Chocolatey Community Repository
0 - Backlog Tests
## Is Your Feature Request Related To A Problem? Please describe. Currently PushCommand tests require an API Key to push to the Chocolatey Community Repository. This means that community members are not able to run the full test suite as they do not have access to this API Key. ## Describe The Solution. Why is it needed? PowerShell has the ability to respond to HTTP requests. We could therefore have our tests listen for Chocolatey to push a package to it and respond with exactly the message we want. ## Additional Context. NA ## Related Issues NA
1.0
PushCommand tests should push to a mocked endpoint and not Chocolatey Community Repository - ## Is Your Feature Request Related To A Problem? Please describe. Currently PushCommand tests require an API Key to push to the Chocolatey Community Repository. This means that community members are not able to run the full test suite as they do not have access to this API Key. ## Describe The Solution. Why is it needed? PowerShell has the ability to respond to HTTP requests. We could therefore have our tests listen for Chocolatey to push a package to it and respond with exactly the message we want. ## Additional Context. NA ## Related Issues NA
test
pushcommand tests should push to a mocked endpoint and not chocolatey community repository is your feature request related to a problem please describe currently pushcommand tests require an api key to push to the chocolatey community repository this means that community members are not able to run the full test suite as they do not have access to this api key describe the solution why is it needed powershell has the ability to respond to http requests we could therefore have our tests listen for chocolatey to push a package to it and respond with exactly the message we want additional context na related issues na
1
152,464
12,109,742,436
IssuesEvent
2020-04-21 09:15:41
WoWManiaUK/Redemption
https://api.github.com/repos/WoWManiaUK/Redemption
closed
[NPC/ICC] Nerub'ar Broodkeeper (Web Wrap)
Fix - Tester Confirmed
**Links:** https://www.wowhead.com/npc=36725/nerubar-broodkeeper#abilities https://www.wowhead.com/spell=70980/web-wrap https://www.wowhead.com/npc=38028/web-wrap **What is Happening:** The [Nerub'ar Broodkeeper](https://www.wowhead.com/npc=36725/nerubar-broodkeeper#abilities) in Icecrown Citadel currently cast their [Web Wraps](https://www.wowhead.com/spell=70980/web-wrap) properly, but only the first part of the spell works. The target gets immobilized for 3 seconds, but are then freed and nothing more happens. **What Should happen:** After the 3s immobilization of the Web Wrap cast time, the target should become encased in a [cocoon](https://www.wowhead.com/npc=38028/web-wrap) .
1.0
[NPC/ICC] Nerub'ar Broodkeeper (Web Wrap) - **Links:** https://www.wowhead.com/npc=36725/nerubar-broodkeeper#abilities https://www.wowhead.com/spell=70980/web-wrap https://www.wowhead.com/npc=38028/web-wrap **What is Happening:** The [Nerub'ar Broodkeeper](https://www.wowhead.com/npc=36725/nerubar-broodkeeper#abilities) in Icecrown Citadel currently cast their [Web Wraps](https://www.wowhead.com/spell=70980/web-wrap) properly, but only the first part of the spell works. The target gets immobilized for 3 seconds, but are then freed and nothing more happens. **What Should happen:** After the 3s immobilization of the Web Wrap cast time, the target should become encased in a [cocoon](https://www.wowhead.com/npc=38028/web-wrap) .
test
nerub ar broodkeeper web wrap links what is happening the in icecrown citadel currently cast their properly but only the first part of the spell works the target gets immobilized for seconds but are then freed and nothing more happens what should happen after the immobilization of the web wrap cast time the target should become encased in a
1
210,205
16,090,673,503
IssuesEvent
2021-04-26 16:18:08
spring-projects/spring-framework
https://api.github.com/repos/spring-projects/spring-framework
closed
MockMvc's MVC_RESULT_ATTRIBUTE lost with HandlerMappingIntrospector and RouterFunctions in use
in: test in: web type: bug
Related to https://github.com/spring-projects/spring-security-samples/issues/9 A contributor shared the following sample application: https://github.com/hantsy/spring-webmvc-auth0-sample The tests result in a `NullPointerException` because the `MockMvc.MVC_RESULT_ATTRIBUTE` is missing. ```java HandlerExecutionChain chain = super.getHandler(request); if (chain != null) { DefaultMvcResult mvcResult = getMvcResult(request); // returns null mvcResult.setHandler(chain.getHandler()); mvcResult.setInterceptors(chain.getInterceptors()); } return chain; ``` It gets removed due to the following arrangement: * Spring Security's `CorsFilter` by default uses `HandlerMappingIntrospector` * `HandlerMappingIntrospector` uses `RequestAttributeChangeIgnoringWrapper` which ignores all but `PATH_ATTRIBUTE` * `RequestPredicates#restoreAttributes` attempts to restore the attributes to a previous state by clearing the attribute set and then re-adding each attribute one by one Before `CorsFilter` runs, `MVC_RESULT_ATTRIBUTE` is present in the request. When `RequestPredicates#restoreAttributes` is run, it removes all attributes. Then, when it tries to add the original set back in, `RequestAttributeChangeIgnoringWrapper` only adds `PATH_ATTRIBUTE` back in. For the specific sample, the tests can be repaired by removing the `CorsFilter` or by exposing a custom `CorsConfigurationSource` bean since either of those will prevent `RequestAttributeChangeIgnoringWrapper` from wrapping the request. I was also able to fix the tests by adding the following to `RequestAttributeChangeIgnoringWrapper`: ```java @Override public void removeAttribute(String name) { if (name.equals(ServletRequestPathUtils.PATH_ATTRIBUTE) || name.equals(UrlPathHelper.PATH_ATTRIBUTE)) { super.removeAttribute(name); } } ``` At least in this isolated case, it seems reasonable that if an attribute cannot be set, it also should not be able to be removed.
1.0
MockMvc's MVC_RESULT_ATTRIBUTE lost with HandlerMappingIntrospector and RouterFunctions in use - Related to https://github.com/spring-projects/spring-security-samples/issues/9 A contributor shared the following sample application: https://github.com/hantsy/spring-webmvc-auth0-sample The tests result in a `NullPointerException` because the `MockMvc.MVC_RESULT_ATTRIBUTE` is missing. ```java HandlerExecutionChain chain = super.getHandler(request); if (chain != null) { DefaultMvcResult mvcResult = getMvcResult(request); // returns null mvcResult.setHandler(chain.getHandler()); mvcResult.setInterceptors(chain.getInterceptors()); } return chain; ``` It gets removed due to the following arrangement: * Spring Security's `CorsFilter` by default uses `HandlerMappingIntrospector` * `HandlerMappingIntrospector` uses `RequestAttributeChangeIgnoringWrapper` which ignores all but `PATH_ATTRIBUTE` * `RequestPredicates#restoreAttributes` attempts to restore the attributes to a previous state by clearing the attribute set and then re-adding each attribute one by one Before `CorsFilter` runs, `MVC_RESULT_ATTRIBUTE` is present in the request. When `RequestPredicates#restoreAttributes` is run, it removes all attributes. Then, when it tries to add the original set back in, `RequestAttributeChangeIgnoringWrapper` only adds `PATH_ATTRIBUTE` back in. For the specific sample, the tests can be repaired by removing the `CorsFilter` or by exposing a custom `CorsConfigurationSource` bean since either of those will prevent `RequestAttributeChangeIgnoringWrapper` from wrapping the request. I was also able to fix the tests by adding the following to `RequestAttributeChangeIgnoringWrapper`: ```java @Override public void removeAttribute(String name) { if (name.equals(ServletRequestPathUtils.PATH_ATTRIBUTE) || name.equals(UrlPathHelper.PATH_ATTRIBUTE)) { super.removeAttribute(name); } } ``` At least in this isolated case, it seems reasonable that if an attribute cannot be set, it also should not be able to be removed.
test
mockmvc s mvc result attribute lost with handlermappingintrospector and routerfunctions in use related to a contributor shared the following sample application the tests result in a nullpointerexception because the mockmvc mvc result attribute is missing java handlerexecutionchain chain super gethandler request if chain null defaultmvcresult mvcresult getmvcresult request returns null mvcresult sethandler chain gethandler mvcresult setinterceptors chain getinterceptors return chain it gets removed due to the following arrangement spring security s corsfilter by default uses handlermappingintrospector handlermappingintrospector uses requestattributechangeignoringwrapper which ignores all but path attribute requestpredicates restoreattributes attempts to restore the attributes to a previous state by clearing the attribute set and then re adding each attribute one by one before corsfilter runs mvc result attribute is present in the request when requestpredicates restoreattributes is run it removes all attributes then when it tries to add the original set back in requestattributechangeignoringwrapper only adds path attribute back in for the specific sample the tests can be repaired by removing the corsfilter or by exposing a custom corsconfigurationsource bean since either of those will prevent requestattributechangeignoringwrapper from wrapping the request i was also able to fix the tests by adding the following to requestattributechangeignoringwrapper java override public void removeattribute string name if name equals servletrequestpathutils path attribute name equals urlpathhelper path attribute super removeattribute name at least in this isolated case it seems reasonable that if an attribute cannot be set it also should not be able to be removed
1
307,646
23,209,893,949
IssuesEvent
2022-08-02 09:13:09
Shard-Labs/starknet-devnet
https://api.github.com/repos/Shard-Labs/starknet-devnet
closed
Error: Installing startnet-devnet on Windows
documentation enhancement help wanted
**Describe the bug** I get this bug when I run the command `pip install starknet-devnet` ![image](https://user-images.githubusercontent.com/31249460/180617846-0f0b8806-9c87-4472-a919-52afeb93197f.png) **Machine specifications** - OS: Windows - CPU: I5 ( Intel64 Family 6 Model 158 Stepping 9 GenuineIntel ~2500 Mhz) - Memory: 16GB RAM
1.0
Error: Installing startnet-devnet on Windows - **Describe the bug** I get this bug when I run the command `pip install starknet-devnet` ![image](https://user-images.githubusercontent.com/31249460/180617846-0f0b8806-9c87-4472-a919-52afeb93197f.png) **Machine specifications** - OS: Windows - CPU: I5 ( Intel64 Family 6 Model 158 Stepping 9 GenuineIntel ~2500 Mhz) - Memory: 16GB RAM
non_test
error installing startnet devnet on windows describe the bug i get this bug when i run the command pip install starknet devnet machine specifications os windows cpu family model stepping genuineintel mhz memory ram
0
52,810
13,771,901,607
IssuesEvent
2020-10-07 23:04:27
jgeraigery/please-del-this-repo
https://api.github.com/repos/jgeraigery/please-del-this-repo
opened
CVE-2017-16137 (Medium) detected in multiple libraries
security vulnerability
## CVE-2017-16137 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>debug-2.2.0.tgz</b>, <b>debug-2.3.3.tgz</b>, <b>debug-2.6.8.tgz</b>, <b>debug-2.6.0.tgz</b></p></summary> <p> <details><summary><b>debug-2.2.0.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p> <p>Path to dependency file: please-del-this-repo/package.json</p> <p>Path to vulnerable library: please-del-this-repo/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - express-4.13.4.tgz (Root Library) - :x: **debug-2.2.0.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.3.3.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.3.3.tgz">https://registry.npmjs.org/debug/-/debug-2.3.3.tgz</a></p> <p>Path to dependency file: please-del-this-repo/package.json</p> <p>Path to vulnerable library: please-del-this-repo/node_modules/socket.io-adapter/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - socket.io-1.7.3.tgz (Root Library) - :x: **debug-2.3.3.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.6.8.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.8.tgz">https://registry.npmjs.org/debug/-/debug-2.6.8.tgz</a></p> <p>Path to dependency file: please-del-this-repo/package.json</p> <p>Path to vulnerable library: please-del-this-repo/node_modules/eslint-module-utils/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - nodemon-1.11.0.tgz (Root Library) - chokidar-1.7.0.tgz - fsevents-1.1.2.tgz - node-pre-gyp-0.6.36.tgz - tar-pack-3.4.0.tgz - :x: **debug-2.6.8.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.6.0.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.0.tgz">https://registry.npmjs.org/debug/-/debug-2.6.0.tgz</a></p> <p>Path to dependency file: please-del-this-repo/package.json</p> <p>Path to vulnerable library: please-del-this-repo/node_modules/mocha/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - mocha-3.4.2.tgz (Root Library) - :x: **debug-2.6.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/please-del-this-repo/commit/e701d80cb459747be96e9c0e79c8640eca26c232">e701d80cb459747be96e9c0e79c8640eca26c232</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 2.6.9</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.2.0","isTransitiveDependency":true,"dependencyTree":"express:4.13.4;debug:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"},{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.3.3","isTransitiveDependency":true,"dependencyTree":"socket.io:1.7.3;debug:2.3.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"},{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.6.8","isTransitiveDependency":true,"dependencyTree":"nodemon:1.11.0;chokidar:1.7.0;fsevents:1.1.2;node-pre-gyp:0.6.36;tar-pack:3.4.0;debug:2.6.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"},{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.6.0","isTransitiveDependency":true,"dependencyTree":"mocha:3.4.2;debug:2.6.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"}],"vulnerabilityIdentifier":"CVE-2017-16137","vulnerabilityDetails":"The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2017-16137 (Medium) detected in multiple libraries - ## CVE-2017-16137 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>debug-2.2.0.tgz</b>, <b>debug-2.3.3.tgz</b>, <b>debug-2.6.8.tgz</b>, <b>debug-2.6.0.tgz</b></p></summary> <p> <details><summary><b>debug-2.2.0.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p> <p>Path to dependency file: please-del-this-repo/package.json</p> <p>Path to vulnerable library: please-del-this-repo/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - express-4.13.4.tgz (Root Library) - :x: **debug-2.2.0.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.3.3.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.3.3.tgz">https://registry.npmjs.org/debug/-/debug-2.3.3.tgz</a></p> <p>Path to dependency file: please-del-this-repo/package.json</p> <p>Path to vulnerable library: please-del-this-repo/node_modules/socket.io-adapter/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - socket.io-1.7.3.tgz (Root Library) - :x: **debug-2.3.3.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.6.8.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.8.tgz">https://registry.npmjs.org/debug/-/debug-2.6.8.tgz</a></p> <p>Path to dependency file: please-del-this-repo/package.json</p> <p>Path to vulnerable library: please-del-this-repo/node_modules/eslint-module-utils/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - nodemon-1.11.0.tgz (Root Library) - chokidar-1.7.0.tgz - fsevents-1.1.2.tgz - node-pre-gyp-0.6.36.tgz - tar-pack-3.4.0.tgz - :x: **debug-2.6.8.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.6.0.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.0.tgz">https://registry.npmjs.org/debug/-/debug-2.6.0.tgz</a></p> <p>Path to dependency file: please-del-this-repo/package.json</p> <p>Path to vulnerable library: please-del-this-repo/node_modules/mocha/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - mocha-3.4.2.tgz (Root Library) - :x: **debug-2.6.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/please-del-this-repo/commit/e701d80cb459747be96e9c0e79c8640eca26c232">e701d80cb459747be96e9c0e79c8640eca26c232</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 2.6.9</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.2.0","isTransitiveDependency":true,"dependencyTree":"express:4.13.4;debug:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"},{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.3.3","isTransitiveDependency":true,"dependencyTree":"socket.io:1.7.3;debug:2.3.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"},{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.6.8","isTransitiveDependency":true,"dependencyTree":"nodemon:1.11.0;chokidar:1.7.0;fsevents:1.1.2;node-pre-gyp:0.6.36;tar-pack:3.4.0;debug:2.6.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"},{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.6.0","isTransitiveDependency":true,"dependencyTree":"mocha:3.4.2;debug:2.6.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"}],"vulnerabilityIdentifier":"CVE-2017-16137","vulnerabilityDetails":"The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries debug tgz debug tgz debug tgz debug tgz debug tgz small debugging utility library home page a href path to dependency file please del this repo package json path to vulnerable library please del this repo node modules debug package json dependency hierarchy express tgz root library x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file please del this repo package json path to vulnerable library please del this repo node modules socket io adapter node modules debug package json dependency hierarchy socket io tgz root library x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file please del this repo package json path to vulnerable library please del this repo node modules eslint module utils node modules debug package json dependency hierarchy nodemon tgz root library chokidar tgz fsevents tgz node pre gyp tgz tar pack tgz x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file please del this repo package json path to vulnerable library please del this repo node modules mocha node modules debug package json dependency hierarchy mocha tgz root library x debug tgz vulnerable library found in head commit a href found in base branch master vulnerability details the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue vulnerabilityurl
0
51,815
6,198,568,817
IssuesEvent
2017-07-05 19:28:43
fedora-infra/bodhi
https://api.github.com/repos/fedora-infra/bodhi
closed
The pydocstyle tests do not work with tox
Tests
I noticed today that the pydocstyle tests do not actually check the paths that are given. I noticed that deleting a docblock from ```bodhi/server/views/generic.py``` doesn't cause any errors when btest is run, and adding all of bodhi to the ```match``` argument in ```tox.ini``` doesn't cause any failures (and should cause plenty).
1.0
The pydocstyle tests do not work with tox - I noticed today that the pydocstyle tests do not actually check the paths that are given. I noticed that deleting a docblock from ```bodhi/server/views/generic.py``` doesn't cause any errors when btest is run, and adding all of bodhi to the ```match``` argument in ```tox.ini``` doesn't cause any failures (and should cause plenty).
test
the pydocstyle tests do not work with tox i noticed today that the pydocstyle tests do not actually check the paths that are given i noticed that deleting a docblock from bodhi server views generic py doesn t cause any errors when btest is run and adding all of bodhi to the match argument in tox ini doesn t cause any failures and should cause plenty
1
256,039
27,552,556,975
IssuesEvent
2023-03-07 15:49:25
billmcchesney1/flow
https://api.github.com/repos/billmcchesney1/flow
closed
CVE-2022-41915 (Medium) detected in netty-codec-http-4.1.49.Final.jar - autoclosed
Mend: dependency security vulnerability
## CVE-2022-41915 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-codec-http-4.1.49.Final.jar</b></p></summary> <p>Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers and clients.</p> <p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-codec-http/4.1.49.Final/netty-codec-http-4.1.49.Final.jar</p> <p> Dependency Hierarchy: - transport-7.9.3.jar (Root Library) - transport-netty4-client-7.9.3.jar - :x: **netty-codec-http-4.1.49.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/flow/commit/eb687271afab9d7c61ca82fce2ed4fdb3d5e1a70">eb687271afab9d7c61ca82fce2ed4fdb3d5e1a70</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Netty project is an event-driven asynchronous network application framework. Starting in version 4.1.83.Final and prior to 4.1.86.Final, when calling `DefaultHttpHeadesr.set` with an _iterator_ of values, header value validation was not performed, allowing malicious header values in the iterator to perform HTTP Response Splitting. This issue has been patched in version 4.1.86.Final. Integrators can work around the issue by changing the `DefaultHttpHeaders.set(CharSequence, Iterator<?>)` call, into a `remove()` call, and call `add()` in a loop over the iterator of values. <p>Publish Date: 2022-12-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41915>CVE-2022-41915</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-12-13</p> <p>Fix Resolution: io.netty:netty-codec-http:4.1.86.Final</p> </p> </details> <p></p>
True
CVE-2022-41915 (Medium) detected in netty-codec-http-4.1.49.Final.jar - autoclosed - ## CVE-2022-41915 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-codec-http-4.1.49.Final.jar</b></p></summary> <p>Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers and clients.</p> <p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-codec-http/4.1.49.Final/netty-codec-http-4.1.49.Final.jar</p> <p> Dependency Hierarchy: - transport-7.9.3.jar (Root Library) - transport-netty4-client-7.9.3.jar - :x: **netty-codec-http-4.1.49.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/flow/commit/eb687271afab9d7c61ca82fce2ed4fdb3d5e1a70">eb687271afab9d7c61ca82fce2ed4fdb3d5e1a70</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Netty project is an event-driven asynchronous network application framework. Starting in version 4.1.83.Final and prior to 4.1.86.Final, when calling `DefaultHttpHeadesr.set` with an _iterator_ of values, header value validation was not performed, allowing malicious header values in the iterator to perform HTTP Response Splitting. This issue has been patched in version 4.1.86.Final. Integrators can work around the issue by changing the `DefaultHttpHeaders.set(CharSequence, Iterator<?>)` call, into a `remove()` call, and call `add()` in a loop over the iterator of values. <p>Publish Date: 2022-12-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41915>CVE-2022-41915</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-12-13</p> <p>Fix Resolution: io.netty:netty-codec-http:4.1.86.Final</p> </p> </details> <p></p>
non_test
cve medium detected in netty codec http final jar autoclosed cve medium severity vulnerability vulnerable library netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository io netty netty codec http final netty codec http final jar dependency hierarchy transport jar root library transport client jar x netty codec http final jar vulnerable library found in head commit a href found in base branch master vulnerability details netty project is an event driven asynchronous network application framework starting in version final and prior to final when calling defaulthttpheadesr set with an iterator of values header value validation was not performed allowing malicious header values in the iterator to perform http response splitting this issue has been patched in version final integrators can work around the issue by changing the defaulthttpheaders set charsequence iterator call into a remove call and call add in a loop over the iterator of values publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution io netty netty codec http final
0
242,515
20,252,599,487
IssuesEvent
2022-02-14 19:27:50
rspott/WAF-test02
https://api.github.com/repos/rspott/WAF-test02
opened
Discover and remediate common risks to improve Secure Score in Azure Security Center
WARP-Import test1 Security Security & Compliance Security Center
<a href="https://docs.microsoft.com/azure/architecture/framework/Security/governance#discover-and-remediate-common-risks">Discover and remediate common risks to improve Secure Score in Azure Security Center</a> <p><b>Why Consider This?</b></p> Secure Score is a snapshot of your security posture relative to Microsoft recommendations. Identifying and remediating common security hygiene risks significantly reduces overall risk to the organization by increasing cost to attackers. <p><b>Context</b></p> <p><span>When routine and well-established attack vectors are mitigated, attackers are forced to acquire and use advanced or untested attack methods. Azure Secure Score in Azure Security Center is one method to easily identify and address well-known issues in Azure.</span></p> <p><b>Suggested Actions</b></p> <p><span>Identify well-known risks for your Azure tenants, remediate those risks, and track your progress using Secure Score</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/azure/architecture/framework/Security/governance#discover-and-remediate-common-risks" target="_blank"><span>Discover and remediate common risks</span></a><span /></p>
1.0
Discover and remediate common risks to improve Secure Score in Azure Security Center - <a href="https://docs.microsoft.com/azure/architecture/framework/Security/governance#discover-and-remediate-common-risks">Discover and remediate common risks to improve Secure Score in Azure Security Center</a> <p><b>Why Consider This?</b></p> Secure Score is a snapshot of your security posture relative to Microsoft recommendations. Identifying and remediating common security hygiene risks significantly reduces overall risk to the organization by increasing cost to attackers. <p><b>Context</b></p> <p><span>When routine and well-established attack vectors are mitigated, attackers are forced to acquire and use advanced or untested attack methods. Azure Secure Score in Azure Security Center is one method to easily identify and address well-known issues in Azure.</span></p> <p><b>Suggested Actions</b></p> <p><span>Identify well-known risks for your Azure tenants, remediate those risks, and track your progress using Secure Score</span></p> <p><b>Learn More</b></p> <p><a href="https://docs.microsoft.com/en-us/azure/architecture/framework/Security/governance#discover-and-remediate-common-risks" target="_blank"><span>Discover and remediate common risks</span></a><span /></p>
test
discover and remediate common risks to improve secure score in azure security center why consider this secure score is a snapshot of your security posture relative to microsoft recommendations identifying and remediating common security hygiene risks significantly reduces overall risk to the organization by increasing cost to attackers context when routine and well established attack vectors are mitigated attackers are forced to acquire and use advanced or untested attack methods azure secure score in azure security center is one method to easily identify and address well known issues in azure suggested actions identify well known risks for your azure tenants remediate those risks and track your progress using secure score learn more discover and remediate common risks
1
80,447
10,013,920,848
IssuesEvent
2019-07-15 16:13:42
Kotlin/kotlinx.coroutines
https://api.github.com/repos/Kotlin/kotlinx.coroutines
opened
ReceiveChannel.asFlow operator
design flow
A missing piece of the Flow puzzle is a connection between regular `ReceiveChannel` and `Flow` (note that we already have a pretty non-obvious `ReceieveChannel.asFlux`) with a proper migration path for channel operators. For example, all operators on top of the channels are deprecated, but users have no **clear** migration path from their channel operators sequence to flow. The main focus here is to provide clear consumption semantics and decide whether we want to introduce such primitive at all
1.0
ReceiveChannel.asFlow operator - A missing piece of the Flow puzzle is a connection between regular `ReceiveChannel` and `Flow` (note that we already have a pretty non-obvious `ReceieveChannel.asFlux`) with a proper migration path for channel operators. For example, all operators on top of the channels are deprecated, but users have no **clear** migration path from their channel operators sequence to flow. The main focus here is to provide clear consumption semantics and decide whether we want to introduce such primitive at all
non_test
receivechannel asflow operator a missing piece of the flow puzzle is a connection between regular receivechannel and flow note that we already have a pretty non obvious receievechannel asflux with a proper migration path for channel operators for example all operators on top of the channels are deprecated but users have no clear migration path from their channel operators sequence to flow the main focus here is to provide clear consumption semantics and decide whether we want to introduce such primitive at all
0
200,199
15,093,543,901
IssuesEvent
2021-02-07 01:09:33
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Out of order tags across active branches are causing tests to fail
kind/failing-test priority/critical-urgent sig/release triage/accepted
_Note: this is the same issue we encountered in https://github.com/kubernetes/kubernetes/issues/86182._ For the `v1.20.0` release, as well as the subsequent patch releases for `v1.17.15` / `v1.18.13` / `v1.19.5`, the release version was tagged either on the same or a later commit than the `rc.0` tag, causing subsequent builds to fail the version regex check in `get-kube.sh`. As summarized in @justaugustus [email to k-dev](https://groups.google.com/g/kubernetes-dev/c/angCw-zHejE/discussion) the last time this happened: > We use a combination of git describe and regexes (of course) to determine the tag and whether or not the tag is a CI tag. As the tags landed on the same commit, our git describe was ambiguous and picked up something in the form of v1.y.z-<number-of-commits-past-tag>+<commit-ish>, which makes our CI version regex unhappy. This will likely need to be solved by manually tagging new versions. This issue thread will be updated with progress towards resolution. An example test failure: https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-cos-k8sbeta-default/1336844198506139648 _Very long_ slack thread from last time this happened with rationale behind doing manual tag and push: https://kubernetes.slack.com/archives/CJH2GBF7Y/p1576268807016400 /cc @kubernetes/release-managers /assign /assign @justaugustus /priority critical-urgent
1.0
Out of order tags across active branches are causing tests to fail - _Note: this is the same issue we encountered in https://github.com/kubernetes/kubernetes/issues/86182._ For the `v1.20.0` release, as well as the subsequent patch releases for `v1.17.15` / `v1.18.13` / `v1.19.5`, the release version was tagged either on the same or a later commit than the `rc.0` tag, causing subsequent builds to fail the version regex check in `get-kube.sh`. As summarized in @justaugustus [email to k-dev](https://groups.google.com/g/kubernetes-dev/c/angCw-zHejE/discussion) the last time this happened: > We use a combination of git describe and regexes (of course) to determine the tag and whether or not the tag is a CI tag. As the tags landed on the same commit, our git describe was ambiguous and picked up something in the form of v1.y.z-<number-of-commits-past-tag>+<commit-ish>, which makes our CI version regex unhappy. This will likely need to be solved by manually tagging new versions. This issue thread will be updated with progress towards resolution. An example test failure: https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-cos-k8sbeta-default/1336844198506139648 _Very long_ slack thread from last time this happened with rationale behind doing manual tag and push: https://kubernetes.slack.com/archives/CJH2GBF7Y/p1576268807016400 /cc @kubernetes/release-managers /assign /assign @justaugustus /priority critical-urgent
test
out of order tags across active branches are causing tests to fail note this is the same issue we encountered in for the release as well as the subsequent patch releases for the release version was tagged either on the same or a later commit than the rc tag causing subsequent builds to fail the version regex check in get kube sh as summarized in justaugustus the last time this happened we use a combination of git describe and regexes of course to determine the tag and whether or not the tag is a ci tag as the tags landed on the same commit our git describe was ambiguous and picked up something in the form of y z which makes our ci version regex unhappy this will likely need to be solved by manually tagging new versions this issue thread will be updated with progress towards resolution an example test failure very long slack thread from last time this happened with rationale behind doing manual tag and push cc kubernetes release managers assign assign justaugustus priority critical urgent
1
233,091
18,947,335,290
IssuesEvent
2021-11-18 11:37:27
hyperledger-labs/go-perun
https://api.github.com/repos/hyperledger-labs/go-perun
opened
[sim/wallet] Address.String() length
good first issue testing-only
### Location `[backend/sim/wallet] Address.String()` function Function only returns the first 4 bytes as string for easier comparability in test output. ### Problem >I am not really convinced that shortening the string representation is a good idea. Your debug sessions may be easier to read, but on the other hand it may create confusion. In any case, this function could be a one-liner: return fmt.Sprintf("0x%x...", a.X.Bytes()[:length]). from @matthiasgeihs https://github.com/hyperledger-labs/go-perun/pull/260#discussion_r751680877 ### Proposal Change to full length.
1.0
[sim/wallet] Address.String() length - ### Location `[backend/sim/wallet] Address.String()` function Function only returns the first 4 bytes as string for easier comparability in test output. ### Problem >I am not really convinced that shortening the string representation is a good idea. Your debug sessions may be easier to read, but on the other hand it may create confusion. In any case, this function could be a one-liner: return fmt.Sprintf("0x%x...", a.X.Bytes()[:length]). from @matthiasgeihs https://github.com/hyperledger-labs/go-perun/pull/260#discussion_r751680877 ### Proposal Change to full length.
test
address string length location address string function function only returns the first bytes as string for easier comparability in test output problem i am not really convinced that shortening the string representation is a good idea your debug sessions may be easier to read but on the other hand it may create confusion in any case this function could be a one liner return fmt sprintf x a x bytes from matthiasgeihs proposal change to full length
1
93,788
15,932,116,301
IssuesEvent
2021-04-14 05:05:02
samq-wsdemo/SecurityShepherd
https://api.github.com/repos/samq-wsdemo/SecurityShepherd
opened
CVE-2020-2933 (Low) detected in mysql-connector-java-5.1.24.jar
security vulnerability
## CVE-2020-2933 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.24.jar</b></p></summary> <p>MySQL JDBC Type 4 driver</p> <p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p> <p>Path to dependency file: SecurityShepherd/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/mysql/mysql-connector-java/5.1.24/mysql-connector-java-5.1.24.jar,SecurityShepherd/target/owaspSecurityShepherd/WEB-INF/lib/mysql-connector-java-5.1.24.jar</p> <p> Dependency Hierarchy: - :x: **mysql-connector-java-5.1.24.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-wsdemo/SecurityShepherd/commit/00289f7750953c5aab43147df68b8dcf4063441b">00289f7750953c5aab43147df68b8dcf4063441b</a></p> <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 5.1.48 and prior. Difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 2.2 (Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:H/UI:N/S:U/C:N/I:N/A:L). <p>Publish Date: 2020-04-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2933>CVE-2020-2933</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://docs.oracle.com/javase/7/docs/api/javax/xml/XMLConstants.html#FEATURE_SECURE_PROCESSING">https://docs.oracle.com/javase/7/docs/api/javax/xml/XMLConstants.html#FEATURE_SECURE_PROCESSING</a></p> <p>Release Date: 2020-04-15</p> <p>Fix Resolution: mysql:mysql-connector-java:5.1.49</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.24","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.24","isMinimumFixVersionAvailable":true,"minimumFixVersion":"mysql:mysql-connector-java:5.1.49"}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-2933","vulnerabilityDetails":"Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 5.1.48 and prior. Difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 2.2 (Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:H/UI:N/S:U/C:N/I:N/A:L).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2933","cvss3Severity":"low","cvss3Score":"2.2","cvss3Metrics":{"A":"Low","AC":"High","PR":"High","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-2933 (Low) detected in mysql-connector-java-5.1.24.jar - ## CVE-2020-2933 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.24.jar</b></p></summary> <p>MySQL JDBC Type 4 driver</p> <p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p> <p>Path to dependency file: SecurityShepherd/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/mysql/mysql-connector-java/5.1.24/mysql-connector-java-5.1.24.jar,SecurityShepherd/target/owaspSecurityShepherd/WEB-INF/lib/mysql-connector-java-5.1.24.jar</p> <p> Dependency Hierarchy: - :x: **mysql-connector-java-5.1.24.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-wsdemo/SecurityShepherd/commit/00289f7750953c5aab43147df68b8dcf4063441b">00289f7750953c5aab43147df68b8dcf4063441b</a></p> <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 5.1.48 and prior. Difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 2.2 (Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:H/UI:N/S:U/C:N/I:N/A:L). <p>Publish Date: 2020-04-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2933>CVE-2020-2933</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://docs.oracle.com/javase/7/docs/api/javax/xml/XMLConstants.html#FEATURE_SECURE_PROCESSING">https://docs.oracle.com/javase/7/docs/api/javax/xml/XMLConstants.html#FEATURE_SECURE_PROCESSING</a></p> <p>Release Date: 2020-04-15</p> <p>Fix Resolution: mysql:mysql-connector-java:5.1.49</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.24","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.24","isMinimumFixVersionAvailable":true,"minimumFixVersion":"mysql:mysql-connector-java:5.1.49"}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-2933","vulnerabilityDetails":"Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 5.1.48 and prior. Difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 2.2 (Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:H/UI:N/S:U/C:N/I:N/A:L).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2933","cvss3Severity":"low","cvss3Score":"2.2","cvss3Metrics":{"A":"Low","AC":"High","PR":"High","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_test
cve low detected in mysql connector java jar cve low severity vulnerability vulnerable library mysql connector java jar mysql jdbc type driver library home page a href path to dependency file securityshepherd pom xml path to vulnerable library canner repository mysql mysql connector java mysql connector java jar securityshepherd target owaspsecurityshepherd web inf lib mysql connector java jar dependency hierarchy x mysql connector java jar vulnerable library found in head commit a href found in base branch dev vulnerability details vulnerability in the mysql connectors product of oracle mysql component connector j supported versions that are affected are and prior difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise mysql connectors successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service partial dos of mysql connectors cvss base score availability impacts cvss vector cvss av n ac h pr h ui n s u c n i n a l publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution mysql mysql connector java rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree mysql mysql connector java isminimumfixversionavailable true minimumfixversion mysql mysql connector java basebranches vulnerabilityidentifier cve vulnerabilitydetails vulnerability in the mysql connectors product of oracle mysql component connector j supported versions that are affected are and prior difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise mysql connectors successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service partial dos of mysql connectors cvss base score availability impacts cvss vector cvss av n ac h pr h ui n s u c n i n a l vulnerabilityurl
0
164,154
6,219,864,529
IssuesEvent
2017-07-09 17:31:03
tferreira/piggydime
https://api.github.com/repos/tferreira/piggydime
closed
Dockerize the app
deployment priority - medium
For easier deployment and best support of a wide range of operating systems, it would be necessary to create mutiple Docker images: - The python server (maybe using uWSGI and Nginx) - The React app - A MySQL DB (using volumes to keep the datas) Using docker-compose is probably the best way to achieve this. Also see: [https://github.com/geerlingguy/docker-examples/tree/master/flask](https://github.com/geerlingguy/docker-examples/tree/master/flask)
1.0
Dockerize the app - For easier deployment and best support of a wide range of operating systems, it would be necessary to create mutiple Docker images: - The python server (maybe using uWSGI and Nginx) - The React app - A MySQL DB (using volumes to keep the datas) Using docker-compose is probably the best way to achieve this. Also see: [https://github.com/geerlingguy/docker-examples/tree/master/flask](https://github.com/geerlingguy/docker-examples/tree/master/flask)
non_test
dockerize the app for easier deployment and best support of a wide range of operating systems it would be necessary to create mutiple docker images the python server maybe using uwsgi and nginx the react app a mysql db using volumes to keep the datas using docker compose is probably the best way to achieve this also see
0
380,786
26,431,097,649
IssuesEvent
2023-01-14 20:36:58
Luanftg/codigo-do-futuro-api-dotnet
https://api.github.com/repos/Luanftg/codigo-do-futuro-api-dotnet
closed
Criação da Conta Google
documentation
### 9-Registrar a chave para utilizar a Api do Google Maps - [ ] Criar conta no google Cloud - [ ] Cria Key para APIs do google maps - [ ] Configurar os acessos das APIs ### Texto Auxiliar [Documentação de Referência APi Google Maps](https://developers.google.com/maps/documentation/maps-static/start?hl=pt-br#Addresses)
1.0
Criação da Conta Google - ### 9-Registrar a chave para utilizar a Api do Google Maps - [ ] Criar conta no google Cloud - [ ] Cria Key para APIs do google maps - [ ] Configurar os acessos das APIs ### Texto Auxiliar [Documentação de Referência APi Google Maps](https://developers.google.com/maps/documentation/maps-static/start?hl=pt-br#Addresses)
non_test
criação da conta google registrar a chave para utilizar a api do google maps criar conta no google cloud cria key para apis do google maps configurar os acessos das apis texto auxiliar
0
68,322
7,093,010,315
IssuesEvent
2018-01-12 18:43:59
truetandem/e-QIP-prototype
https://api.github.com/repos/truetandem/e-QIP-prototype
opened
Safari - Foreign activities - review screen
cross browser testing - safari
Foreign activities - review screen dropped user midway down the screen instead of at top. ![screen shot 2018-01-12 at 1 41 52 pm](https://user-images.githubusercontent.com/19935974/34890058-7735cb44-f79e-11e7-92b6-f51118ad8bf6.jpg)
1.0
Safari - Foreign activities - review screen - Foreign activities - review screen dropped user midway down the screen instead of at top. ![screen shot 2018-01-12 at 1 41 52 pm](https://user-images.githubusercontent.com/19935974/34890058-7735cb44-f79e-11e7-92b6-f51118ad8bf6.jpg)
test
safari foreign activities review screen foreign activities review screen dropped user midway down the screen instead of at top
1
87,569
8,100,760,973
IssuesEvent
2018-08-12 03:41:53
codercom/bugs
https://api.github.com/repos/codercom/bugs
closed
Project data not removed when the project is deleted.
bug testing tracking
**Describe the bug** Data is not removed when a project is deleted. **To Reproduce** Steps to reproduce the behavior: 1. Create a project. 2. Create a file in the project. 3. Delete the project. 4. Create a new project with the exact same name and the files will still be there. **Expected behavior** A new project would be created without old data as it is supposed to delete when the project is deleted. **Desktop (please complete the following information):** - OS: Windows 10 - Browser: Google Chrome - Version: 67.0.3396.99 (Official Build) (64-bit)
1.0
Project data not removed when the project is deleted. - **Describe the bug** Data is not removed when a project is deleted. **To Reproduce** Steps to reproduce the behavior: 1. Create a project. 2. Create a file in the project. 3. Delete the project. 4. Create a new project with the exact same name and the files will still be there. **Expected behavior** A new project would be created without old data as it is supposed to delete when the project is deleted. **Desktop (please complete the following information):** - OS: Windows 10 - Browser: Google Chrome - Version: 67.0.3396.99 (Official Build) (64-bit)
test
project data not removed when the project is deleted describe the bug data is not removed when a project is deleted to reproduce steps to reproduce the behavior create a project create a file in the project delete the project create a new project with the exact same name and the files will still be there expected behavior a new project would be created without old data as it is supposed to delete when the project is deleted desktop please complete the following information os windows browser google chrome version official build bit
1
31,830
6,643,248,573
IssuesEvent
2017-09-27 10:31:57
contao/core-bundle
https://api.github.com/repos/contao/core-bundle
closed
Important part cannot be set within picker overlay
defect
If you go to the file manager within the picker overlay and edit an image, you are not able to set the important part with your cursor anymore. Instead, only the input fields to manually set the important part appear. Possibly related to #1092 ?
1.0
Important part cannot be set within picker overlay - If you go to the file manager within the picker overlay and edit an image, you are not able to set the important part with your cursor anymore. Instead, only the input fields to manually set the important part appear. Possibly related to #1092 ?
non_test
important part cannot be set within picker overlay if you go to the file manager within the picker overlay and edit an image you are not able to set the important part with your cursor anymore instead only the input fields to manually set the important part appear possibly related to
0
699,548
24,020,807,536
IssuesEvent
2022-09-15 07:29:11
PrefectHQ/prefect
https://api.github.com/repos/PrefectHQ/prefect
closed
Only one default parameter is deployed when multiple are specified (in deployment.py)
bug status:backlog priority:high component:deployment
### First check - [X] I added a descriptive title to this issue. - [X] I used the GitHub search to find a similar issue and didn't find it. - [X] I searched the Prefect documentation for this issue. - [X] I checked that this issue is related to Prefect and not one of its dependencies. ### Bug summary I am scheduling a dbt flow with Prefect. I tried to specify two parameters in my deployment.py file, but only the first of them is stored in the database after deployment. I have not tried deployment via CLI/yaml yet, but though this is already worth mentioning. ### Reproduction ```python # deployment.py from dbt_debug import dbt_debug_flow from prefect.deployments import Deployment deployment = Deployment.build_from_flow( flow=dbt_debug_flow, description ="Testing a prefect-based dbt workflow.", name="dbt-debug-test", parameters={'file_path': 'C:/Users/Public/File', 'project_dir':'C:/Users/User/Documents/Git/Project/dbt_dev'}, infra_overrides={"env": {"PREFECT_LOGGING_LEVEL": "DEBUG"}}, work_queue_name="test", ) if __name__ == "__main__": deployment.apply() # deploy via python deployment.py ``` ### Error # On Prefect Orion ![image](https://user-images.githubusercontent.com/31634583/190200767-2055bc90-a4fe-4c14-921c-6066850e9a92.png) ``` Flow run 'cuddly-lyrebird' - Validation of flow parameters failed with error: SignatureMismatchError("Function expects parameters ['file_path'] but was provided with parameters ['file_path', 'project_dir']") ``` ### Versions ``` Version: 2.4.0 API version: 0.8.0 Python version: 3.10.7 Git commit: 513639e8 Built: Tue, Sep 13, 2022 2:15 PM OS/Arch: win32/AMD64 Profile: default Server type: ephemeral Server: Database: sqlite SQLite version: 3.37.2 ``` ### Additional context _No response_
1.0
Only one default parameter is deployed when multiple are specified (in deployment.py) - ### First check - [X] I added a descriptive title to this issue. - [X] I used the GitHub search to find a similar issue and didn't find it. - [X] I searched the Prefect documentation for this issue. - [X] I checked that this issue is related to Prefect and not one of its dependencies. ### Bug summary I am scheduling a dbt flow with Prefect. I tried to specify two parameters in my deployment.py file, but only the first of them is stored in the database after deployment. I have not tried deployment via CLI/yaml yet, but though this is already worth mentioning. ### Reproduction ```python # deployment.py from dbt_debug import dbt_debug_flow from prefect.deployments import Deployment deployment = Deployment.build_from_flow( flow=dbt_debug_flow, description ="Testing a prefect-based dbt workflow.", name="dbt-debug-test", parameters={'file_path': 'C:/Users/Public/File', 'project_dir':'C:/Users/User/Documents/Git/Project/dbt_dev'}, infra_overrides={"env": {"PREFECT_LOGGING_LEVEL": "DEBUG"}}, work_queue_name="test", ) if __name__ == "__main__": deployment.apply() # deploy via python deployment.py ``` ### Error # On Prefect Orion ![image](https://user-images.githubusercontent.com/31634583/190200767-2055bc90-a4fe-4c14-921c-6066850e9a92.png) ``` Flow run 'cuddly-lyrebird' - Validation of flow parameters failed with error: SignatureMismatchError("Function expects parameters ['file_path'] but was provided with parameters ['file_path', 'project_dir']") ``` ### Versions ``` Version: 2.4.0 API version: 0.8.0 Python version: 3.10.7 Git commit: 513639e8 Built: Tue, Sep 13, 2022 2:15 PM OS/Arch: win32/AMD64 Profile: default Server type: ephemeral Server: Database: sqlite SQLite version: 3.37.2 ``` ### Additional context _No response_
non_test
only one default parameter is deployed when multiple are specified in deployment py first check i added a descriptive title to this issue i used the github search to find a similar issue and didn t find it i searched the prefect documentation for this issue i checked that this issue is related to prefect and not one of its dependencies bug summary i am scheduling a dbt flow with prefect i tried to specify two parameters in my deployment py file but only the first of them is stored in the database after deployment i have not tried deployment via cli yaml yet but though this is already worth mentioning reproduction python deployment py from dbt debug import dbt debug flow from prefect deployments import deployment deployment deployment build from flow flow dbt debug flow description testing a prefect based dbt workflow name dbt debug test parameters file path c users public file project dir c users user documents git project dbt dev infra overrides env prefect logging level debug work queue name test if name main deployment apply deploy via python deployment py error on prefect orion flow run cuddly lyrebird validation of flow parameters failed with error signaturemismatcherror function expects parameters but was provided with parameters versions version api version python version git commit built tue sep pm os arch profile default server type ephemeral server database sqlite sqlite version additional context no response
0
807,332
29,995,841,602
IssuesEvent
2023-06-26 05:24:13
KingSupernova31/RulesGuru
https://api.github.com/repos/KingSupernova31/RulesGuru
opened
Talk to Kyle Ryc about their phone issues
bug medium priority
Canadian L2 Kyle Ryc was telling me about several issues they were having with RG on their phone. I or someone else needs to reach out to them, get a list of the issues, and fix them.
1.0
Talk to Kyle Ryc about their phone issues - Canadian L2 Kyle Ryc was telling me about several issues they were having with RG on their phone. I or someone else needs to reach out to them, get a list of the issues, and fix them.
non_test
talk to kyle ryc about their phone issues canadian kyle ryc was telling me about several issues they were having with rg on their phone i or someone else needs to reach out to them get a list of the issues and fix them
0
57,034
3,081,232,950
IssuesEvent
2015-08-22 14:21:05
bitfighter/bitfighter
https://api.github.com/repos/bitfighter/bitfighter
closed
Multi-line text items
020 bug imported Priority-Medium
_From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on October 23, 2013 07:40:11_ Allow users to enter multi-line text items, with line breaks specified with \n. _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=265_
1.0
Multi-line text items - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on October 23, 2013 07:40:11_ Allow users to enter multi-line text items, with line breaks specified with \n. _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=265_
non_test
multi line text items from on october allow users to enter multi line text items with line breaks specified with n original issue
0
424,205
12,307,575,547
IssuesEvent
2020-05-12 05:03:21
googleapis/java-billing
https://api.github.com/repos/googleapis/java-billing
opened
Synthesis failed for java-billing
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate java-billing. :broken_heart: Here's the output from running `synth.py`: ``` 2020-05-11 21:52:44,383 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/java-billing 2020-05-11 21:52:44,900 autosynth > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2020-05-11 21:52:44,905 autosynth > Running: git config user.name yoshi-automation 2020-05-11 21:52:44,910 autosynth > Running: git config user.email yoshi-automation@google.com 2020-05-11 21:52:44,915 autosynth > Running: git config push.default simple 2020-05-11 21:52:44,920 autosynth > Running: git branch -f autosynth 2020-05-11 21:52:44,926 autosynth > Running: git checkout autosynth 2020-05-11 21:52:44,939 autosynth > autosynth flags: 2020-05-11 21:52:44,940 autosynth > AUTOSYNTH_MULTIPLE_COMMITS: true 2020-05-11 21:52:44,940 autosynth > AUTOSYNTH_MULTIPLE_PRS: true 2020-05-11 21:52:44,941 autosynth > Running: git rev-parse --show-toplevel 2020-05-11 21:52:44,946 autosynth > Running: git log -1 --pretty=%H 2020-05-11 21:52:44,951 autosynth > Running: git remote get-url origin 2020-05-11 21:52:44,956 autosynth > Running: git clone --single-branch https://github.com/googleapis/googleapis.git -- /tmpfs/tmp/tmpaaavgl1n/googleapis 2020-05-11 21:52:46,450 autosynth > Running: git log d741cd976975c745d0199987aff0e908b8352992..HEAD --pretty=%H --no-decorate 2020-05-11 21:52:46,457 autosynth > Running: git clone --single-branch https://github.com/googleapis/synthtool.git -- /tmpfs/tmp/tmpaaavgl1n/synthtool 2020-05-11 21:52:47,010 autosynth > Running: git log f8a9933e5e98202b04ef427f28d1d79221190fa4..HEAD --pretty=%H --no-decorate 2020-05-11 21:52:47,017 autosynth > Running: git log -1 --pretty=%at d741cd976975c745d0199987aff0e908b8352992 2020-05-11 21:52:47,022 autosynth > Running: git log -1 --pretty=%at e4f4b23e07315492b533746e6a9255a1e6b3e748 2020-05-11 21:52:47,028 autosynth > Running: git log -1 --pretty=%at ee4ea76504aa60c2bff9b7c11269c155d8c21e0d 2020-05-11 21:52:47,033 autosynth > Running: git log -1 --pretty=%at 6114e2d30c0792522720d1c2cf8f5e111be7545b 2020-05-11 21:52:47,038 autosynth > Running: git log -1 --pretty=%at af71ea52236b05ce73253f6bc935fcf35c866664 2020-05-11 21:52:47,044 autosynth > Running: git log -1 --pretty=%at c451a4457f38e5582e9794a0094c773355a9eaad 2020-05-11 21:52:47,050 autosynth > Running: git log -1 --pretty=%at 9af7212bc2cba8a7184f032e1ab6662ccf604b3a 2020-05-11 21:52:47,055 autosynth > Running: git log -1 --pretty=%at ac1f904979899bc598fb0567cf7a367f08b40b43 2020-05-11 21:52:47,061 autosynth > Running: git log -1 --pretty=%at fe8b17aff5b419ffb418a77228fa049190488233 2020-05-11 21:52:47,067 autosynth > Running: git log -1 --pretty=%at fd35dc6865fa9f47f283d5a6c37c4901162007cf 2020-05-11 21:52:47,072 autosynth > Running: git log -1 --pretty=%at 96c0db42dbcb3bb0116cfa55c054dd139ffb3912 2020-05-11 21:52:47,077 autosynth > Running: git log -1 --pretty=%at b8c7c4e8da56547fc32bee9b563d306fadabdb6f 2020-05-11 21:52:47,082 autosynth > Running: git log -1 --pretty=%at b21bc9b46e3263cd4ebe50f60c61589e3fe797f2 2020-05-11 21:52:47,087 autosynth > Running: git log -1 --pretty=%at 41d5dee23527df3fc9127838a027df393d79ff3c 2020-05-11 21:52:47,092 autosynth > Running: git log -1 --pretty=%at 801c3fb47b1a533989a1fe5af4aedf02af5f2eae 2020-05-11 21:52:47,097 autosynth > Running: git log -1 --pretty=%at 1692c3d51ee7d4366ce81892df1b9ca9846480d3 2020-05-11 21:52:47,102 autosynth > Running: git log -1 --pretty=%at 1df336ab4a989e7b0bc363baa9fc91eb66b0625e 2020-05-11 21:52:47,107 autosynth > Running: git log -1 --pretty=%at 1e92c343790ae63f88817afde60bd2d5bc775f36 2020-05-11 21:52:47,111 autosynth > Running: git log -1 --pretty=%at cf18ab5200e41e9d0cfef88c6c4bcdd394faf1ca 2020-05-11 21:52:47,116 autosynth > Running: git log -1 --pretty=%at ab869e90c6ced0ef30fcc8e8fbe8e527a695d3cb 2020-05-11 21:52:47,121 autosynth > Running: git log -1 --pretty=%at 13736ead90e0fa65103eb03fea775a1c010564af 2020-05-11 21:52:47,126 autosynth > Running: git log -1 --pretty=%at f9d76ac088f5d444b16d5e17fe58a9eea57b7a99 2020-05-11 21:52:47,131 autosynth > Running: git log -1 --pretty=%at 1bd77e8ce6f953ac641af7966d0c52646afc16a8 2020-05-11 21:52:47,136 autosynth > Running: git log -1 --pretty=%at 3824f547aa44df459580615c73cabb30a2a78ee0 2020-05-11 21:52:47,141 autosynth > Running: git log -1 --pretty=%at 4d61e1cb40184a7ad63ef37b1813f6608718674a 2020-05-11 21:52:47,147 autosynth > Running: git log -1 --pretty=%at 9ea892ebe8a7cd66da9bd62eb7086b8dc2ddd559 2020-05-11 21:52:47,152 autosynth > Running: git log -1 --pretty=%at 0feb08065d91d6deebffef3019751f0619a72140 2020-05-11 21:52:47,157 autosynth > Running: git log -1 --pretty=%at 789fd7e0680b04a589447139811d55c5d1f34afe 2020-05-11 21:52:47,163 autosynth > Running: git log -1 --pretty=%at 4efd71279f0781579ced60c374336ce1b8b8cea8 2020-05-11 21:52:47,168 autosynth > Running: git log -1 --pretty=%at a34f4eb96a94bfd5b3e7f8de253db28d0a3e9237 2020-05-11 21:52:47,173 autosynth > Running: git log -1 --pretty=%at f561f57eea013327be27d8f6799b3cc2d18ee804 2020-05-11 21:52:47,177 autosynth > Running: git log -1 --pretty=%at c9942c00e8f35b5b37cd589ef7a0323aeda264e2 2020-05-11 21:52:47,182 autosynth > Running: git log -1 --pretty=%at e2a59ff2f861b1b92b2f15b242780dccddd77ab9 2020-05-11 21:52:47,188 autosynth > Running: git log -1 --pretty=%at df955af5d9508cf9215de6f1fbdad033d5b8fef6 2020-05-11 21:52:47,193 autosynth > Running: git log -1 --pretty=%at dea1f52034d188ca371c3a70c716ca2a6b3f2aa1 2020-05-11 21:52:47,198 autosynth > Running: git log -1 --pretty=%at 35672e7936b923eff048d0ee1529faa163cc81b8 2020-05-11 21:52:47,203 autosynth > Running: git log -1 --pretty=%at 26523a96798ce1a6caa1b3c912119059cfcc98a7 2020-05-11 21:52:47,208 autosynth > Running: git log -1 --pretty=%at 04eff659376f40a7111bf0e8e49fbb38bc7739f7 2020-05-11 21:52:47,213 autosynth > Running: git log -1 --pretty=%at 3028060618e8024af9a32b3ab3456c160091ecb7 2020-05-11 21:52:47,218 autosynth > Running: git log -1 --pretty=%at b5aa7d668d9513cf0d305c8b23262ee0ea9b850c 2020-05-11 21:52:47,223 autosynth > Running: git log -1 --pretty=%at c0a072159eda3c7d5a98201d6500feaf78a18073 2020-05-11 21:52:47,228 autosynth > Running: git log -1 --pretty=%at fea22b1d9f27f86ef355c1d0dba00e0791a08a19 2020-05-11 21:52:47,233 autosynth > Running: git log -1 --pretty=%at 80dac49dd5f407b14f15ac86b4922967b5dd6508 2020-05-11 21:52:47,238 autosynth > Running: git log -1 --pretty=%at b8495354535dc1370377cbb15a3ef7a07ec3a114 2020-05-11 21:52:47,243 autosynth > Running: git log -1 --pretty=%at ab5d07c2883026eec395d73ff59a17f4ad0153db 2020-05-11 21:52:47,249 autosynth > Running: git log -1 --pretty=%at 160e3187dc93265ed635ad68e12c483c8ab4f7be 2020-05-11 21:52:47,254 autosynth > Running: git log -1 --pretty=%at 629e6bea8418ad4a5d081ab9eb48bb327ad368e1 2020-05-11 21:52:47,259 autosynth > Running: git log -1 --pretty=%at c3f3cdeda70f60d808eaaae8fea44d306b6d5cc2 2020-05-11 21:52:47,264 autosynth > Running: git log -1 --pretty=%at cdf13efacdea0649e940452f9c5d320b93735974 2020-05-11 21:52:47,269 autosynth > Running: git log -1 --pretty=%at 101d31acd73076c52d78e18322be01f3debe8cb5 2020-05-11 21:52:47,274 autosynth > Running: git log -1 --pretty=%at cab24c7c9b570e2b483e4f742a7a4a111167238b 2020-05-11 21:52:47,279 autosynth > Running: git log -1 --pretty=%at 6e0ac551e057e20a6c41ab0b21ffa9623e463326 2020-05-11 21:52:47,285 autosynth > Running: git log -1 --pretty=%at 69b4dc56d21394e7c39bc62fe11c76a88aa5511f 2020-05-11 21:52:47,290 autosynth > Running: git log -1 --pretty=%at bcc476396e799806d3355e87246c6becf6250a70 2020-05-11 21:52:47,295 autosynth > Running: git log -1 --pretty=%at b5e9b0b09cc15c263f07da92fa64db31b2e4dd07 2020-05-11 21:52:47,300 autosynth > Running: git log -1 --pretty=%at c09ec5156db66d326cce56ed3a2a8092e02c3abd 2020-05-11 21:52:47,306 autosynth > Running: git log -1 --pretty=%at 42ee97c1b93a0e3759bbba3013da309f670a90ab 2020-05-11 21:52:47,310 autosynth > Running: git log -1 --pretty=%at 45b363d3ab4692115cc5018eb26dcdaa49b320a8 2020-05-11 21:52:47,315 autosynth > Running: git log -1 --pretty=%at 012c553112fcceb7b69822e24335575bf99ec62b 2020-05-11 21:52:47,320 autosynth > Running: git log -1 --pretty=%at 2bf9f22da3f101aeb536e1e88fe036323b8e9188 2020-05-11 21:52:47,326 autosynth > Running: git log -1 --pretty=%at 862cd840dbb0101a0630913217bf8b6d4101c84c 2020-05-11 21:52:47,331 autosynth > Running: git log -1 --pretty=%at b3bcfd059ec684445a38db693f6e36f14a1d2618 2020-05-11 21:52:47,336 autosynth > Running: git log -1 --pretty=%at 1d520eaa7bbd8b40f53942ff03b5496fff887b53 2020-05-11 21:52:47,341 autosynth > Running: git log -1 --pretty=%at 0d7c3565dd942d0cc41b16e895046486ae29e44e 2020-05-11 21:52:47,347 autosynth > Running: git log -1 --pretty=%at d14bf59a446c14ef16e9931ebfc8e63ab549bf07 2020-05-11 21:52:47,352 autosynth > Running: git log -1 --pretty=%at 28e76243c23cc282efbb288cb558c174e3e5e9ee 2020-05-11 21:52:47,357 autosynth > Running: git log -1 --pretty=%at 5dca7f1079a4c127245c0f1dbb8293b434fbd0c8 2020-05-11 21:52:47,363 autosynth > Running: git log -1 --pretty=%at c2bfceedfc9f01356229bd206ad25e250aa9551c 2020-05-11 21:52:47,368 autosynth > Running: git log -1 --pretty=%at 371206424e6db4df2cb3734bb1013dcaac9dbe4b 2020-05-11 21:52:47,373 autosynth > Running: git log -1 --pretty=%at ff17eefa184f6985e5d9a8ee3e725da95a06cd23 2020-05-11 21:52:47,379 autosynth > Running: git log -1 --pretty=%at c284e4f849e4b085a297d336cef8721ab69ba013 2020-05-11 21:52:47,384 autosynth > Running: git log -1 --pretty=%at 756b174de4a122461993c1c583345533d819936d 2020-05-11 21:52:47,390 autosynth > Running: git log -1 --pretty=%at 3f5f8a2258c6a41f9fbf7b80acbca631dda0a952 2020-05-11 21:52:47,395 autosynth > Running: git log -1 --pretty=%at 7e1c7603c8cb2f05ef29842405654ff8a9d3d108 2020-05-11 21:52:47,400 autosynth > Running: git log -1 --pretty=%at f78da0d54c78aa8f66d52a448db1f7ec0e0fd591 2020-05-11 21:52:47,406 autosynth > Running: git log -1 --pretty=%at aaff764c185e18a6c73227357c3df5fa60fec85a 2020-05-11 21:52:47,412 autosynth > Running: git log -1 --pretty=%at ee9e8e4e67f06aba881392a1a2956fbe7a42d216 2020-05-11 21:52:47,418 autosynth > Running: git log -1 --pretty=%at 93b25607689c4bcdb3cb35c657fd3c16249dcd54 2020-05-11 21:52:47,423 autosynth > Running: git log -1 --pretty=%at ab3468cf63abc521be76c9eceb620d61c5ddd76a 2020-05-11 21:52:47,428 autosynth > Running: git log -1 --pretty=%at a737a68e214d1cf927457c996b73e2600b14a949 2020-05-11 21:52:47,433 autosynth > Running: git log -1 --pretty=%at 4ca990cb1a6771f11469b1fdf2fc578271703db0 2020-05-11 21:52:47,439 autosynth > Running: git log -1 --pretty=%at e0f9d9e1f9de890db765be46f45ca8490723e3eb 2020-05-11 21:52:47,445 autosynth > Running: git log -1 --pretty=%at 1b5a8d2bbb69c5a04db26bd546d2888e609c6bab 2020-05-11 21:52:47,450 autosynth > Running: git log -1 --pretty=%at 4bef0001ac7040431ea24b6187424fdec9c08b1b 2020-05-11 21:52:47,456 autosynth > Running: git log -1 --pretty=%at a3a0bf0f6291d69f2ff3df7fcd63d28ee20ac727 2020-05-11 21:52:47,461 autosynth > Running: git log -1 --pretty=%at 6dfd72d028a0d0a43764e060f7b15e004385c3a1 2020-05-11 21:52:47,466 autosynth > Running: git log -1 --pretty=%at 839fae42335ee1bb1e70767f3e6c51738683892b 2020-05-11 21:52:47,471 autosynth > Running: git log -1 --pretty=%at 30cfca094376e4904e32e71c838a81169fd4a2e2 2020-05-11 21:52:47,476 autosynth > Running: git log -1 --pretty=%at c08dcec05ce1c181bcdbce59cabba36e0e541ff6 2020-05-11 21:52:47,481 autosynth > Running: git log -1 --pretty=%at 2fc2caaacb15949c7f80426bfc7dafdd41dbc333 2020-05-11 21:52:47,486 autosynth > Running: git log -1 --pretty=%at 684dfea7decfeca7a7526ea96a8e9256694dd5d8 2020-05-11 21:52:47,491 autosynth > Running: git log -1 --pretty=%at fb8f62b6784f43faf4b64179c57ce4b4931b1a00 2020-05-11 21:52:47,495 autosynth > Running: git log -1 --pretty=%at d3cc7bbf41afdfd6677cc1ab0d53965625cc073b 2020-05-11 21:52:47,500 autosynth > Running: git log -1 --pretty=%at 83816bb3093686a28af2891db5b7506614a820b1 2020-05-11 21:52:47,505 autosynth > Running: git log -1 --pretty=%at 73d4b5d9a791f8b1ee63d439ffe909bb8ffa07f7 2020-05-11 21:52:47,510 autosynth > Running: git log -1 --pretty=%at 482e5206e05e4ba8cd79738f6fb7a521c8c23555 2020-05-11 21:52:47,516 autosynth > Running: git log -1 --pretty=%at 1019b0d9dbd999e1107f90247fcc478678d1105c 2020-05-11 21:52:47,521 autosynth > Running: git log -1 --pretty=%at 4186d3e5424edc088a0f4dbce78f02dd860f3de8 2020-05-11 21:52:47,526 autosynth > Running: git log -1 --pretty=%at aed11c01e52921613b9ee469c2d85f5f33175fb7 2020-05-11 21:52:47,531 autosynth > Running: git log -1 --pretty=%at edd3b80fb770548d6ad780105f1782de6ff73ea0 2020-05-11 21:52:47,536 autosynth > Running: git log -1 --pretty=%at f8a9933e5e98202b04ef427f28d1d79221190fa4 2020-05-11 21:52:47,541 autosynth > Running: git log -1 --pretty=%at 19465d3ec5e5acdb01521d8f3bddd311bcbee28d 2020-05-11 21:52:47,546 autosynth > Running: git log -1 --pretty=%at f5e4c17dc78a966dbf29961dd01f9bbd63e20a04 2020-05-11 21:52:47,551 autosynth > Running: git log -1 --pretty=%at 21ce80f9d4785dc75aa36d26944bddd5a1b7e25b 2020-05-11 21:52:47,556 autosynth > Running: git log -1 --pretty=%at 716f741f2d307b48cbe8a5bc3bc883571212344a 2020-05-11 21:52:47,561 autosynth > Running: git log -1 --pretty=%at 01b6f23d24b27878b48667ce597876d66b59780e 2020-05-11 21:52:47,566 autosynth > Running: git log -1 --pretty=%at 969a2340e74c73227e7c1638ed7650abcac22ee4 2020-05-11 21:52:47,571 autosynth > Running: git log -1 --pretty=%at 720b7bff0da03ed9e67cb2362f2d291241fa59ae 2020-05-11 21:52:47,576 autosynth > Running: git log -1 --pretty=%at 7c1a92d00de654db2fb440243129c61393a694e2 2020-05-11 21:52:47,581 autosynth > Running: git log -1 --pretty=%at 46fb62a3ebda62cc6c1c98e4f1eaf91965eab580 2020-05-11 21:52:47,586 autosynth > Running: git log -1 --pretty=%at 6b685a261479e27fd959dd7b1ff253d6cf595172 2020-05-11 21:52:47,591 autosynth > Running: git log -1 --pretty=%at 958a803ee0c43523fa3e6e0bb016276a69614c8e 2020-05-11 21:52:47,596 autosynth > Running: git log -1 --pretty=%at cdddf139b36000b3a7c65fd2a7781e253262359a 2020-05-11 21:52:47,601 autosynth > Running: git log -1 --pretty=%at 04cb397eb7590ea1e6c10a39fca3d8fe0fb3d256 2020-05-11 21:52:47,605 autosynth > Running: git log -1 --pretty=%at a2c1c4f89a5e220e7b39420ebea33623c7c72804 2020-05-11 21:52:47,610 autosynth > Running: git log -1 --pretty=%at ea3ae392fced25d7d6f5c7ef9c8a1f24768a0936 2020-05-11 21:52:47,614 autosynth > Running: git log -1 --pretty=%at 09c48461232ce929c34386259eb59018ad2d8eef 2020-05-11 21:52:47,619 autosynth > Running: git log -1 --pretty=%at ab883569eb0257bbf16a6d825fd018b3adde3912 2020-05-11 21:52:47,624 autosynth > Running: git log -1 --pretty=%at 756bc4dfc24e8bc4c5dd4116daa41a0440ebf5a0 2020-05-11 21:52:47,629 autosynth > Running: git log -1 --pretty=%at 558bb0d70fa98ea228483b44885704d8941a6a80 2020-05-11 21:52:47,634 autosynth > Running: git log -1 --pretty=%at be74d3e532faa47eb59f1a0eaebde0860d1d8ab4 2020-05-11 21:52:47,638 autosynth > Running: git log -1 --pretty=%at 4674113712c0c7ada19e6a8219d7963ff174b392 2020-05-11 21:52:47,643 autosynth > Running: git log -1 --pretty=%at 5bbfd095faedfe273819d266f21e402192a29041 2020-05-11 21:52:47,648 autosynth > Running: git log -1 --pretty=%at 4fa923bd3dafb91df8613accbe2230299cc5b98e 2020-05-11 21:52:47,652 autosynth > Running: git log -1 --pretty=%at 55cdc844877d97139f25004229842624a6a86a02 2020-05-11 21:52:47,657 autosynth > Running: git log -1 --pretty=%at 98c50772ec23295c64cf0d2ddf199ea52961fd4c 2020-05-11 21:52:47,662 autosynth > Forking toolbox 2020-05-11 21:52:47,662 autosynth > forking: autosynth-self 2020-05-11 21:52:47,663 autosynth > Running: git branch -f autosynth-self 2020-05-11 21:52:47,668 autosynth > forking: autosynth-googleapis 2020-05-11 21:52:47,669 autosynth > Running: git branch -f autosynth-googleapis 2020-05-11 21:52:47,674 autosynth > forking: autosynth-synthtool 2020-05-11 21:52:47,675 autosynth > Running: git branch -f autosynth-synthtool 2020-05-11 21:52:47,679 autosynth > Forked branch: autosynth-self 2020-05-11 21:52:47,997 autosynth > Running: git checkout autosynth-self 2020-05-11 21:52:48,012 autosynth > Building most recent versions 2020-05-11 21:52:48,013 autosynth > Running: git checkout 50afb0e301f97194bf58c78b0133606cf140784f 2020-05-11 21:52:48,020 autosynth > Running: git checkout d741cd976975c745d0199987aff0e908b8352992 2020-05-11 21:52:48,176 autosynth > Running: git checkout f8a9933e5e98202b04ef427f28d1d79221190fa4 2020-05-11 21:52:48,189 autosynth > Running: git branch -f autosynth-self-2 2020-05-11 21:52:48,196 autosynth > Running: git checkout autosynth-self-2 2020-05-11 21:52:48,204 autosynth > Running synthtool 2020-05-11 21:52:48,204 autosynth > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py -- 2020-05-11 21:52:48,204 autosynth > -> /tmpfs/src/github/synthtool/logs/googleapis/java-billing/self/2/sponge_log.log 2020-05-11 21:53:10,684 autosynth > Failed executing /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py -- 2020-05-11 21:53:10,685 autosynth > Synthesis failed 2020-05-11 21:53:10,688 autosynth > Running: git reset --hard HEAD 2020-05-11 21:53:10,817 autosynth > Running: git checkout autosynth-self 2020-05-11 21:53:10,829 autosynth > Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. 2020-05-11 21:53:10,830 autosynth > Building most recent versions 2020-05-11 21:53:10,831 autosynth > Running: git checkout 50afb0e301f97194bf58c78b0133606cf140784f 2020-05-11 21:53:10,839 autosynth > Running: git checkout edd3b80fb770548d6ad780105f1782de6ff73ea0 2020-05-11 21:53:11,242 autosynth > Running: git checkout 98c50772ec23295c64cf0d2ddf199ea52961fd4c 2020-05-11 21:53:11,363 autosynth > Running: git branch -f autosynth-126 2020-05-11 21:53:11,371 autosynth > Running: git checkout autosynth-126 2020-05-11 21:53:11,378 autosynth > Running synthtool 2020-05-11 21:53:11,378 autosynth > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py -- 2020-05-11 21:53:11,378 autosynth > -> /tmpfs/src/github/synthtool/logs/googleapis/java-billing/126/sponge_log.log 2020-05-11 22:03:18,709 autosynth > Failed executing /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py -- 2020-05-11 22:03:18,711 autosynth > Synthesis failed 2020-05-11 22:03:18,717 autosynth > Running: git reset --hard HEAD 2020-05-11 22:03:18,911 autosynth > Running: git checkout autosynth Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 642, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 489, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 624, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 379, in synthesize_loop synthesize_inner_loop(toolbox, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 390, in synthesize_inner_loop synthesizer, len(toolbox.versions) - 1 File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 274, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 110, in synthesize proc.check_returncode() File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](https://sponge/c5f7338e-3c88-4aa0-9102-e6d02a16d4ef).
1.0
Synthesis failed for java-billing - Hello! Autosynth couldn't regenerate java-billing. :broken_heart: Here's the output from running `synth.py`: ``` 2020-05-11 21:52:44,383 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/java-billing 2020-05-11 21:52:44,900 autosynth > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2020-05-11 21:52:44,905 autosynth > Running: git config user.name yoshi-automation 2020-05-11 21:52:44,910 autosynth > Running: git config user.email yoshi-automation@google.com 2020-05-11 21:52:44,915 autosynth > Running: git config push.default simple 2020-05-11 21:52:44,920 autosynth > Running: git branch -f autosynth 2020-05-11 21:52:44,926 autosynth > Running: git checkout autosynth 2020-05-11 21:52:44,939 autosynth > autosynth flags: 2020-05-11 21:52:44,940 autosynth > AUTOSYNTH_MULTIPLE_COMMITS: true 2020-05-11 21:52:44,940 autosynth > AUTOSYNTH_MULTIPLE_PRS: true 2020-05-11 21:52:44,941 autosynth > Running: git rev-parse --show-toplevel 2020-05-11 21:52:44,946 autosynth > Running: git log -1 --pretty=%H 2020-05-11 21:52:44,951 autosynth > Running: git remote get-url origin 2020-05-11 21:52:44,956 autosynth > Running: git clone --single-branch https://github.com/googleapis/googleapis.git -- /tmpfs/tmp/tmpaaavgl1n/googleapis 2020-05-11 21:52:46,450 autosynth > Running: git log d741cd976975c745d0199987aff0e908b8352992..HEAD --pretty=%H --no-decorate 2020-05-11 21:52:46,457 autosynth > Running: git clone --single-branch https://github.com/googleapis/synthtool.git -- /tmpfs/tmp/tmpaaavgl1n/synthtool 2020-05-11 21:52:47,010 autosynth > Running: git log f8a9933e5e98202b04ef427f28d1d79221190fa4..HEAD --pretty=%H --no-decorate 2020-05-11 21:52:47,017 autosynth > Running: git log -1 --pretty=%at d741cd976975c745d0199987aff0e908b8352992 2020-05-11 21:52:47,022 autosynth > Running: git log -1 --pretty=%at e4f4b23e07315492b533746e6a9255a1e6b3e748 2020-05-11 21:52:47,028 autosynth > Running: git log -1 --pretty=%at ee4ea76504aa60c2bff9b7c11269c155d8c21e0d 2020-05-11 21:52:47,033 autosynth > Running: git log -1 --pretty=%at 6114e2d30c0792522720d1c2cf8f5e111be7545b 2020-05-11 21:52:47,038 autosynth > Running: git log -1 --pretty=%at af71ea52236b05ce73253f6bc935fcf35c866664 2020-05-11 21:52:47,044 autosynth > Running: git log -1 --pretty=%at c451a4457f38e5582e9794a0094c773355a9eaad 2020-05-11 21:52:47,050 autosynth > Running: git log -1 --pretty=%at 9af7212bc2cba8a7184f032e1ab6662ccf604b3a 2020-05-11 21:52:47,055 autosynth > Running: git log -1 --pretty=%at ac1f904979899bc598fb0567cf7a367f08b40b43 2020-05-11 21:52:47,061 autosynth > Running: git log -1 --pretty=%at fe8b17aff5b419ffb418a77228fa049190488233 2020-05-11 21:52:47,067 autosynth > Running: git log -1 --pretty=%at fd35dc6865fa9f47f283d5a6c37c4901162007cf 2020-05-11 21:52:47,072 autosynth > Running: git log -1 --pretty=%at 96c0db42dbcb3bb0116cfa55c054dd139ffb3912 2020-05-11 21:52:47,077 autosynth > Running: git log -1 --pretty=%at b8c7c4e8da56547fc32bee9b563d306fadabdb6f 2020-05-11 21:52:47,082 autosynth > Running: git log -1 --pretty=%at b21bc9b46e3263cd4ebe50f60c61589e3fe797f2 2020-05-11 21:52:47,087 autosynth > Running: git log -1 --pretty=%at 41d5dee23527df3fc9127838a027df393d79ff3c 2020-05-11 21:52:47,092 autosynth > Running: git log -1 --pretty=%at 801c3fb47b1a533989a1fe5af4aedf02af5f2eae 2020-05-11 21:52:47,097 autosynth > Running: git log -1 --pretty=%at 1692c3d51ee7d4366ce81892df1b9ca9846480d3 2020-05-11 21:52:47,102 autosynth > Running: git log -1 --pretty=%at 1df336ab4a989e7b0bc363baa9fc91eb66b0625e 2020-05-11 21:52:47,107 autosynth > Running: git log -1 --pretty=%at 1e92c343790ae63f88817afde60bd2d5bc775f36 2020-05-11 21:52:47,111 autosynth > Running: git log -1 --pretty=%at cf18ab5200e41e9d0cfef88c6c4bcdd394faf1ca 2020-05-11 21:52:47,116 autosynth > Running: git log -1 --pretty=%at ab869e90c6ced0ef30fcc8e8fbe8e527a695d3cb 2020-05-11 21:52:47,121 autosynth > Running: git log -1 --pretty=%at 13736ead90e0fa65103eb03fea775a1c010564af 2020-05-11 21:52:47,126 autosynth > Running: git log -1 --pretty=%at f9d76ac088f5d444b16d5e17fe58a9eea57b7a99 2020-05-11 21:52:47,131 autosynth > Running: git log -1 --pretty=%at 1bd77e8ce6f953ac641af7966d0c52646afc16a8 2020-05-11 21:52:47,136 autosynth > Running: git log -1 --pretty=%at 3824f547aa44df459580615c73cabb30a2a78ee0 2020-05-11 21:52:47,141 autosynth > Running: git log -1 --pretty=%at 4d61e1cb40184a7ad63ef37b1813f6608718674a 2020-05-11 21:52:47,147 autosynth > Running: git log -1 --pretty=%at 9ea892ebe8a7cd66da9bd62eb7086b8dc2ddd559 2020-05-11 21:52:47,152 autosynth > Running: git log -1 --pretty=%at 0feb08065d91d6deebffef3019751f0619a72140 2020-05-11 21:52:47,157 autosynth > Running: git log -1 --pretty=%at 789fd7e0680b04a589447139811d55c5d1f34afe 2020-05-11 21:52:47,163 autosynth > Running: git log -1 --pretty=%at 4efd71279f0781579ced60c374336ce1b8b8cea8 2020-05-11 21:52:47,168 autosynth > Running: git log -1 --pretty=%at a34f4eb96a94bfd5b3e7f8de253db28d0a3e9237 2020-05-11 21:52:47,173 autosynth > Running: git log -1 --pretty=%at f561f57eea013327be27d8f6799b3cc2d18ee804 2020-05-11 21:52:47,177 autosynth > Running: git log -1 --pretty=%at c9942c00e8f35b5b37cd589ef7a0323aeda264e2 2020-05-11 21:52:47,182 autosynth > Running: git log -1 --pretty=%at e2a59ff2f861b1b92b2f15b242780dccddd77ab9 2020-05-11 21:52:47,188 autosynth > Running: git log -1 --pretty=%at df955af5d9508cf9215de6f1fbdad033d5b8fef6 2020-05-11 21:52:47,193 autosynth > Running: git log -1 --pretty=%at dea1f52034d188ca371c3a70c716ca2a6b3f2aa1 2020-05-11 21:52:47,198 autosynth > Running: git log -1 --pretty=%at 35672e7936b923eff048d0ee1529faa163cc81b8 2020-05-11 21:52:47,203 autosynth > Running: git log -1 --pretty=%at 26523a96798ce1a6caa1b3c912119059cfcc98a7 2020-05-11 21:52:47,208 autosynth > Running: git log -1 --pretty=%at 04eff659376f40a7111bf0e8e49fbb38bc7739f7 2020-05-11 21:52:47,213 autosynth > Running: git log -1 --pretty=%at 3028060618e8024af9a32b3ab3456c160091ecb7 2020-05-11 21:52:47,218 autosynth > Running: git log -1 --pretty=%at b5aa7d668d9513cf0d305c8b23262ee0ea9b850c 2020-05-11 21:52:47,223 autosynth > Running: git log -1 --pretty=%at c0a072159eda3c7d5a98201d6500feaf78a18073 2020-05-11 21:52:47,228 autosynth > Running: git log -1 --pretty=%at fea22b1d9f27f86ef355c1d0dba00e0791a08a19 2020-05-11 21:52:47,233 autosynth > Running: git log -1 --pretty=%at 80dac49dd5f407b14f15ac86b4922967b5dd6508 2020-05-11 21:52:47,238 autosynth > Running: git log -1 --pretty=%at b8495354535dc1370377cbb15a3ef7a07ec3a114 2020-05-11 21:52:47,243 autosynth > Running: git log -1 --pretty=%at ab5d07c2883026eec395d73ff59a17f4ad0153db 2020-05-11 21:52:47,249 autosynth > Running: git log -1 --pretty=%at 160e3187dc93265ed635ad68e12c483c8ab4f7be 2020-05-11 21:52:47,254 autosynth > Running: git log -1 --pretty=%at 629e6bea8418ad4a5d081ab9eb48bb327ad368e1 2020-05-11 21:52:47,259 autosynth > Running: git log -1 --pretty=%at c3f3cdeda70f60d808eaaae8fea44d306b6d5cc2 2020-05-11 21:52:47,264 autosynth > Running: git log -1 --pretty=%at cdf13efacdea0649e940452f9c5d320b93735974 2020-05-11 21:52:47,269 autosynth > Running: git log -1 --pretty=%at 101d31acd73076c52d78e18322be01f3debe8cb5 2020-05-11 21:52:47,274 autosynth > Running: git log -1 --pretty=%at cab24c7c9b570e2b483e4f742a7a4a111167238b 2020-05-11 21:52:47,279 autosynth > Running: git log -1 --pretty=%at 6e0ac551e057e20a6c41ab0b21ffa9623e463326 2020-05-11 21:52:47,285 autosynth > Running: git log -1 --pretty=%at 69b4dc56d21394e7c39bc62fe11c76a88aa5511f 2020-05-11 21:52:47,290 autosynth > Running: git log -1 --pretty=%at bcc476396e799806d3355e87246c6becf6250a70 2020-05-11 21:52:47,295 autosynth > Running: git log -1 --pretty=%at b5e9b0b09cc15c263f07da92fa64db31b2e4dd07 2020-05-11 21:52:47,300 autosynth > Running: git log -1 --pretty=%at c09ec5156db66d326cce56ed3a2a8092e02c3abd 2020-05-11 21:52:47,306 autosynth > Running: git log -1 --pretty=%at 42ee97c1b93a0e3759bbba3013da309f670a90ab 2020-05-11 21:52:47,310 autosynth > Running: git log -1 --pretty=%at 45b363d3ab4692115cc5018eb26dcdaa49b320a8 2020-05-11 21:52:47,315 autosynth > Running: git log -1 --pretty=%at 012c553112fcceb7b69822e24335575bf99ec62b 2020-05-11 21:52:47,320 autosynth > Running: git log -1 --pretty=%at 2bf9f22da3f101aeb536e1e88fe036323b8e9188 2020-05-11 21:52:47,326 autosynth > Running: git log -1 --pretty=%at 862cd840dbb0101a0630913217bf8b6d4101c84c 2020-05-11 21:52:47,331 autosynth > Running: git log -1 --pretty=%at b3bcfd059ec684445a38db693f6e36f14a1d2618 2020-05-11 21:52:47,336 autosynth > Running: git log -1 --pretty=%at 1d520eaa7bbd8b40f53942ff03b5496fff887b53 2020-05-11 21:52:47,341 autosynth > Running: git log -1 --pretty=%at 0d7c3565dd942d0cc41b16e895046486ae29e44e 2020-05-11 21:52:47,347 autosynth > Running: git log -1 --pretty=%at d14bf59a446c14ef16e9931ebfc8e63ab549bf07 2020-05-11 21:52:47,352 autosynth > Running: git log -1 --pretty=%at 28e76243c23cc282efbb288cb558c174e3e5e9ee 2020-05-11 21:52:47,357 autosynth > Running: git log -1 --pretty=%at 5dca7f1079a4c127245c0f1dbb8293b434fbd0c8 2020-05-11 21:52:47,363 autosynth > Running: git log -1 --pretty=%at c2bfceedfc9f01356229bd206ad25e250aa9551c 2020-05-11 21:52:47,368 autosynth > Running: git log -1 --pretty=%at 371206424e6db4df2cb3734bb1013dcaac9dbe4b 2020-05-11 21:52:47,373 autosynth > Running: git log -1 --pretty=%at ff17eefa184f6985e5d9a8ee3e725da95a06cd23 2020-05-11 21:52:47,379 autosynth > Running: git log -1 --pretty=%at c284e4f849e4b085a297d336cef8721ab69ba013 2020-05-11 21:52:47,384 autosynth > Running: git log -1 --pretty=%at 756b174de4a122461993c1c583345533d819936d 2020-05-11 21:52:47,390 autosynth > Running: git log -1 --pretty=%at 3f5f8a2258c6a41f9fbf7b80acbca631dda0a952 2020-05-11 21:52:47,395 autosynth > Running: git log -1 --pretty=%at 7e1c7603c8cb2f05ef29842405654ff8a9d3d108 2020-05-11 21:52:47,400 autosynth > Running: git log -1 --pretty=%at f78da0d54c78aa8f66d52a448db1f7ec0e0fd591 2020-05-11 21:52:47,406 autosynth > Running: git log -1 --pretty=%at aaff764c185e18a6c73227357c3df5fa60fec85a 2020-05-11 21:52:47,412 autosynth > Running: git log -1 --pretty=%at ee9e8e4e67f06aba881392a1a2956fbe7a42d216 2020-05-11 21:52:47,418 autosynth > Running: git log -1 --pretty=%at 93b25607689c4bcdb3cb35c657fd3c16249dcd54 2020-05-11 21:52:47,423 autosynth > Running: git log -1 --pretty=%at ab3468cf63abc521be76c9eceb620d61c5ddd76a 2020-05-11 21:52:47,428 autosynth > Running: git log -1 --pretty=%at a737a68e214d1cf927457c996b73e2600b14a949 2020-05-11 21:52:47,433 autosynth > Running: git log -1 --pretty=%at 4ca990cb1a6771f11469b1fdf2fc578271703db0 2020-05-11 21:52:47,439 autosynth > Running: git log -1 --pretty=%at e0f9d9e1f9de890db765be46f45ca8490723e3eb 2020-05-11 21:52:47,445 autosynth > Running: git log -1 --pretty=%at 1b5a8d2bbb69c5a04db26bd546d2888e609c6bab 2020-05-11 21:52:47,450 autosynth > Running: git log -1 --pretty=%at 4bef0001ac7040431ea24b6187424fdec9c08b1b 2020-05-11 21:52:47,456 autosynth > Running: git log -1 --pretty=%at a3a0bf0f6291d69f2ff3df7fcd63d28ee20ac727 2020-05-11 21:52:47,461 autosynth > Running: git log -1 --pretty=%at 6dfd72d028a0d0a43764e060f7b15e004385c3a1 2020-05-11 21:52:47,466 autosynth > Running: git log -1 --pretty=%at 839fae42335ee1bb1e70767f3e6c51738683892b 2020-05-11 21:52:47,471 autosynth > Running: git log -1 --pretty=%at 30cfca094376e4904e32e71c838a81169fd4a2e2 2020-05-11 21:52:47,476 autosynth > Running: git log -1 --pretty=%at c08dcec05ce1c181bcdbce59cabba36e0e541ff6 2020-05-11 21:52:47,481 autosynth > Running: git log -1 --pretty=%at 2fc2caaacb15949c7f80426bfc7dafdd41dbc333 2020-05-11 21:52:47,486 autosynth > Running: git log -1 --pretty=%at 684dfea7decfeca7a7526ea96a8e9256694dd5d8 2020-05-11 21:52:47,491 autosynth > Running: git log -1 --pretty=%at fb8f62b6784f43faf4b64179c57ce4b4931b1a00 2020-05-11 21:52:47,495 autosynth > Running: git log -1 --pretty=%at d3cc7bbf41afdfd6677cc1ab0d53965625cc073b 2020-05-11 21:52:47,500 autosynth > Running: git log -1 --pretty=%at 83816bb3093686a28af2891db5b7506614a820b1 2020-05-11 21:52:47,505 autosynth > Running: git log -1 --pretty=%at 73d4b5d9a791f8b1ee63d439ffe909bb8ffa07f7 2020-05-11 21:52:47,510 autosynth > Running: git log -1 --pretty=%at 482e5206e05e4ba8cd79738f6fb7a521c8c23555 2020-05-11 21:52:47,516 autosynth > Running: git log -1 --pretty=%at 1019b0d9dbd999e1107f90247fcc478678d1105c 2020-05-11 21:52:47,521 autosynth > Running: git log -1 --pretty=%at 4186d3e5424edc088a0f4dbce78f02dd860f3de8 2020-05-11 21:52:47,526 autosynth > Running: git log -1 --pretty=%at aed11c01e52921613b9ee469c2d85f5f33175fb7 2020-05-11 21:52:47,531 autosynth > Running: git log -1 --pretty=%at edd3b80fb770548d6ad780105f1782de6ff73ea0 2020-05-11 21:52:47,536 autosynth > Running: git log -1 --pretty=%at f8a9933e5e98202b04ef427f28d1d79221190fa4 2020-05-11 21:52:47,541 autosynth > Running: git log -1 --pretty=%at 19465d3ec5e5acdb01521d8f3bddd311bcbee28d 2020-05-11 21:52:47,546 autosynth > Running: git log -1 --pretty=%at f5e4c17dc78a966dbf29961dd01f9bbd63e20a04 2020-05-11 21:52:47,551 autosynth > Running: git log -1 --pretty=%at 21ce80f9d4785dc75aa36d26944bddd5a1b7e25b 2020-05-11 21:52:47,556 autosynth > Running: git log -1 --pretty=%at 716f741f2d307b48cbe8a5bc3bc883571212344a 2020-05-11 21:52:47,561 autosynth > Running: git log -1 --pretty=%at 01b6f23d24b27878b48667ce597876d66b59780e 2020-05-11 21:52:47,566 autosynth > Running: git log -1 --pretty=%at 969a2340e74c73227e7c1638ed7650abcac22ee4 2020-05-11 21:52:47,571 autosynth > Running: git log -1 --pretty=%at 720b7bff0da03ed9e67cb2362f2d291241fa59ae 2020-05-11 21:52:47,576 autosynth > Running: git log -1 --pretty=%at 7c1a92d00de654db2fb440243129c61393a694e2 2020-05-11 21:52:47,581 autosynth > Running: git log -1 --pretty=%at 46fb62a3ebda62cc6c1c98e4f1eaf91965eab580 2020-05-11 21:52:47,586 autosynth > Running: git log -1 --pretty=%at 6b685a261479e27fd959dd7b1ff253d6cf595172 2020-05-11 21:52:47,591 autosynth > Running: git log -1 --pretty=%at 958a803ee0c43523fa3e6e0bb016276a69614c8e 2020-05-11 21:52:47,596 autosynth > Running: git log -1 --pretty=%at cdddf139b36000b3a7c65fd2a7781e253262359a 2020-05-11 21:52:47,601 autosynth > Running: git log -1 --pretty=%at 04cb397eb7590ea1e6c10a39fca3d8fe0fb3d256 2020-05-11 21:52:47,605 autosynth > Running: git log -1 --pretty=%at a2c1c4f89a5e220e7b39420ebea33623c7c72804 2020-05-11 21:52:47,610 autosynth > Running: git log -1 --pretty=%at ea3ae392fced25d7d6f5c7ef9c8a1f24768a0936 2020-05-11 21:52:47,614 autosynth > Running: git log -1 --pretty=%at 09c48461232ce929c34386259eb59018ad2d8eef 2020-05-11 21:52:47,619 autosynth > Running: git log -1 --pretty=%at ab883569eb0257bbf16a6d825fd018b3adde3912 2020-05-11 21:52:47,624 autosynth > Running: git log -1 --pretty=%at 756bc4dfc24e8bc4c5dd4116daa41a0440ebf5a0 2020-05-11 21:52:47,629 autosynth > Running: git log -1 --pretty=%at 558bb0d70fa98ea228483b44885704d8941a6a80 2020-05-11 21:52:47,634 autosynth > Running: git log -1 --pretty=%at be74d3e532faa47eb59f1a0eaebde0860d1d8ab4 2020-05-11 21:52:47,638 autosynth > Running: git log -1 --pretty=%at 4674113712c0c7ada19e6a8219d7963ff174b392 2020-05-11 21:52:47,643 autosynth > Running: git log -1 --pretty=%at 5bbfd095faedfe273819d266f21e402192a29041 2020-05-11 21:52:47,648 autosynth > Running: git log -1 --pretty=%at 4fa923bd3dafb91df8613accbe2230299cc5b98e 2020-05-11 21:52:47,652 autosynth > Running: git log -1 --pretty=%at 55cdc844877d97139f25004229842624a6a86a02 2020-05-11 21:52:47,657 autosynth > Running: git log -1 --pretty=%at 98c50772ec23295c64cf0d2ddf199ea52961fd4c 2020-05-11 21:52:47,662 autosynth > Forking toolbox 2020-05-11 21:52:47,662 autosynth > forking: autosynth-self 2020-05-11 21:52:47,663 autosynth > Running: git branch -f autosynth-self 2020-05-11 21:52:47,668 autosynth > forking: autosynth-googleapis 2020-05-11 21:52:47,669 autosynth > Running: git branch -f autosynth-googleapis 2020-05-11 21:52:47,674 autosynth > forking: autosynth-synthtool 2020-05-11 21:52:47,675 autosynth > Running: git branch -f autosynth-synthtool 2020-05-11 21:52:47,679 autosynth > Forked branch: autosynth-self 2020-05-11 21:52:47,997 autosynth > Running: git checkout autosynth-self 2020-05-11 21:52:48,012 autosynth > Building most recent versions 2020-05-11 21:52:48,013 autosynth > Running: git checkout 50afb0e301f97194bf58c78b0133606cf140784f 2020-05-11 21:52:48,020 autosynth > Running: git checkout d741cd976975c745d0199987aff0e908b8352992 2020-05-11 21:52:48,176 autosynth > Running: git checkout f8a9933e5e98202b04ef427f28d1d79221190fa4 2020-05-11 21:52:48,189 autosynth > Running: git branch -f autosynth-self-2 2020-05-11 21:52:48,196 autosynth > Running: git checkout autosynth-self-2 2020-05-11 21:52:48,204 autosynth > Running synthtool 2020-05-11 21:52:48,204 autosynth > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py -- 2020-05-11 21:52:48,204 autosynth > -> /tmpfs/src/github/synthtool/logs/googleapis/java-billing/self/2/sponge_log.log 2020-05-11 21:53:10,684 autosynth > Failed executing /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py -- 2020-05-11 21:53:10,685 autosynth > Synthesis failed 2020-05-11 21:53:10,688 autosynth > Running: git reset --hard HEAD 2020-05-11 21:53:10,817 autosynth > Running: git checkout autosynth-self 2020-05-11 21:53:10,829 autosynth > Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. 2020-05-11 21:53:10,830 autosynth > Building most recent versions 2020-05-11 21:53:10,831 autosynth > Running: git checkout 50afb0e301f97194bf58c78b0133606cf140784f 2020-05-11 21:53:10,839 autosynth > Running: git checkout edd3b80fb770548d6ad780105f1782de6ff73ea0 2020-05-11 21:53:11,242 autosynth > Running: git checkout 98c50772ec23295c64cf0d2ddf199ea52961fd4c 2020-05-11 21:53:11,363 autosynth > Running: git branch -f autosynth-126 2020-05-11 21:53:11,371 autosynth > Running: git checkout autosynth-126 2020-05-11 21:53:11,378 autosynth > Running synthtool 2020-05-11 21:53:11,378 autosynth > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py -- 2020-05-11 21:53:11,378 autosynth > -> /tmpfs/src/github/synthtool/logs/googleapis/java-billing/126/sponge_log.log 2020-05-11 22:03:18,709 autosynth > Failed executing /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py -- 2020-05-11 22:03:18,711 autosynth > Synthesis failed 2020-05-11 22:03:18,717 autosynth > Running: git reset --hard HEAD 2020-05-11 22:03:18,911 autosynth > Running: git checkout autosynth Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 642, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 489, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 624, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 379, in synthesize_loop synthesize_inner_loop(toolbox, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 390, in synthesize_inner_loop synthesizer, len(toolbox.versions) - 1 File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 274, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 110, in synthesize proc.check_returncode() File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](https://sponge/c5f7338e-3c88-4aa0-9102-e6d02a16d4ef).
non_test
synthesis failed for java billing hello autosynth couldn t regenerate java billing broken heart here s the output from running synth py autosynth logs will be written to tmpfs src github synthtool logs googleapis java billing autosynth running git config global core excludesfile home kbuilder autosynth gitignore autosynth running git config user name yoshi automation autosynth running git config user email yoshi automation google com autosynth running git config push default simple autosynth running git branch f autosynth autosynth running git checkout autosynth autosynth autosynth flags autosynth autosynth multiple commits true autosynth autosynth multiple prs true autosynth running git rev parse show toplevel autosynth running git log pretty h autosynth running git remote get url origin autosynth running git clone single branch tmpfs tmp googleapis autosynth running git log head pretty h no decorate autosynth running git clone single branch tmpfs tmp synthtool autosynth running git log head pretty h no decorate autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth forking toolbox autosynth forking autosynth self autosynth running git branch f autosynth self autosynth forking autosynth googleapis autosynth running git branch f autosynth googleapis autosynth forking autosynth synthtool autosynth running git branch f autosynth synthtool autosynth forked branch autosynth self autosynth running git checkout autosynth self autosynth building most recent versions autosynth running git checkout autosynth running git checkout autosynth running git checkout autosynth running git branch f autosynth self autosynth running git checkout autosynth self autosynth running synthtool autosynth running tmpfs src github synthtool env bin m synthtool metadata synth metadata synth py autosynth tmpfs src github synthtool logs googleapis java billing self sponge log log autosynth failed executing tmpfs src github synthtool env bin m synthtool metadata synth metadata synth py autosynth synthesis failed autosynth running git reset hard head autosynth running git checkout autosynth self autosynth command returned non zero exit status autosynth building most recent versions autosynth running git checkout autosynth running git checkout autosynth running git checkout autosynth running git branch f autosynth autosynth running git checkout autosynth autosynth running synthtool autosynth running tmpfs src github synthtool env bin m synthtool metadata synth metadata synth py autosynth tmpfs src github synthtool logs googleapis java billing sponge log log autosynth failed executing tmpfs src github synthtool env bin m synthtool metadata synth metadata synth py autosynth synthesis failed autosynth running git reset hard head autosynth running git checkout autosynth traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop synthesize inner loop toolbox synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize inner loop synthesizer len toolbox versions file tmpfs src github synthtool autosynth synth py line in synthesize version in new branch synthesizer synthesize synth log path self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize proc check returncode file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
339,294
30,389,702,466
IssuesEvent
2023-07-13 05:46:42
dealii/dealii
https://api.github.com/repos/dealii/dealii
opened
Regression tester regressed 79ac9c
High Priority :warning: Regression tester
[start automated regression testsuite report] Dear all, this is the automated regression testsuite reporting a new regression between - new revision: 79ac9cf0deb8c76604c4fd0dee46c185ff26803c [regression testsuite results](https://cdash.dealii.org/index.php?project=deal.II&filtercount=1&showfilters=1&field1=revision&compare1=63&value1=79ac9c) - old revision: 9e4231ab33c5ee1feb527a01c8687506ea1c3049 [regression testsuite results](https://cdash.dealii.org/index.php?project=deal.II&filtercount=1&showfilters=1&field1=revision&compare1=63&value1=9e4231) Summary: - :rage: build errors for configuration [GNU-9.4.0-master-ubuntu-lts-20.04](https://cdash.dealii.org/build/352) on site tester-tng I have identified the following pull requests as possible candidates: - <!-- Use readability-qualified-auto clang-tidy check -->#15705 (authored-by: @masterleinad; merged-by: @bangerth) - <!-- MatrixFree: Memory consumption of mapping data faces by cells in MappingInfo -->#15729 (authored-by: @bergbauer; merged-by: @bangerth) Notes: - The regression reported above is a subset of the following full set of regressions compared to the baseline: - :rage: build errors for configuration [GNU-9.4.0-master-ubuntu-lts-20.04](https://cdash.dealii.org/build/352) on site tester-tng - I will close and unpin this issue automatically once a full run is complete and compares cleanly to the baseline. - If closed I will reopen the issue if the testsuite run identified an additional regression compared to what I have reported so far. [end automated regression testsuite report]
1.0
Regression tester regressed 79ac9c - [start automated regression testsuite report] Dear all, this is the automated regression testsuite reporting a new regression between - new revision: 79ac9cf0deb8c76604c4fd0dee46c185ff26803c [regression testsuite results](https://cdash.dealii.org/index.php?project=deal.II&filtercount=1&showfilters=1&field1=revision&compare1=63&value1=79ac9c) - old revision: 9e4231ab33c5ee1feb527a01c8687506ea1c3049 [regression testsuite results](https://cdash.dealii.org/index.php?project=deal.II&filtercount=1&showfilters=1&field1=revision&compare1=63&value1=9e4231) Summary: - :rage: build errors for configuration [GNU-9.4.0-master-ubuntu-lts-20.04](https://cdash.dealii.org/build/352) on site tester-tng I have identified the following pull requests as possible candidates: - <!-- Use readability-qualified-auto clang-tidy check -->#15705 (authored-by: @masterleinad; merged-by: @bangerth) - <!-- MatrixFree: Memory consumption of mapping data faces by cells in MappingInfo -->#15729 (authored-by: @bergbauer; merged-by: @bangerth) Notes: - The regression reported above is a subset of the following full set of regressions compared to the baseline: - :rage: build errors for configuration [GNU-9.4.0-master-ubuntu-lts-20.04](https://cdash.dealii.org/build/352) on site tester-tng - I will close and unpin this issue automatically once a full run is complete and compares cleanly to the baseline. - If closed I will reopen the issue if the testsuite run identified an additional regression compared to what I have reported so far. [end automated regression testsuite report]
test
regression tester regressed dear all this is the automated regression testsuite reporting a new regression between new revision old revision summary rage build errors for configuration on site tester tng i have identified the following pull requests as possible candidates authored by masterleinad merged by bangerth authored by bergbauer merged by bangerth notes the regression reported above is a subset of the following full set of regressions compared to the baseline rage build errors for configuration on site tester tng i will close and unpin this issue automatically once a full run is complete and compares cleanly to the baseline if closed i will reopen the issue if the testsuite run identified an additional regression compared to what i have reported so far
1
16,832
4,099,731,546
IssuesEvent
2016-06-03 13:49:04
F5Networks/f5-openstack-lbaasv2-driver
https://api.github.com/repos/F5Networks/f5-openstack-lbaasv2-driver
opened
Documentation link in the README is not correct.
critical documentation
#### Agent Version N/A #### Operating System N/A #### OpenStack Release Liberty #### Description The link to the project documentation is not valid. When you click on it you get the RTD maze for page not found. Here is the link that is in the README: http://f5-openstack-lbaasv2.readthedocs.io/en/ #### Deployment N/A
1.0
Documentation link in the README is not correct. - #### Agent Version N/A #### Operating System N/A #### OpenStack Release Liberty #### Description The link to the project documentation is not valid. When you click on it you get the RTD maze for page not found. Here is the link that is in the README: http://f5-openstack-lbaasv2.readthedocs.io/en/ #### Deployment N/A
non_test
documentation link in the readme is not correct agent version n a operating system n a openstack release liberty description the link to the project documentation is not valid when you click on it you get the rtd maze for page not found here is the link that is in the readme deployment n a
0