Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
49,087
| 6,008,707,275
|
IssuesEvent
|
2017-06-06 08:38:02
|
owncloud/client
|
https://api.github.com/repos/owncloud/client
|
closed
|
Client asks for account password after certificate is rejected
|
bug ReadyToTest
|
When a SSL server certificate is presented to the user and rejected the client will not be able to communicate with that server. However, the client will still pop up a dialog asking for the user's password *after* the certificate was rejected. That is pointless and confusing.
Steps:
1. Have a server with a self-signed certificate (or mitmproxy one)
2. Set up an account, log out of it, quit the client.
3. Remove the certificate authorization from the config file. (or restart mitmproxy)
4. Start the client.
5. Reject the server's certificate.
6. Notice that the "Please enter your password" dialog pops up for the account.
@guruz
|
1.0
|
Client asks for account password after certificate is rejected - When a SSL server certificate is presented to the user and rejected the client will not be able to communicate with that server. However, the client will still pop up a dialog asking for the user's password *after* the certificate was rejected. That is pointless and confusing.
Steps:
1. Have a server with a self-signed certificate (or mitmproxy one)
2. Set up an account, log out of it, quit the client.
3. Remove the certificate authorization from the config file. (or restart mitmproxy)
4. Start the client.
5. Reject the server's certificate.
6. Notice that the "Please enter your password" dialog pops up for the account.
@guruz
|
test
|
client asks for account password after certificate is rejected when a ssl server certificate is presented to the user and rejected the client will not be able to communicate with that server however the client will still pop up a dialog asking for the user s password after the certificate was rejected that is pointless and confusing steps have a server with a self signed certificate or mitmproxy one set up an account log out of it quit the client remove the certificate authorization from the config file or restart mitmproxy start the client reject the server s certificate notice that the please enter your password dialog pops up for the account guruz
| 1
|
409,035
| 11,955,756,517
|
IssuesEvent
|
2020-04-04 06:39:38
|
googleapis/google-cloud-dotnet
|
https://api.github.com/repos/googleapis/google-cloud-dotnet
|
closed
|
Synthesis failed for Google.Cloud.PubSub.V1
|
autosynth failure priority: p1 type: bug
|
Hello! Autosynth couldn't regenerate Google.Cloud.PubSub.V1. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1'
Cloning into '/tmpfs/tmp/tmporf5fmop/googleapis'...
Note: checking out '9d4a3ad084027c414a5670ff358a65917c81cdeb'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 9d4a3ad08 chore: set Ruby namespace in proto options
Note: checking out 'd6cb4997910eda04c0c66c0f2fd043eeaa0f660d'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at d6cb4997 chore: enable gapic v2 and proto annotation for documentai API.
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-60'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:48:35,484 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
Building new version of Google.Cloud.PubSub.V1 for compatibility checking
Changes in Google.Cloud.PubSub.V1:
Diff level: Identical
2020-04-03 22:48:50,918 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/Google.Cloud.PubSub.V1/Pubsub.cs
M apis/Google.Cloud.PubSub.V1/synth.metadata
[autosynth-Google.Cloud.PubSub.V1-60 b9f5bffbe] ignored
2 files changed, 4 insertions(+), 3 deletions(-)
HEAD is now at b9f5bffbe ignored
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
Previous HEAD position was d6cb4997 chore: enable gapic v2 and proto annotation for documentai API.
HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes
Note: checking out '98024617efce32982bd763ad14f00c9bc0819bea'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 98024617e Touch all the synth.metadata files again
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-0'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:48:51,462 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking
2020-04-03 22:48:57,493 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/synth.metadata
HEAD is now at 98024617e Touch all the synth.metadata files again
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
Note: checking out '9d4a3ad084027c414a5670ff358a65917c81cdeb'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 9d4a3ad08 chore: set Ruby namespace in proto options
Previous HEAD position was 7be2811a fix: Update gapic-generator version to pickup discogapic fixes
HEAD is now at 17cfae00 Add a new AuthorizationType for Data Source Definition.
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-30'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:48:58,081 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking
2020-04-03 22:49:04,186 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/synth.metadata
HEAD is now at 9d4a3ad08 chore: set Ruby namespace in proto options
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
Previous HEAD position was 17cfae00 Add a new AuthorizationType for Data Source Definition.
HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes
Note: checking out '5a41fb5f1c7f5329cc981b77cae1f4762d705002'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 5a41fb5f1 Dialogflow weekly v2 library update
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-15'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:49:04,714 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking
2020-04-03 22:49:10,840 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/synth.metadata
HEAD is now at 5a41fb5f1 Dialogflow weekly v2 library update
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes
Note: checking out 'edebc2b7a22574b76412a9c1cf6832719b9c85e8'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at edebc2b7a fix Dataproc: add missing `REQUIRED` annotation.
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-7'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:49:11,371 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking
2020-04-03 22:49:17,464 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/synth.metadata
HEAD is now at edebc2b7a fix Dataproc: add missing `REQUIRED` annotation.
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes
Note: checking out '38fb38aeb841fb08e12ff366c2159fb6669b45d8'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 38fb38aeb Touch synth.metadata to get SecretManager APIs building
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-3'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:49:18,004 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking
2020-04-03 22:49:24,391 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/synth.metadata
HEAD is now at 38fb38aeb Touch synth.metadata to get SecretManager APIs building
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes
Note: checking out '0c88ce0da0e66910735643832b237c9f873b05e6'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 0c88ce0da Add a new AuthorizationType for Data Source Definition.
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-1'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:49:24,939 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking
2020-04-03 22:49:30,979 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/synth.metadata
HEAD is now at 0c88ce0da Add a new AuthorizationType for Data Source Definition.
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
On branch autosynth-Google.Cloud.PubSub.V1
nothing to commit, working tree clean
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 484, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 370, in main
return _inner_main(temp_dir)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 474, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 284, in synthesize_loop
synthesize_range(toolbox, synthesizer)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 306, in synthesize_range
toolbox.patch_merge_version(young)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 146, in patch_merge_version
comment or self.versions[index].version.get_comment()
File "/tmpfs/src/git/autosynth/autosynth/git.py", line 95, in commit_all_changes
subprocess.check_call(["git", "commit", "-m", message])
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['git', 'commit', '-m', 'Add a new AuthorizationType for Data Source Definition.\n\nhttps://github.com/googleapis/google-cloud-dotnet/commit/0c88ce0da0e66910735643832b237c9f873b05e6\ncommit 0c88ce0da0e66910735643832b237c9f873b05e6\nAuthor: yoshi-automation <yoshi-automation@google.com>\nDate: Wed Apr 1 02:27:58 2020 -0700\n\n Add a new AuthorizationType for Data Source Definition.\n \n https://github.com/googleapis/googleapis/commit/17cfae00f2bb51cb1683f017da7e295a1b0f01a8\n commit 17cfae00f2bb51cb1683f017da7e295a1b0f01a8\n Author: Google APIs <noreply@google.com>\n Date: Tue Mar 31 10:21:11 2020 -0700\n \n Add a new AuthorizationType for Data Source Definition.\n \n PiperOrigin-RevId: 303992863']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/89ee1c7b-40f0-4f7a-a7cd-f68bdd6f31c1).
|
1.0
|
Synthesis failed for Google.Cloud.PubSub.V1 - Hello! Autosynth couldn't regenerate Google.Cloud.PubSub.V1. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1'
Cloning into '/tmpfs/tmp/tmporf5fmop/googleapis'...
Note: checking out '9d4a3ad084027c414a5670ff358a65917c81cdeb'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 9d4a3ad08 chore: set Ruby namespace in proto options
Note: checking out 'd6cb4997910eda04c0c66c0f2fd043eeaa0f660d'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at d6cb4997 chore: enable gapic v2 and proto annotation for documentai API.
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-60'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:48:35,484 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
Building new version of Google.Cloud.PubSub.V1 for compatibility checking
Changes in Google.Cloud.PubSub.V1:
Diff level: Identical
2020-04-03 22:48:50,918 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/Google.Cloud.PubSub.V1/Pubsub.cs
M apis/Google.Cloud.PubSub.V1/synth.metadata
[autosynth-Google.Cloud.PubSub.V1-60 b9f5bffbe] ignored
2 files changed, 4 insertions(+), 3 deletions(-)
HEAD is now at b9f5bffbe ignored
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
Previous HEAD position was d6cb4997 chore: enable gapic v2 and proto annotation for documentai API.
HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes
Note: checking out '98024617efce32982bd763ad14f00c9bc0819bea'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 98024617e Touch all the synth.metadata files again
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-0'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:48:51,462 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking
2020-04-03 22:48:57,493 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/synth.metadata
HEAD is now at 98024617e Touch all the synth.metadata files again
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
Note: checking out '9d4a3ad084027c414a5670ff358a65917c81cdeb'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 9d4a3ad08 chore: set Ruby namespace in proto options
Previous HEAD position was 7be2811a fix: Update gapic-generator version to pickup discogapic fixes
HEAD is now at 17cfae00 Add a new AuthorizationType for Data Source Definition.
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-30'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:48:58,081 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking
2020-04-03 22:49:04,186 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/synth.metadata
HEAD is now at 9d4a3ad08 chore: set Ruby namespace in proto options
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
Previous HEAD position was 17cfae00 Add a new AuthorizationType for Data Source Definition.
HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes
Note: checking out '5a41fb5f1c7f5329cc981b77cae1f4762d705002'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 5a41fb5f1 Dialogflow weekly v2 library update
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-15'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:49:04,714 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking
2020-04-03 22:49:10,840 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/synth.metadata
HEAD is now at 5a41fb5f1 Dialogflow weekly v2 library update
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes
Note: checking out 'edebc2b7a22574b76412a9c1cf6832719b9c85e8'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at edebc2b7a fix Dataproc: add missing `REQUIRED` annotation.
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-7'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:49:11,371 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking
2020-04-03 22:49:17,464 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/synth.metadata
HEAD is now at edebc2b7a fix Dataproc: add missing `REQUIRED` annotation.
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes
Note: checking out '38fb38aeb841fb08e12ff366c2159fb6669b45d8'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 38fb38aeb Touch synth.metadata to get SecretManager APIs building
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-3'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:49:18,004 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking
2020-04-03 22:49:24,391 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/synth.metadata
HEAD is now at 38fb38aeb Touch synth.metadata to get SecretManager APIs building
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
HEAD is now at 7be2811a fix: Update gapic-generator version to pickup discogapic fixes
Note: checking out '0c88ce0da0e66910735643832b237c9f873b05e6'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 0c88ce0da Add a new AuthorizationType for Data Source Definition.
Switched to a new branch 'autosynth-Google.Cloud.PubSub.V1-1'
Running synthtool
['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-04-03 22:49:24,939 synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.Cloud.PubSub.V1/synth.py.
Skipping microgenerator fetch/build: already built, and running on Kokoro
Building existing version of Google.Cloud.PubSub.V1 for compatibility checking
Generating Google.Cloud.PubSub.V1
Running mid-micro-generation script for Google.Cloud.PubSub.V1
Running post-generation script for Google.Cloud.PubSub.V1
Updated 1 path from the index
Updated 1 path from the index
git detects no change in Google.Cloud.PubSub.V1; skipping compatibility checking
2020-04-03 22:49:30,979 synthtool > Wrote metadata to synth.metadata.
Changed files:
M apis/Google.Cloud.PubSub.V1/synth.metadata
HEAD is now at 0c88ce0da Add a new AuthorizationType for Data Source Definition.
Switched to branch 'autosynth-Google.Cloud.PubSub.V1'
On branch autosynth-Google.Cloud.PubSub.V1
nothing to commit, working tree clean
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 484, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 370, in main
return _inner_main(temp_dir)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 474, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 284, in synthesize_loop
synthesize_range(toolbox, synthesizer)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 306, in synthesize_range
toolbox.patch_merge_version(young)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 146, in patch_merge_version
comment or self.versions[index].version.get_comment()
File "/tmpfs/src/git/autosynth/autosynth/git.py", line 95, in commit_all_changes
subprocess.check_call(["git", "commit", "-m", message])
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['git', 'commit', '-m', 'Add a new AuthorizationType for Data Source Definition.\n\nhttps://github.com/googleapis/google-cloud-dotnet/commit/0c88ce0da0e66910735643832b237c9f873b05e6\ncommit 0c88ce0da0e66910735643832b237c9f873b05e6\nAuthor: yoshi-automation <yoshi-automation@google.com>\nDate: Wed Apr 1 02:27:58 2020 -0700\n\n Add a new AuthorizationType for Data Source Definition.\n \n https://github.com/googleapis/googleapis/commit/17cfae00f2bb51cb1683f017da7e295a1b0f01a8\n commit 17cfae00f2bb51cb1683f017da7e295a1b0f01a8\n Author: Google APIs <noreply@google.com>\n Date: Tue Mar 31 10:21:11 2020 -0700\n \n Add a new AuthorizationType for Data Source Definition.\n \n PiperOrigin-RevId: 303992863']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/89ee1c7b-40f0-4f7a-a7cd-f68bdd6f31c1).
|
non_test
|
synthesis failed for google cloud pubsub hello autosynth couldn t regenerate google cloud pubsub broken heart here s the output from running synth py cloning into working repo switched to a new branch autosynth google cloud pubsub cloning into tmpfs tmp googleapis note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at chore set ruby namespace in proto options note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at chore enable gapic and proto annotation for documentai api switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index building new version of google cloud pubsub for compatibility checking changes in google cloud pubsub diff level identical synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub google cloud pubsub pubsub cs m apis google cloud pubsub synth metadata ignored files changed insertions deletions head is now at ignored switched to branch autosynth google cloud pubsub previous head position was chore enable gapic and proto annotation for documentai api head is now at fix update gapic generator version to pickup discogapic fixes note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at touch all the synth metadata files again switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index git detects no change in google cloud pubsub skipping compatibility checking synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub synth metadata head is now at touch all the synth metadata files again switched to branch autosynth google cloud pubsub note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at chore set ruby namespace in proto options previous head position was fix update gapic generator version to pickup discogapic fixes head is now at add a new authorizationtype for data source definition switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index git detects no change in google cloud pubsub skipping compatibility checking synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub synth metadata head is now at chore set ruby namespace in proto options switched to branch autosynth google cloud pubsub previous head position was add a new authorizationtype for data source definition head is now at fix update gapic generator version to pickup discogapic fixes note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at dialogflow weekly library update switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index git detects no change in google cloud pubsub skipping compatibility checking synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub synth metadata head is now at dialogflow weekly library update switched to branch autosynth google cloud pubsub head is now at fix update gapic generator version to pickup discogapic fixes note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at fix dataproc add missing required annotation switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index git detects no change in google cloud pubsub skipping compatibility checking synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub synth metadata head is now at fix dataproc add missing required annotation switched to branch autosynth google cloud pubsub head is now at fix update gapic generator version to pickup discogapic fixes note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at touch synth metadata to get secretmanager apis building switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index git detects no change in google cloud pubsub skipping compatibility checking synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub synth metadata head is now at touch synth metadata to get secretmanager apis building switched to branch autosynth google cloud pubsub head is now at fix update gapic generator version to pickup discogapic fixes note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at add a new authorizationtype for data source definition switched to a new branch autosynth google cloud pubsub running synthtool synthtool executing tmpfs src git autosynth working repo apis google cloud pubsub synth py skipping microgenerator fetch build already built and running on kokoro building existing version of google cloud pubsub for compatibility checking generating google cloud pubsub running mid micro generation script for google cloud pubsub running post generation script for google cloud pubsub updated path from the index updated path from the index git detects no change in google cloud pubsub skipping compatibility checking synthtool wrote metadata to synth metadata changed files m apis google cloud pubsub synth metadata head is now at add a new authorizationtype for data source definition switched to branch autosynth google cloud pubsub on branch autosynth google cloud pubsub nothing to commit working tree clean traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main return inner main temp dir file tmpfs src git autosynth autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src git autosynth autosynth synth py line in synthesize loop synthesize range toolbox synthesizer file tmpfs src git autosynth autosynth synth py line in synthesize range toolbox patch merge version young file tmpfs src git autosynth autosynth synth py line in patch merge version comment or self versions version get comment file tmpfs src git autosynth autosynth git py line in commit all changes subprocess check call file home kbuilder pyenv versions lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
| 0
|
477,059
| 13,754,824,790
|
IssuesEvent
|
2020-10-06 17:30:46
|
red-hat-storage/ocs-ci
|
https://api.github.com/repos/red-hat-storage/ocs-ci
|
closed
|
Ensure ocscilib module adheres to log level set by user
|
High Priority team/ecosystem
|
Currently the logger used in [ocscilib](https://github.com/red-hat-storage/ocs-ci/blob/master/ocs_ci/framework/pytest_customization/ocscilib.py) is only logging `WARNING` and greater log levels. This should adhere to the log level set by the user (default is INFO). I suspect this is due to how pytest loads and calls this module / these functions, but we should still find a way to ensure we can log properly here as this code is critical.
|
1.0
|
Ensure ocscilib module adheres to log level set by user - Currently the logger used in [ocscilib](https://github.com/red-hat-storage/ocs-ci/blob/master/ocs_ci/framework/pytest_customization/ocscilib.py) is only logging `WARNING` and greater log levels. This should adhere to the log level set by the user (default is INFO). I suspect this is due to how pytest loads and calls this module / these functions, but we should still find a way to ensure we can log properly here as this code is critical.
|
non_test
|
ensure ocscilib module adheres to log level set by user currently the logger used in is only logging warning and greater log levels this should adhere to the log level set by the user default is info i suspect this is due to how pytest loads and calls this module these functions but we should still find a way to ensure we can log properly here as this code is critical
| 0
|
344,710
| 10,348,813,595
|
IssuesEvent
|
2019-09-04 20:43:32
|
mozilla/addons-code-manager
|
https://api.github.com/repos/mozilla/addons-code-manager
|
opened
|
Add API for creating / updating a comment
|
priority: p3 qa: not needed
|
The thunk for managing a comment (https://github.com/mozilla/addons-code-manager/issues/988) needs some methods to work with the [API](https://addons-server.readthedocs.io/en/latest/topics/api/reviewers.html#post--api-v4-reviewers-addon-(int-addon_id)-versions-(int-version_id)-draft_comments-).
|
1.0
|
Add API for creating / updating a comment - The thunk for managing a comment (https://github.com/mozilla/addons-code-manager/issues/988) needs some methods to work with the [API](https://addons-server.readthedocs.io/en/latest/topics/api/reviewers.html#post--api-v4-reviewers-addon-(int-addon_id)-versions-(int-version_id)-draft_comments-).
|
non_test
|
add api for creating updating a comment the thunk for managing a comment needs some methods to work with the
| 0
|
280,143
| 8,678,339,895
|
IssuesEvent
|
2018-11-30 19:36:11
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
closed
|
Synthesis failed for redis
|
autosynth failure priority: p1 type: bug
|
Hello! Autosynth couldn't regenerate redis. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to branch 'autosynth-redis'
[35msynthtool > [31m[43mYou are running the synthesis script directly, this will be disabled in a future release of Synthtool. Please use python3 -m synthtool instead.[0m
[35msynthtool > [36mEnsuring dependencies.[0m
[35msynthtool > [36mPulling artman image.[0m
latest: Pulling from googleapis/artman
Digest: sha256:2f6b261ee7fe1aedf238991c93a20b3820de37a343d0cacf3e3e9555c2aaf2ea
Status: Image is up to date for googleapis/artman:latest
[35msynthtool > [36mCloning googleapis.[0m
[35msynthtool > [36mRunning generator for google/cloud/redis/artman_redis_v1beta1.yaml.[0m
[35msynthtool > [32mGenerated code into /home/kbuilder/.cache/synthtool/googleapis/artman-genfiles/python/redis-v1beta1.[0m
[35msynthtool > [36mRunning generator for google/cloud/redis/artman_redis_v1.yaml.[0m
[35msynthtool > [32mGenerated code into /home/kbuilder/.cache/synthtool/googleapis/artman-genfiles/python/redis-v1.[0m
[35msynthtool > [33mNo replacements made in google/cloud/**/cloud_redis_client.py for pattern resources of the form:\n ``, maybe replacement is not longer needed?[0m
[35msynthtool > [33mNo replacements made in google/cloud/**/cloud_redis_client.py for pattern
parent \(str\): Required. The resource name of the instance location using the form:
::
`projects/{project_id}/locations/{location_id}`
where ``location_id`` refers to a GCP region, maybe replacement is not longer needed?[0m
[35msynthtool > [33mNo replacements made in google/cloud/**/cloud_redis_client.py for pattern
with the following restrictions:
\* Must contain only lowercase letters, numbers, and hyphens\., maybe replacement is not longer needed?[0m
[35msynthtool > [33mNo replacements made in google/cloud/**/cloud_redis_client.py for pattern
name \(str\): Required. Redis instance resource name using the form:
::
`projects/{project_id}/locations/{location_id}/instances/{instance_id}`
where ``location_id`` refers to a GCP region, maybe replacement is not longer needed?[0m
[35msynthtool > [33mNo replacements made in google/cloud/**/cloud_redis_client.py for pattern
fields from ``Instance``:
\* ``displayName``
\* ``labels``
\* ``memorySizeGb``
\* ``redisConfig``, maybe replacement is not longer needed?[0m
.coveragerc
.flake8
MANIFEST.in
noxfile.py.j2
setup.cfg
Traceback (most recent call last):
File "synth.py", line 101, in <module>
s.shell.run(["nox", "-s", "blacken"], hide_output=False)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run
with Popen(*popenargs, **kwargs) as process:
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 707, in __init__
restore_signals, start_new_session)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 1326, in _execute_child
raise child_exception_type(errno_num, err_msg)
FileNotFoundError: [Errno 2] No such file or directory: 'nox'
[35msynthtool > [36mCleaned up 2 temporary directories.[0m
Synthesis failed
```
Google internal developers can see the full log [here](https://sponge/3abb58ef-0652-4854-846e-e1d9be78e73e).
|
1.0
|
Synthesis failed for redis - Hello! Autosynth couldn't regenerate redis. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to branch 'autosynth-redis'
[35msynthtool > [31m[43mYou are running the synthesis script directly, this will be disabled in a future release of Synthtool. Please use python3 -m synthtool instead.[0m
[35msynthtool > [36mEnsuring dependencies.[0m
[35msynthtool > [36mPulling artman image.[0m
latest: Pulling from googleapis/artman
Digest: sha256:2f6b261ee7fe1aedf238991c93a20b3820de37a343d0cacf3e3e9555c2aaf2ea
Status: Image is up to date for googleapis/artman:latest
[35msynthtool > [36mCloning googleapis.[0m
[35msynthtool > [36mRunning generator for google/cloud/redis/artman_redis_v1beta1.yaml.[0m
[35msynthtool > [32mGenerated code into /home/kbuilder/.cache/synthtool/googleapis/artman-genfiles/python/redis-v1beta1.[0m
[35msynthtool > [36mRunning generator for google/cloud/redis/artman_redis_v1.yaml.[0m
[35msynthtool > [32mGenerated code into /home/kbuilder/.cache/synthtool/googleapis/artman-genfiles/python/redis-v1.[0m
[35msynthtool > [33mNo replacements made in google/cloud/**/cloud_redis_client.py for pattern resources of the form:\n ``, maybe replacement is not longer needed?[0m
[35msynthtool > [33mNo replacements made in google/cloud/**/cloud_redis_client.py for pattern
parent \(str\): Required. The resource name of the instance location using the form:
::
`projects/{project_id}/locations/{location_id}`
where ``location_id`` refers to a GCP region, maybe replacement is not longer needed?[0m
[35msynthtool > [33mNo replacements made in google/cloud/**/cloud_redis_client.py for pattern
with the following restrictions:
\* Must contain only lowercase letters, numbers, and hyphens\., maybe replacement is not longer needed?[0m
[35msynthtool > [33mNo replacements made in google/cloud/**/cloud_redis_client.py for pattern
name \(str\): Required. Redis instance resource name using the form:
::
`projects/{project_id}/locations/{location_id}/instances/{instance_id}`
where ``location_id`` refers to a GCP region, maybe replacement is not longer needed?[0m
[35msynthtool > [33mNo replacements made in google/cloud/**/cloud_redis_client.py for pattern
fields from ``Instance``:
\* ``displayName``
\* ``labels``
\* ``memorySizeGb``
\* ``redisConfig``, maybe replacement is not longer needed?[0m
.coveragerc
.flake8
MANIFEST.in
noxfile.py.j2
setup.cfg
Traceback (most recent call last):
File "synth.py", line 101, in <module>
s.shell.run(["nox", "-s", "blacken"], hide_output=False)
File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run
with Popen(*popenargs, **kwargs) as process:
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 707, in __init__
restore_signals, start_new_session)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 1326, in _execute_child
raise child_exception_type(errno_num, err_msg)
FileNotFoundError: [Errno 2] No such file or directory: 'nox'
[35msynthtool > [36mCleaned up 2 temporary directories.[0m
Synthesis failed
```
Google internal developers can see the full log [here](https://sponge/3abb58ef-0652-4854-846e-e1d9be78e73e).
|
non_test
|
synthesis failed for redis hello autosynth couldn t regenerate redis broken heart here s the output from running synth py cloning into working repo switched to branch autosynth redis are running the synthesis script directly this will be disabled in a future release of synthtool please use m synthtool instead dependencies artman image latest pulling from googleapis artman digest status image is up to date for googleapis artman latest googleapis generator for google cloud redis artman redis yaml code into home kbuilder cache synthtool googleapis artman genfiles python redis generator for google cloud redis artman redis yaml code into home kbuilder cache synthtool googleapis artman genfiles python redis replacements made in google cloud cloud redis client py for pattern resources of the form n maybe replacement is not longer needed replacements made in google cloud cloud redis client py for pattern parent str required the resource name of the instance location using the form projects project id locations location id where location id refers to a gcp region maybe replacement is not longer needed replacements made in google cloud cloud redis client py for pattern with the following restrictions must contain only lowercase letters numbers and hyphens maybe replacement is not longer needed replacements made in google cloud cloud redis client py for pattern name str required redis instance resource name using the form projects project id locations location id instances instance id where location id refers to a gcp region maybe replacement is not longer needed replacements made in google cloud cloud redis client py for pattern fields from instance displayname labels memorysizegb redisconfig maybe replacement is not longer needed coveragerc manifest in noxfile py setup cfg traceback most recent call last file synth py line in s shell run hide output false file tmpfs src git autosynth env lib site packages synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run with popen popenargs kwargs as process file home kbuilder pyenv versions lib subprocess py line in init restore signals start new session file home kbuilder pyenv versions lib subprocess py line in execute child raise child exception type errno num err msg filenotfounderror no such file or directory nox up temporary directories synthesis failed google internal developers can see the full log
| 0
|
342,744
| 30,636,743,374
|
IssuesEvent
|
2023-07-24 18:22:58
|
Loweredgames/Skyblock-Classic-Edition
|
https://api.github.com/repos/Loweredgames/Skyblock-Classic-Edition
|
closed
|
Aggiustata e migliorata la funzione Legacy per la 5.0.0
|
bug confirmed low test fix performance
|
- [x] Testare
---------------------------------------------
sara l'ultima volta che verra aggiustata.
vedere: #202
|
1.0
|
Aggiustata e migliorata la funzione Legacy per la 5.0.0 - - [x] Testare
---------------------------------------------
sara l'ultima volta che verra aggiustata.
vedere: #202
|
test
|
aggiustata e migliorata la funzione legacy per la testare sara l ultima volta che verra aggiustata vedere
| 1
|
313,456
| 23,475,716,681
|
IssuesEvent
|
2022-08-17 05:40:58
|
ToolJet/ToolJet
|
https://api.github.com/repos/ToolJet/ToolJet
|
opened
|
[docs]: White labelling - EE feature
|
documentation
|
### Summary
Add documentation for White labelling - the feature added to EE in which customer can configure`CUSTOM_LOGO_URL` - url of custom logo.
best viewed :
maximum width 130px
height: 26px
### Issue Type
Documentation feature/update
### The entire URL of the documentation with the issue
_No response_
### Steps to reproduce the issue
_No response_
### Additional Information
_No response_
### If the issue is confirmed, would you be willing to submit a pull request?
No
### Code of Conduct
- [X] I agree to follow the ToolJet Code of Conduct
|
1.0
|
[docs]: White labelling - EE feature - ### Summary
Add documentation for White labelling - the feature added to EE in which customer can configure`CUSTOM_LOGO_URL` - url of custom logo.
best viewed :
maximum width 130px
height: 26px
### Issue Type
Documentation feature/update
### The entire URL of the documentation with the issue
_No response_
### Steps to reproduce the issue
_No response_
### Additional Information
_No response_
### If the issue is confirmed, would you be willing to submit a pull request?
No
### Code of Conduct
- [X] I agree to follow the ToolJet Code of Conduct
|
non_test
|
white labelling ee feature summary add documentation for white labelling the feature added to ee in which customer can configure custom logo url url of custom logo best viewed maximum width height issue type documentation feature update the entire url of the documentation with the issue no response steps to reproduce the issue no response additional information no response if the issue is confirmed would you be willing to submit a pull request no code of conduct i agree to follow the tooljet code of conduct
| 0
|
289,403
| 24,987,099,255
|
IssuesEvent
|
2022-11-02 15:48:55
|
vegaprotocol/vega
|
https://api.github.com/repos/vegaprotocol/vega
|
opened
|
Update feature tests for 0005-COLL-001 for additional events
|
feature tests
|
As part of the spec reviews it was mentioned that for the AC 0005-COLL-001 we would add additional tests for other events as per:
- https://github.com/vegaprotocol/specs/issues/1321
|
1.0
|
Update feature tests for 0005-COLL-001 for additional events - As part of the spec reviews it was mentioned that for the AC 0005-COLL-001 we would add additional tests for other events as per:
- https://github.com/vegaprotocol/specs/issues/1321
|
test
|
update feature tests for coll for additional events as part of the spec reviews it was mentioned that for the ac coll we would add additional tests for other events as per
| 1
|
160,516
| 12,515,966,032
|
IssuesEvent
|
2020-06-03 08:36:48
|
dasch-swiss/knora-app
|
https://api.github.com/repos/dasch-swiss/knora-app
|
closed
|
Removing a search criteria e.g. property
|
bug user-testing
|
*Advanced Search, not logged in
I created a long search with several selection properties.
There seems to be a limit in criterias (see also image)
Now I decide, that the first of the properties is not so important and I wish to remove it from my list (in this example im picture the property "Color - exists).
There seems to be no possibility to remove that property individually but I have to remove all others afterwards in the list as well.

**Desktop (please complete the following information):**
* OS: Windows 10 Pro N, 1903 - 18362.418
* Browser Google Chrome, 78.0.3904.87
* Version Knora App, 1.6.2
|
1.0
|
Removing a search criteria e.g. property - *Advanced Search, not logged in
I created a long search with several selection properties.
There seems to be a limit in criterias (see also image)
Now I decide, that the first of the properties is not so important and I wish to remove it from my list (in this example im picture the property "Color - exists).
There seems to be no possibility to remove that property individually but I have to remove all others afterwards in the list as well.

**Desktop (please complete the following information):**
* OS: Windows 10 Pro N, 1903 - 18362.418
* Browser Google Chrome, 78.0.3904.87
* Version Knora App, 1.6.2
|
test
|
removing a search criteria e g property advanced search not logged in i created a long search with several selection properties there seems to be a limit in criterias see also image now i decide that the first of the properties is not so important and i wish to remove it from my list in this example im picture the property color exists there seems to be no possibility to remove that property individually but i have to remove all others afterwards in the list as well desktop please complete the following information os windows pro n browser google chrome version knora app
| 1
|
288,135
| 31,857,033,785
|
IssuesEvent
|
2023-09-15 08:14:04
|
nidhi7598/linux-4.19.72_CVE-2022-3564
|
https://api.github.com/repos/nidhi7598/linux-4.19.72_CVE-2022-3564
|
closed
|
CVE-2020-25704 (Medium) detected in linuxlinux-4.19.294 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2020-25704 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72_CVE-2022-3564/commit/454c7dacf6fa9a6de86d4067f5a08f25cffa519b">454c7dacf6fa9a6de86d4067f5a08f25cffa519b</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw memory leak in the Linux kernel performance monitoring subsystem was found in the way if using PERF_EVENT_IOC_SET_FILTER. A local user could use this flaw to starve the resources causing denial of service.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25704>CVE-2020-25704</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25704">https://www.linuxkernelcves.com/cves/CVE-2020-25704</a></p>
<p>Release Date: 2020-12-02</p>
<p>Fix Resolution: v4.9.244, v4.14.207, v4.19.156, v5.4.76, v5.9.7, v5.10-rc3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-25704 (Medium) detected in linuxlinux-4.19.294 - autoclosed - ## CVE-2020-25704 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72_CVE-2022-3564/commit/454c7dacf6fa9a6de86d4067f5a08f25cffa519b">454c7dacf6fa9a6de86d4067f5a08f25cffa519b</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw memory leak in the Linux kernel performance monitoring subsystem was found in the way if using PERF_EVENT_IOC_SET_FILTER. A local user could use this flaw to starve the resources causing denial of service.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25704>CVE-2020-25704</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25704">https://www.linuxkernelcves.com/cves/CVE-2020-25704</a></p>
<p>Release Date: 2020-12-02</p>
<p>Fix Resolution: v4.9.244, v4.14.207, v4.19.156, v5.4.76, v5.9.7, v5.10-rc3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files vulnerability details a flaw memory leak in the linux kernel performance monitoring subsystem was found in the way if using perf event ioc set filter a local user could use this flaw to starve the resources causing denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
777,562
| 27,285,220,220
|
IssuesEvent
|
2023-02-23 13:03:00
|
wso2/api-manager
|
https://api.github.com/repos/wso2/api-manager
|
closed
|
No Server ALPNProcessors! error when running APIEndpointCertificateTestCase
|
Type/Bug Priority/Normal Component/APIM Affected/4.2.0-Beta
|
### Description
- When running the `APIEndpointCertificateTestCase` in the integration test module, there is an error occuring in the form of `No Server ALPNProcessors!`.
- When debugged this was occuring due to a startup failure in the `Wire Mock Server` library.
### Steps to Reproduce
1. Run the `APIEndpointCertificateTestCase` by enabling only it in the testng.xml
### Affected Component
APIM
### Version
4.2.0
### Environment Details (with versions)
MacOS Ventura with M1 chip
### Relevant Log Output
```shell
INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - Jacoco class file paths analyzed
INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - -- repository/components/plugins
INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - -- repository/deployment/server/webapps
INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - -- lib/runtimes/cxf3
INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - Jacoco coverage HTML report path : /Users/dinithherath/Desktop/Work/Personal/product-apim/modules/integration/tests-integration/tests-backend/target/jacoco/coverage/index.html
[ERROR] Tests run: 53, Failures: 4, Errors: 0, Skipped: 20, Time elapsed: 1,087.093 s <<< FAILURE! - in TestSuite
[ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize Time elapsed: 349.106 s <<< FAILURE!
java.lang.IllegalStateException: No Server ALPNProcessors!
at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.startSecureEndpoint(APIEndpointCertificateTestCase.java:317)
at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize(APIEndpointCertificateTestCase.java:110)
[ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy Time elapsed: 349.107 s <<< FAILURE!
org.wso2.am.integration.clients.publisher.api.ApiException: Missing the required parameter 'apiId' when calling deleteAPI(Async)
at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy(APIEndpointCertificateTestCase.java:327)
[ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize Time elapsed: 351.543 s <<< FAILURE!
java.lang.IllegalStateException: No Server ALPNProcessors!
at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.startSecureEndpoint(APIEndpointCertificateTestCase.java:317)
at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize(APIEndpointCertificateTestCase.java:110)
[ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy Time elapsed: 351.543 s <<< FAILURE!
org.wso2.am.integration.clients.publisher.api.ApiException: Missing the required parameter 'apiId' when calling deleteAPI(Async)
at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy(APIEndpointCertificateTestCase.java:327)
```
### Related Issues
_No response_
### Suggested Labels
_No response_
|
1.0
|
No Server ALPNProcessors! error when running APIEndpointCertificateTestCase - ### Description
- When running the `APIEndpointCertificateTestCase` in the integration test module, there is an error occuring in the form of `No Server ALPNProcessors!`.
- When debugged this was occuring due to a startup failure in the `Wire Mock Server` library.
### Steps to Reproduce
1. Run the `APIEndpointCertificateTestCase` by enabling only it in the testng.xml
### Affected Component
APIM
### Version
4.2.0
### Environment Details (with versions)
MacOS Ventura with M1 chip
### Relevant Log Output
```shell
INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - Jacoco class file paths analyzed
INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - -- repository/components/plugins
INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - -- repository/deployment/server/webapps
INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - -- lib/runtimes/cxf3
INFO [org.wso2.carbon.automation.extensions.servers.carbonserver.CarbonServerManager] - Jacoco coverage HTML report path : /Users/dinithherath/Desktop/Work/Personal/product-apim/modules/integration/tests-integration/tests-backend/target/jacoco/coverage/index.html
[ERROR] Tests run: 53, Failures: 4, Errors: 0, Skipped: 20, Time elapsed: 1,087.093 s <<< FAILURE! - in TestSuite
[ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize Time elapsed: 349.106 s <<< FAILURE!
java.lang.IllegalStateException: No Server ALPNProcessors!
at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.startSecureEndpoint(APIEndpointCertificateTestCase.java:317)
at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize(APIEndpointCertificateTestCase.java:110)
[ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy Time elapsed: 349.107 s <<< FAILURE!
org.wso2.am.integration.clients.publisher.api.ApiException: Missing the required parameter 'apiId' when calling deleteAPI(Async)
at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy(APIEndpointCertificateTestCase.java:327)
[ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize Time elapsed: 351.543 s <<< FAILURE!
java.lang.IllegalStateException: No Server ALPNProcessors!
at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.startSecureEndpoint(APIEndpointCertificateTestCase.java:317)
at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.initialize(APIEndpointCertificateTestCase.java:110)
[ERROR] org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy Time elapsed: 351.543 s <<< FAILURE!
org.wso2.am.integration.clients.publisher.api.ApiException: Missing the required parameter 'apiId' when calling deleteAPI(Async)
at org.wso2.am.integration.tests.api.lifecycle.APIEndpointCertificateTestCase.destroy(APIEndpointCertificateTestCase.java:327)
```
### Related Issues
_No response_
### Suggested Labels
_No response_
|
non_test
|
no server alpnprocessors error when running apiendpointcertificatetestcase description when running the apiendpointcertificatetestcase in the integration test module there is an error occuring in the form of no server alpnprocessors when debugged this was occuring due to a startup failure in the wire mock server library steps to reproduce run the apiendpointcertificatetestcase by enabling only it in the testng xml affected component apim version environment details with versions macos ventura with chip relevant log output shell info jacoco class file paths analyzed info repository components plugins info repository deployment server webapps info lib runtimes info jacoco coverage html report path users dinithherath desktop work personal product apim modules integration tests integration tests backend target jacoco coverage index html tests run failures errors skipped time elapsed s failure in testsuite org am integration tests api lifecycle apiendpointcertificatetestcase initialize time elapsed s failure java lang illegalstateexception no server alpnprocessors at org am integration tests api lifecycle apiendpointcertificatetestcase startsecureendpoint apiendpointcertificatetestcase java at org am integration tests api lifecycle apiendpointcertificatetestcase initialize apiendpointcertificatetestcase java org am integration tests api lifecycle apiendpointcertificatetestcase destroy time elapsed s failure org am integration clients publisher api apiexception missing the required parameter apiid when calling deleteapi async at org am integration tests api lifecycle apiendpointcertificatetestcase destroy apiendpointcertificatetestcase java org am integration tests api lifecycle apiendpointcertificatetestcase initialize time elapsed s failure java lang illegalstateexception no server alpnprocessors at org am integration tests api lifecycle apiendpointcertificatetestcase startsecureendpoint apiendpointcertificatetestcase java at org am integration tests api lifecycle apiendpointcertificatetestcase initialize apiendpointcertificatetestcase java org am integration tests api lifecycle apiendpointcertificatetestcase destroy time elapsed s failure org am integration clients publisher api apiexception missing the required parameter apiid when calling deleteapi async at org am integration tests api lifecycle apiendpointcertificatetestcase destroy apiendpointcertificatetestcase java related issues no response suggested labels no response
| 0
|
84,289
| 7,909,990,545
|
IssuesEvent
|
2018-07-03 03:08:42
|
elastic/beats
|
https://api.github.com/repos/elastic/beats
|
opened
|
Test Filebeat modules from packages
|
:Testing Filebeat
|
We don't have any tests in place to detect if the Filebeat modules work from the packages we produce. I noticed this because they weren't included in the packages and none of our tests failed. I suggest adding some tests in these areas:
- [ ] Check that packages have `modules` and `modules.d` at a minimum. Currently we check file permissions and ownership, but we don't assert they exist. I suggest adding a flag like `--modules` or `--dir` that causes the check to test that the directories exist in all packages.
- [ ] Run some tests on the Filebeat modules from beats-tester. Maybe execute the enable commands.
|
1.0
|
Test Filebeat modules from packages - We don't have any tests in place to detect if the Filebeat modules work from the packages we produce. I noticed this because they weren't included in the packages and none of our tests failed. I suggest adding some tests in these areas:
- [ ] Check that packages have `modules` and `modules.d` at a minimum. Currently we check file permissions and ownership, but we don't assert they exist. I suggest adding a flag like `--modules` or `--dir` that causes the check to test that the directories exist in all packages.
- [ ] Run some tests on the Filebeat modules from beats-tester. Maybe execute the enable commands.
|
test
|
test filebeat modules from packages we don t have any tests in place to detect if the filebeat modules work from the packages we produce i noticed this because they weren t included in the packages and none of our tests failed i suggest adding some tests in these areas check that packages have modules and modules d at a minimum currently we check file permissions and ownership but we don t assert they exist i suggest adding a flag like modules or dir that causes the check to test that the directories exist in all packages run some tests on the filebeat modules from beats tester maybe execute the enable commands
| 1
|
27,269
| 12,535,511,669
|
IssuesEvent
|
2020-06-04 21:31:24
|
MicrosoftDocs/powerbi-docs
|
https://api.github.com/repos/MicrosoftDocs/powerbi-docs
|
closed
|
PostgreSQL requires gateway
|
assigned-to-author doc-bug powerbi-service/subsvc powerbi/svc pri1
|
The table is incorrect; PostgreSQL requires the data gateway to connect and refresh from the Power BI Service.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 3c34c249-033f-4156-0abc-8d16cbe355a0
* Version Independent ID: 20dc29c2-9cb7-2322-e80b-3f801bc34cfe
* Content: [Power BI data sources - Power BI](https://docs.microsoft.com/en-us/power-bi/power-bi-data-sources#feedback)
* Content Source: [powerbi-docs/power-bi-data-sources.md](https://github.com/MicrosoftDocs/powerbi-docs/blob/live/powerbi-docs/power-bi-data-sources.md)
* Service: **powerbi**
* Sub-service: **powerbi-service**
* GitHub Login: @kfollis
* Microsoft Alias: **kfollis**
|
1.0
|
PostgreSQL requires gateway - The table is incorrect; PostgreSQL requires the data gateway to connect and refresh from the Power BI Service.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 3c34c249-033f-4156-0abc-8d16cbe355a0
* Version Independent ID: 20dc29c2-9cb7-2322-e80b-3f801bc34cfe
* Content: [Power BI data sources - Power BI](https://docs.microsoft.com/en-us/power-bi/power-bi-data-sources#feedback)
* Content Source: [powerbi-docs/power-bi-data-sources.md](https://github.com/MicrosoftDocs/powerbi-docs/blob/live/powerbi-docs/power-bi-data-sources.md)
* Service: **powerbi**
* Sub-service: **powerbi-service**
* GitHub Login: @kfollis
* Microsoft Alias: **kfollis**
|
non_test
|
postgresql requires gateway the table is incorrect postgresql requires the data gateway to connect and refresh from the power bi service document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service powerbi sub service powerbi service github login kfollis microsoft alias kfollis
| 0
|
50,622
| 13,541,511,051
|
IssuesEvent
|
2020-09-16 15:58:43
|
oracle/helidon
|
https://api.github.com/repos/oracle/helidon
|
closed
|
Fix javadoc of EvictableCache
|
2.x P3 security
|
EvictableCache now implies some implementation details in its javadoc, such as backing it by a `ConcurrentHashMap` - this should not be done, as anybody can implement the interface.
Fix javadoc of the class and its methods + add details to `factory` methods and builder that explain how the default implementation works.
Examples of wrong docs:
`Generic cache with eviction and max size` - the interface itself does not imply max size
javadoc of `size` method.
|
True
|
Fix javadoc of EvictableCache - EvictableCache now implies some implementation details in its javadoc, such as backing it by a `ConcurrentHashMap` - this should not be done, as anybody can implement the interface.
Fix javadoc of the class and its methods + add details to `factory` methods and builder that explain how the default implementation works.
Examples of wrong docs:
`Generic cache with eviction and max size` - the interface itself does not imply max size
javadoc of `size` method.
|
non_test
|
fix javadoc of evictablecache evictablecache now implies some implementation details in its javadoc such as backing it by a concurrenthashmap this should not be done as anybody can implement the interface fix javadoc of the class and its methods add details to factory methods and builder that explain how the default implementation works examples of wrong docs generic cache with eviction and max size the interface itself does not imply max size javadoc of size method
| 0
|
204,164
| 7,084,874,615
|
IssuesEvent
|
2018-01-11 08:59:40
|
wulkano/kap
|
https://api.github.com/repos/wulkano/kap
|
opened
|
Blacklist redundant apps from the window selector.
|
Priority: Medium
|
**Platform:** MacOS
**Kap Version:** 2.0.0-beta.4 (2.0.0-beta.4.556)
### Steps to reproduce
Currently, a vast amount of redundant apps show up in the window selector that are currently not recordable.
### Workaround
Suggestion: Create a blacklist of known redundant apps that shouldn't show up in the menu.
|
1.0
|
Blacklist redundant apps from the window selector. - **Platform:** MacOS
**Kap Version:** 2.0.0-beta.4 (2.0.0-beta.4.556)
### Steps to reproduce
Currently, a vast amount of redundant apps show up in the window selector that are currently not recordable.
### Workaround
Suggestion: Create a blacklist of known redundant apps that shouldn't show up in the menu.
|
non_test
|
blacklist redundant apps from the window selector platform macos kap version beta beta steps to reproduce currently a vast amount of redundant apps show up in the window selector that are currently not recordable workaround suggestion create a blacklist of known redundant apps that shouldn t show up in the menu
| 0
|
61,155
| 25,383,311,305
|
IssuesEvent
|
2022-11-21 19:29:41
|
BCDevOps/developer-experience
|
https://api.github.com/repos/BCDevOps/developer-experience
|
closed
|
SSO realm merge planning - Rocketchat
|
ops and shared services
|
**Describe the issue**
we will be merging the realm that RC uses to another realm in Gold. This ticket is to find out what's the impact and if the merge is doable for this service.
**What is the plan? How will this get completed?**
Discussion with Service Lead, testing, planning
**Definition of done**
- [x] identify service's SSO usage
- [x] discuss if realm merge is doable
- [x] come up with testing plan
|
1.0
|
SSO realm merge planning - Rocketchat - **Describe the issue**
we will be merging the realm that RC uses to another realm in Gold. This ticket is to find out what's the impact and if the merge is doable for this service.
**What is the plan? How will this get completed?**
Discussion with Service Lead, testing, planning
**Definition of done**
- [x] identify service's SSO usage
- [x] discuss if realm merge is doable
- [x] come up with testing plan
|
non_test
|
sso realm merge planning rocketchat describe the issue we will be merging the realm that rc uses to another realm in gold this ticket is to find out what s the impact and if the merge is doable for this service what is the plan how will this get completed discussion with service lead testing planning definition of done identify service s sso usage discuss if realm merge is doable come up with testing plan
| 0
|
145,980
| 11,716,637,697
|
IssuesEvent
|
2020-03-09 15:56:57
|
waterloop/teamhub
|
https://api.github.com/repos/waterloop/teamhub
|
closed
|
Add config.json for tests running on circle-ci
|
backend testing
|
Need to add mongodb+srv url for when tests run on circle-ci.
|
1.0
|
Add config.json for tests running on circle-ci - Need to add mongodb+srv url for when tests run on circle-ci.
|
test
|
add config json for tests running on circle ci need to add mongodb srv url for when tests run on circle ci
| 1
|
292,120
| 25,202,059,142
|
IssuesEvent
|
2022-11-13 08:18:50
|
david3214/Team_2_CS_2450-601
|
https://api.github.com/repos/david3214/Team_2_CS_2450-601
|
closed
|
Functional/Non-Functional Testing
|
testing
|
According to the textbook:
**Functional test**—Deals with features the application provides. These are generally listed in the requirements.
**Nonfunctional test**—Studies application characteristics that aren’t related to specific functions the users will perform. For example, these tests might check performance under a heavy user load, with limited memory, or with missing network connections. These often identify minimal requirements.
|
1.0
|
Functional/Non-Functional Testing - According to the textbook:
**Functional test**—Deals with features the application provides. These are generally listed in the requirements.
**Nonfunctional test**—Studies application characteristics that aren’t related to specific functions the users will perform. For example, these tests might check performance under a heavy user load, with limited memory, or with missing network connections. These often identify minimal requirements.
|
test
|
functional non functional testing according to the textbook functional test —deals with features the application provides these are generally listed in the requirements nonfunctional test —studies application characteristics that aren’t related to specific functions the users will perform for example these tests might check performance under a heavy user load with limited memory or with missing network connections these often identify minimal requirements
| 1
|
256,971
| 22,139,086,764
|
IssuesEvent
|
2022-06-03 03:55:01
|
cmwall0605/cse112-sp22-group8
|
https://api.github.com/repos/cmwall0605/cse112-sp22-group8
|
opened
|
[Testing]: Add Overall Testing Suite for Cypress
|
difficult issue Priority: High testing task-11-testing
|
### Describe the feature to implement
Create a comprehensive suite of testing for the user
### Breakdown of what needs to be completed
- [ ] Create a test where a user enters the timer page, starts, and does not create a task. Ensure that the create task dialog pops up, and ensure that the timer runs for 30 minutes (25 work + 5 break). Ensure that after the break, that there is no task-item created in the task page. Ensure that the stats timer page has the correct info
- [ ] Create a test where a user enters the timer page, starts, and does create a task. Ensure that the create task dialog pops up, and ensure that the timer runs for 30 minutes (25 work + 5 break). Ensure that after the break, that there is a task-item created in the task page with the correct title and the correct ratio. Ensure that the stats timer page has the correct info.
- [ ] Create a test where a user enters the timer page, goes to the stats page, creates a task, and then goes to the timer page through the header (not by playing the task). Ensure that it goes through the same process as the first process.
- [ ] Create a test where a user enters the timer page, goes to the stats page, creates a task, and then goes to the timer page through the task item (not through the header). Click the start and ensure that the create task does not show up. After the session, go to the task page and ensure that the task chosen updates its ratio and that the stats page is updated.
- [ ] Create a test where the user edits the work, short and long break and then starts a test. Make sure that the work, short, and long sessions are the correct amount of time.
- [ ] Enable auto timer and run through 4 work sessions, ensuring that that the system runs through them automatically and that no interaction needs to happen.
### Additional Context if Applicable
|
2.0
|
[Testing]: Add Overall Testing Suite for Cypress - ### Describe the feature to implement
Create a comprehensive suite of testing for the user
### Breakdown of what needs to be completed
- [ ] Create a test where a user enters the timer page, starts, and does not create a task. Ensure that the create task dialog pops up, and ensure that the timer runs for 30 minutes (25 work + 5 break). Ensure that after the break, that there is no task-item created in the task page. Ensure that the stats timer page has the correct info
- [ ] Create a test where a user enters the timer page, starts, and does create a task. Ensure that the create task dialog pops up, and ensure that the timer runs for 30 minutes (25 work + 5 break). Ensure that after the break, that there is a task-item created in the task page with the correct title and the correct ratio. Ensure that the stats timer page has the correct info.
- [ ] Create a test where a user enters the timer page, goes to the stats page, creates a task, and then goes to the timer page through the header (not by playing the task). Ensure that it goes through the same process as the first process.
- [ ] Create a test where a user enters the timer page, goes to the stats page, creates a task, and then goes to the timer page through the task item (not through the header). Click the start and ensure that the create task does not show up. After the session, go to the task page and ensure that the task chosen updates its ratio and that the stats page is updated.
- [ ] Create a test where the user edits the work, short and long break and then starts a test. Make sure that the work, short, and long sessions are the correct amount of time.
- [ ] Enable auto timer and run through 4 work sessions, ensuring that that the system runs through them automatically and that no interaction needs to happen.
### Additional Context if Applicable
|
test
|
add overall testing suite for cypress describe the feature to implement create a comprehensive suite of testing for the user breakdown of what needs to be completed create a test where a user enters the timer page starts and does not create a task ensure that the create task dialog pops up and ensure that the timer runs for minutes work break ensure that after the break that there is no task item created in the task page ensure that the stats timer page has the correct info create a test where a user enters the timer page starts and does create a task ensure that the create task dialog pops up and ensure that the timer runs for minutes work break ensure that after the break that there is a task item created in the task page with the correct title and the correct ratio ensure that the stats timer page has the correct info create a test where a user enters the timer page goes to the stats page creates a task and then goes to the timer page through the header not by playing the task ensure that it goes through the same process as the first process create a test where a user enters the timer page goes to the stats page creates a task and then goes to the timer page through the task item not through the header click the start and ensure that the create task does not show up after the session go to the task page and ensure that the task chosen updates its ratio and that the stats page is updated create a test where the user edits the work short and long break and then starts a test make sure that the work short and long sessions are the correct amount of time enable auto timer and run through work sessions ensuring that that the system runs through them automatically and that no interaction needs to happen additional context if applicable
| 1
|
280,164
| 24,281,791,791
|
IssuesEvent
|
2022-09-28 18:05:35
|
Ascentroid/Ascentroid
|
https://api.github.com/repos/Ascentroid/Ascentroid
|
closed
|
Aftershock splash damage not killing sometimes, when it should
|
bug fixed needs testing
|
(In Shock Henge) Aftershock splash damage not killing sometimes, when it should
|
1.0
|
Aftershock splash damage not killing sometimes, when it should - (In Shock Henge) Aftershock splash damage not killing sometimes, when it should
|
test
|
aftershock splash damage not killing sometimes when it should in shock henge aftershock splash damage not killing sometimes when it should
| 1
|
165,785
| 12,879,872,091
|
IssuesEvent
|
2020-07-12 01:27:28
|
osquery/osquery
|
https://api.github.com/repos/osquery/osquery
|
closed
|
Create tests for table `shell_history`
|
FreeBSD Linux good-first-issue macOS test
|
## Create tests for table `shell_history`
- Create header file for the table implementation, if one is not exists.
- In test, query the table and check if retrieved columns (name and types) match the columns from table spec.
- If there is any guarantee to number of rows (e.g. only 1 record in every query result, more than 3 records or something else) check it.
- Test the implementation details of the table, if it possible.
Table spec: `specs/posix/shell_history.table`
Source files:
- `osquery/tables/system/posix/shell_history.h`
Table generating function: `genShellHistory()`
|
1.0
|
Create tests for table `shell_history` - ## Create tests for table `shell_history`
- Create header file for the table implementation, if one is not exists.
- In test, query the table and check if retrieved columns (name and types) match the columns from table spec.
- If there is any guarantee to number of rows (e.g. only 1 record in every query result, more than 3 records or something else) check it.
- Test the implementation details of the table, if it possible.
Table spec: `specs/posix/shell_history.table`
Source files:
- `osquery/tables/system/posix/shell_history.h`
Table generating function: `genShellHistory()`
|
test
|
create tests for table shell history create tests for table shell history create header file for the table implementation if one is not exists in test query the table and check if retrieved columns name and types match the columns from table spec if there is any guarantee to number of rows e g only record in every query result more than records or something else check it test the implementation details of the table if it possible table spec specs posix shell history table source files osquery tables system posix shell history h table generating function genshellhistory
| 1
|
315,940
| 27,120,548,247
|
IssuesEvent
|
2023-02-15 22:26:26
|
filecoin-project/ref-fvm
|
https://api.github.com/repos/filecoin-project/ref-fvm
|
opened
|
Eth API: Write full test case for EthGetStorageAt
|
Topic: Testing
|
Specifically, we need to test:
1. Looking up storage for a non-existent actor (returns 32 zero bytes).
2. Looking up storage for a non-evm actor (returns 32 zero bytes).
3. Looking up an unset slot (returns 32 zero bytes).
4. Looking up a set slot (returns the value).
5. Looking up a set slot _after_ deleting an actor (selfdestruct).
We already have some test coverage here, but it's not thorough.
|
1.0
|
Eth API: Write full test case for EthGetStorageAt - Specifically, we need to test:
1. Looking up storage for a non-existent actor (returns 32 zero bytes).
2. Looking up storage for a non-evm actor (returns 32 zero bytes).
3. Looking up an unset slot (returns 32 zero bytes).
4. Looking up a set slot (returns the value).
5. Looking up a set slot _after_ deleting an actor (selfdestruct).
We already have some test coverage here, but it's not thorough.
|
test
|
eth api write full test case for ethgetstorageat specifically we need to test looking up storage for a non existent actor returns zero bytes looking up storage for a non evm actor returns zero bytes looking up an unset slot returns zero bytes looking up a set slot returns the value looking up a set slot after deleting an actor selfdestruct we already have some test coverage here but it s not thorough
| 1
|
551,886
| 16,190,678,355
|
IssuesEvent
|
2021-05-04 08:02:10
|
epiphany-platform/epiphany
|
https://api.github.com/repos/epiphany-platform/epiphany
|
closed
|
[BUG] [Ubuntu] elasticsearch-curator in version 5.8.3 is not available from APT repo
|
area/logs priority/critical type/bug
|
**Describe the bug**
Epicli installation fails while downloading requirements as the `elasticsearch-curator` package in version 5.8.3 is no longer available in the Curator APT repository.
**How to reproduce**
Steps to reproduce the behavior:
1. execute `epicli apply` (repository vm is enough)
**Expected behavior**
Repository has been set up properly.
**Config files**
**Environment**
- OS: [Ubuntu]
**epicli version**: [`epicli --version`]
All epicli versions affected
**Additional context**
```
root@ip-10-1-11-41:~# apt-cache policy elasticsearch-curator
elasticsearch-curator:
Installed: (none)
Candidate: 5.8.4
Version table:
5.8.4 500
500 https://packages.elastic.co/curator/5/debian stable/main amd64 Packages
5.2.0-1 500
500 http://eu-west-1.ec2.archive.ubuntu.com/ubuntu bionic/universe amd64 Packages
```
---
**DoD checklist**
* [x] Changelog updated (if affected version was released)
* [x] COMPONENTS.md updated / doesn't need to be updated
* [x] Automated tests passed (QA pipelines)
* [x] apply
* [ ] upgrade
* [x] Case covered by automated test (if possible)
* [ ] Idempotency tested
* [x] Documentation updated / doesn't need to be updated
* [x] All conversations in PR resolved
* [x] Backport tasks created / doesn't need to be backported
|
1.0
|
[BUG] [Ubuntu] elasticsearch-curator in version 5.8.3 is not available from APT repo - **Describe the bug**
Epicli installation fails while downloading requirements as the `elasticsearch-curator` package in version 5.8.3 is no longer available in the Curator APT repository.
**How to reproduce**
Steps to reproduce the behavior:
1. execute `epicli apply` (repository vm is enough)
**Expected behavior**
Repository has been set up properly.
**Config files**
**Environment**
- OS: [Ubuntu]
**epicli version**: [`epicli --version`]
All epicli versions affected
**Additional context**
```
root@ip-10-1-11-41:~# apt-cache policy elasticsearch-curator
elasticsearch-curator:
Installed: (none)
Candidate: 5.8.4
Version table:
5.8.4 500
500 https://packages.elastic.co/curator/5/debian stable/main amd64 Packages
5.2.0-1 500
500 http://eu-west-1.ec2.archive.ubuntu.com/ubuntu bionic/universe amd64 Packages
```
---
**DoD checklist**
* [x] Changelog updated (if affected version was released)
* [x] COMPONENTS.md updated / doesn't need to be updated
* [x] Automated tests passed (QA pipelines)
* [x] apply
* [ ] upgrade
* [x] Case covered by automated test (if possible)
* [ ] Idempotency tested
* [x] Documentation updated / doesn't need to be updated
* [x] All conversations in PR resolved
* [x] Backport tasks created / doesn't need to be backported
|
non_test
|
elasticsearch curator in version is not available from apt repo describe the bug epicli installation fails while downloading requirements as the elasticsearch curator package in version is no longer available in the curator apt repository how to reproduce steps to reproduce the behavior execute epicli apply repository vm is enough expected behavior repository has been set up properly config files environment os epicli version all epicli versions affected additional context root ip apt cache policy elasticsearch curator elasticsearch curator installed none candidate version table stable main packages bionic universe packages dod checklist changelog updated if affected version was released components md updated doesn t need to be updated automated tests passed qa pipelines apply upgrade case covered by automated test if possible idempotency tested documentation updated doesn t need to be updated all conversations in pr resolved backport tasks created doesn t need to be backported
| 0
|
273,927
| 23,795,534,578
|
IssuesEvent
|
2022-09-02 19:11:01
|
pypa/pip
|
https://api.github.com/repos/pypa/pip
|
closed
|
Test Suite: Tests are not using the latest setuptools
|
type: enhancement C: tests
|
I'm struggling with running the tests while implementing #5407 (https://travis-ci.org/pfmoore/pip/jobs/420017719). The failure "error: 'egg_base' must be a directory name" is a bug in setuptools, that is fixed in the latest version, 40.2.0. But after a lot of debugging, I've found that the build environment used by `test_pip_wheel_with_pep518_build_reqs_no_isolation` has setuptools 39.1.0 installed. I've no idea why - there's no mention of that version in the pip sources, and as far as I can see that version isn't installed on my PC.
So I'm somewhat baffled as to where the test suite is getting that version from. If anyone has any suggestions, I'd really appreciate it. Ultimately, this is a bug in the test suite as it doesn't seem to be using the latest version of setuptools, but in the short term my main concern is that I can't progress on the PEP 517 implementation until I get this sorted.
@pypa/pip-committers any suggestions? Or anyone else?
|
1.0
|
Test Suite: Tests are not using the latest setuptools - I'm struggling with running the tests while implementing #5407 (https://travis-ci.org/pfmoore/pip/jobs/420017719). The failure "error: 'egg_base' must be a directory name" is a bug in setuptools, that is fixed in the latest version, 40.2.0. But after a lot of debugging, I've found that the build environment used by `test_pip_wheel_with_pep518_build_reqs_no_isolation` has setuptools 39.1.0 installed. I've no idea why - there's no mention of that version in the pip sources, and as far as I can see that version isn't installed on my PC.
So I'm somewhat baffled as to where the test suite is getting that version from. If anyone has any suggestions, I'd really appreciate it. Ultimately, this is a bug in the test suite as it doesn't seem to be using the latest version of setuptools, but in the short term my main concern is that I can't progress on the PEP 517 implementation until I get this sorted.
@pypa/pip-committers any suggestions? Or anyone else?
|
test
|
test suite tests are not using the latest setuptools i m struggling with running the tests while implementing the failure error egg base must be a directory name is a bug in setuptools that is fixed in the latest version but after a lot of debugging i ve found that the build environment used by test pip wheel with build reqs no isolation has setuptools installed i ve no idea why there s no mention of that version in the pip sources and as far as i can see that version isn t installed on my pc so i m somewhat baffled as to where the test suite is getting that version from if anyone has any suggestions i d really appreciate it ultimately this is a bug in the test suite as it doesn t seem to be using the latest version of setuptools but in the short term my main concern is that i can t progress on the pep implementation until i get this sorted pypa pip committers any suggestions or anyone else
| 1
|
308,040
| 26,575,727,506
|
IssuesEvent
|
2023-01-21 19:50:04
|
TheRenegadeCoder/sample-programs
|
https://api.github.com/repos/TheRenegadeCoder/sample-programs
|
closed
|
Add Dijkstra in Mathematica
|
enhancement needs tests dijkstra
|
To request a new code snippet, please fill out the following:
Project name: Dijkstra
Project link: https://sampleprograms.io/projects/dijkstra
Language: Mathematica
If this code snippet will also be the first in a new language, please fill out the following:
Official Language Style Guide:
Official Language Website:
Official Language Docker Image: https://hub.docker.com/r/repository/image
> The project link must be a link to a project on the [Sample Programs Website project list][sample-programs-project-list].
> If you would like to add a new project to the Sample Programs Website,
> please first make a pull request to the [Sample Programs Website][sample-programs-website].
[sample-programs-website]: https://github.com/TheRenegadeCoder/sample-programs-website
[sample-programs-project-list]: https://sampleprograms.io/projects/
|
1.0
|
Add Dijkstra in Mathematica - To request a new code snippet, please fill out the following:
Project name: Dijkstra
Project link: https://sampleprograms.io/projects/dijkstra
Language: Mathematica
If this code snippet will also be the first in a new language, please fill out the following:
Official Language Style Guide:
Official Language Website:
Official Language Docker Image: https://hub.docker.com/r/repository/image
> The project link must be a link to a project on the [Sample Programs Website project list][sample-programs-project-list].
> If you would like to add a new project to the Sample Programs Website,
> please first make a pull request to the [Sample Programs Website][sample-programs-website].
[sample-programs-website]: https://github.com/TheRenegadeCoder/sample-programs-website
[sample-programs-project-list]: https://sampleprograms.io/projects/
|
test
|
add dijkstra in mathematica to request a new code snippet please fill out the following project name dijkstra project link language mathematica if this code snippet will also be the first in a new language please fill out the following official language style guide official language website official language docker image the project link must be a link to a project on the if you would like to add a new project to the sample programs website please first make a pull request to the
| 1
|
309,491
| 23,297,202,429
|
IssuesEvent
|
2022-08-06 19:18:54
|
r-hyperspec/hyperSpec
|
https://api.github.com/repos/r-hyperspec/hyperSpec
|
opened
|
Review the examples with faux_cell if they make sense
|
Topic: documentation :blue_book:
|
After changing from `chondro` into `faux_cell` in examples and vignettes, the examples should be carefully reviewed to make sure that the results make sense.
https://github.com/cbeleites/hyperSpec/issues/236
|
1.0
|
Review the examples with faux_cell if they make sense - After changing from `chondro` into `faux_cell` in examples and vignettes, the examples should be carefully reviewed to make sure that the results make sense.
https://github.com/cbeleites/hyperSpec/issues/236
|
non_test
|
review the examples with faux cell if they make sense after changing from chondro into faux cell in examples and vignettes the examples should be carefully reviewed to make sure that the results make sense
| 0
|
100,794
| 8,755,136,686
|
IssuesEvent
|
2018-12-14 14:00:18
|
LiskHQ/lisk-hub
|
https://api.github.com/repos/LiskHQ/lisk-hub
|
closed
|
Fix unit tests for send transaction with HW wallet - Followup #1317
|
type: test
|
<!--- Provide a descriptive summary in the Title above -->
### Description of the problem
<!-- A description of the problem you want to solve, including why you think this is a problem -->
Create tests covering files below so test coverage is passing
```'src/components/passphraseSteps/index.js',
'src/components/resultBox/resultBox.js',
'src/components/send/steps/confirm/confirm.js',
'src/actions/peers.js', ,```
|
1.0
|
Fix unit tests for send transaction with HW wallet - Followup #1317 - <!--- Provide a descriptive summary in the Title above -->
### Description of the problem
<!-- A description of the problem you want to solve, including why you think this is a problem -->
Create tests covering files below so test coverage is passing
```'src/components/passphraseSteps/index.js',
'src/components/resultBox/resultBox.js',
'src/components/send/steps/confirm/confirm.js',
'src/actions/peers.js', ,```
|
test
|
fix unit tests for send transaction with hw wallet followup description of the problem create tests covering files below so test coverage is passing src components passphrasesteps index js src components resultbox resultbox js src components send steps confirm confirm js src actions peers js
| 1
|
114,266
| 14,544,025,415
|
IssuesEvent
|
2020-12-15 17:37:05
|
cloudfour/cloudfour.com-patterns
|
https://api.github.com/repos/cloudfour/cloudfour.com-patterns
|
closed
|
Grid/layout patterns
|
size:10 ⁉️ question ✋ blocked 🎨 design
|
Create a grid system that can be used to layout content.
<img width="636" alt="grids" src="https://user-images.githubusercontent.com/42841342/74988623-1d2f0300-53f3-11ea-8f7d-1f1fd05db3ca.png">
|
1.0
|
Grid/layout patterns - Create a grid system that can be used to layout content.
<img width="636" alt="grids" src="https://user-images.githubusercontent.com/42841342/74988623-1d2f0300-53f3-11ea-8f7d-1f1fd05db3ca.png">
|
non_test
|
grid layout patterns create a grid system that can be used to layout content img width alt grids src
| 0
|
222,518
| 7,433,136,136
|
IssuesEvent
|
2018-03-26 06:03:32
|
telerik/kendo-ui-core
|
https://api.github.com/repos/telerik/kendo-ui-core
|
opened
|
The popup edit form is not closed on the first update click if there was a invalid value before entering a valid one(SASS themes).
|
Bug C: Grid Kendo1 Priority 1 SEV: Low
|
### Bug report
The popup edit form is not closed on the first update click if there was an invalid value before entering a valid one(SASS themes).
### Reproduction of the problem
The issue can be observed in our demo:
https://demos.telerik.com/kendo-ui/grid/editing-popup
I made a video showcasing the issue: https://www.screencast.com/t/FiR6zWhPp
### Environment
* **Kendo UI version:** 2018.2.221
* **Browser:** [all ]
|
1.0
|
The popup edit form is not closed on the first update click if there was a invalid value before entering a valid one(SASS themes). - ### Bug report
The popup edit form is not closed on the first update click if there was an invalid value before entering a valid one(SASS themes).
### Reproduction of the problem
The issue can be observed in our demo:
https://demos.telerik.com/kendo-ui/grid/editing-popup
I made a video showcasing the issue: https://www.screencast.com/t/FiR6zWhPp
### Environment
* **Kendo UI version:** 2018.2.221
* **Browser:** [all ]
|
non_test
|
the popup edit form is not closed on the first update click if there was a invalid value before entering a valid one sass themes bug report the popup edit form is not closed on the first update click if there was an invalid value before entering a valid one sass themes reproduction of the problem the issue can be observed in our demo i made a video showcasing the issue environment kendo ui version browser
| 0
|
48,281
| 5,950,975,081
|
IssuesEvent
|
2017-05-26 18:12:55
|
openbmc/openbmc-test-automation
|
https://api.github.com/repos/openbmc/openbmc-test-automation
|
opened
|
[Network Test] Upload BMC work-around
|
SYS_TEST Test
|
The /tmp/images/ are specially reserved for code update. It doesnt' allow to upload unqualified file..
|
2.0
|
[Network Test] Upload BMC work-around - The /tmp/images/ are specially reserved for code update. It doesnt' allow to upload unqualified file..
|
test
|
upload bmc work around the tmp images are specially reserved for code update it doesnt allow to upload unqualified file
| 1
|
248,870
| 21,081,700,726
|
IssuesEvent
|
2022-04-03 01:33:19
|
Uuvana-Studios/longvinter-windows-client
|
https://api.github.com/repos/Uuvana-Studios/longvinter-windows-client
|
opened
|
Turrets not working even though electricity needs are met
|
Bug Not Tested
|
**Describe the bug**
Turrets not working even though electricity needs are met.
**To Reproduce**
Steps to reproduce the behavior:
1. Fully upgrade house down sustainability path.
2. Install one turret.
4. Turret does not work during night.
5. Install second turret.
6. Second turret does not work during the day.
**Expected behavior**
Meeting electricity requirements would mean that turrets work.
**Screenshots**

Fully upgraded sustainable power path

Turrets stop working at night

Daytime with one turret on, one turret off
**Desktop (please complete the following information):**
- OS: Windows
- Game Version 1.0.2
- Steam Version
**Additional context**
Other solar powered houses have multiple fully working turrets.
|
1.0
|
Turrets not working even though electricity needs are met - **Describe the bug**
Turrets not working even though electricity needs are met.
**To Reproduce**
Steps to reproduce the behavior:
1. Fully upgrade house down sustainability path.
2. Install one turret.
4. Turret does not work during night.
5. Install second turret.
6. Second turret does not work during the day.
**Expected behavior**
Meeting electricity requirements would mean that turrets work.
**Screenshots**

Fully upgraded sustainable power path

Turrets stop working at night

Daytime with one turret on, one turret off
**Desktop (please complete the following information):**
- OS: Windows
- Game Version 1.0.2
- Steam Version
**Additional context**
Other solar powered houses have multiple fully working turrets.
|
test
|
turrets not working even though electricity needs are met describe the bug turrets not working even though electricity needs are met to reproduce steps to reproduce the behavior fully upgrade house down sustainability path install one turret turret does not work during night install second turret second turret does not work during the day expected behavior meeting electricity requirements would mean that turrets work screenshots fully upgraded sustainable power path turrets stop working at night daytime with one turret on one turret off desktop please complete the following information os windows game version steam version additional context other solar powered houses have multiple fully working turrets
| 1
|
34,692
| 4,940,139,814
|
IssuesEvent
|
2016-11-29 16:06:44
|
flow123d/flow123d
|
https://api.github.com/repos/flow123d/flow123d
|
opened
|
Integration test update
|
New Tests
|
Go through all test input files:
- Add header in form:
# Test goal:
# Author:
# Source for reference: {text only, analytical, other SW, qualitative}
- update description
- [ ] Flow tests.
- [ ] Transport tests.
- [ ] Other tests.
|
1.0
|
Integration test update - Go through all test input files:
- Add header in form:
# Test goal:
# Author:
# Source for reference: {text only, analytical, other SW, qualitative}
- update description
- [ ] Flow tests.
- [ ] Transport tests.
- [ ] Other tests.
|
test
|
integration test update go through all test input files add header in form test goal author source for reference text only analytical other sw qualitative update description flow tests transport tests other tests
| 1
|
272,933
| 23,714,256,507
|
IssuesEvent
|
2022-08-30 10:22:58
|
marco-at-bit/test-quality-sdk
|
https://api.github.com/repos/marco-at-bit/test-quality-sdk
|
closed
|
Testing out status
|
bug TestQuality
|
## Precondition
This condition must be met
## Steps to Reproduce:
### Step 1 <span style="color:#FCF0BA"> **[Retest]** </span>
Step uno
#### Expected Result
Goes 1
### Step 2 <span style="color:#58b880"> **[Pass]** </span>
Second goes
#### Expected Result
Goes thwo
### Step 3 <span style="color:#58b880"> **[Pass]** </span>
The classic third
#### Expected Result
goe three
|
1.0
|
Testing out status - ## Precondition
This condition must be met
## Steps to Reproduce:
### Step 1 <span style="color:#FCF0BA"> **[Retest]** </span>
Step uno
#### Expected Result
Goes 1
### Step 2 <span style="color:#58b880"> **[Pass]** </span>
Second goes
#### Expected Result
Goes thwo
### Step 3 <span style="color:#58b880"> **[Pass]** </span>
The classic third
#### Expected Result
goe three
|
test
|
testing out status precondition this condition must be met steps to reproduce step step uno expected result goes step second goes expected result goes thwo step the classic third expected result goe three
| 1
|
91,760
| 10,727,992,907
|
IssuesEvent
|
2019-10-28 13:02:16
|
Timotheeee/PSIT3-HS19-IT18a_ZH-TeamIP2
|
https://api.github.com/repos/Timotheeee/PSIT3-HS19-IT18a_ZH-TeamIP2
|
closed
|
Lösungsarchitektur (M2)
|
documentation help wanted
|
− Use-Case-Modell
− Zusätzliche Anforderungen
− Domänenmodell
− Softwarearchitektur
− Design-Artefakte
− Implementation
− Projektmanagement
− Glossar
|
1.0
|
Lösungsarchitektur (M2) - − Use-Case-Modell
− Zusätzliche Anforderungen
− Domänenmodell
− Softwarearchitektur
− Design-Artefakte
− Implementation
− Projektmanagement
− Glossar
|
non_test
|
lösungsarchitektur − use case modell − zusätzliche anforderungen − domänenmodell − softwarearchitektur − design artefakte − implementation − projektmanagement − glossar
| 0
|
321,637
| 27,545,163,472
|
IssuesEvent
|
2023-03-07 11:13:49
|
giantswarm/roadmap
|
https://api.github.com/repos/giantswarm/roadmap
|
opened
|
Automated testing of CAPI management clusters
|
kind/epic area/kaas epic/capi component/mc-bootstrap topic/testing
|
## User Story
- As a Giant Swarm Engineer I want to run tests on Giant Swarm management clusters so that I can ensure their quality and reliability.
## Details, Background
Previous efforts from team Rocket around this topic can be tracked [here](https://github.com/giantswarm/roadmap/issues/685).
```[tasklist]
- [ ] https://github.com/giantswarm/roadmap/issues/2118
```
|
1.0
|
Automated testing of CAPI management clusters - ## User Story
- As a Giant Swarm Engineer I want to run tests on Giant Swarm management clusters so that I can ensure their quality and reliability.
## Details, Background
Previous efforts from team Rocket around this topic can be tracked [here](https://github.com/giantswarm/roadmap/issues/685).
```[tasklist]
- [ ] https://github.com/giantswarm/roadmap/issues/2118
```
|
test
|
automated testing of capi management clusters user story as a giant swarm engineer i want to run tests on giant swarm management clusters so that i can ensure their quality and reliability details background previous efforts from team rocket around this topic can be tracked
| 1
|
27,699
| 2,695,233,688
|
IssuesEvent
|
2015-04-02 02:48:45
|
neuropoly/spinalcordtoolbox
|
https://api.github.com/repos/neuropoly/spinalcordtoolbox
|
opened
|
smooth centerline not perfect at edge
|
enhancement priority: high sct_straighten_spinalcord
|
Data:
~~~~
git clone https://github.com/neuropoly/sct_example_data.git
~~~~
Syntax:
~~~~
sct_straighten_spinalcord -i t2.nii.gz -c t2_seg.nii.gz -v 3
~~~~
Output:

|
1.0
|
smooth centerline not perfect at edge - Data:
~~~~
git clone https://github.com/neuropoly/sct_example_data.git
~~~~
Syntax:
~~~~
sct_straighten_spinalcord -i t2.nii.gz -c t2_seg.nii.gz -v 3
~~~~
Output:

|
non_test
|
smooth centerline not perfect at edge data git clone syntax sct straighten spinalcord i nii gz c seg nii gz v output
| 0
|
368,221
| 25,782,754,444
|
IssuesEvent
|
2022-12-09 17:21:54
|
Peruz/icsd
|
https://api.github.com/repos/Peruz/icsd
|
opened
|
Create config file for CI tests with python 3.9 and 3.10
|
documentation enhancement
|
Up to now, the coverage test using the [CI workflow](https://github.com/Peruz/icsd/blob/main/.github/workflows/tests_package.yml) is **only done for python 3.8 with os ubuntu**. We need to **extend it to 3.9 and 3.10 and to macos and windows os**.
Adding the following two files in the test folder:
- tests/envs/39-env.yaml
```
name: test
channels:
- conda-forge
dependencies:
- python=3.9
- ...
```
- tests/envs/310-env.yaml
```
name: test
channels:
- conda-forge
dependencies:
- python=3.10
- ...
```
And the following strategy command in the [CI workflow](https://github.com/Peruz/icsd/blob/main/.github/workflows/tests_package.yml)
```
strategy:
fail-fast: false
matrix:
include:
- os: ubuntu-latest
env: tests/envs/38-env.yaml
py: py3.8
- os: ubuntu-latest
env: tests/envs/39-env.yaml
py: py3.9
- os: ubuntu-latest
env: tests/envs/310-env.yaml
py: py3.10
- os: macos-latest
env: tests/envs/310-env.yaml
py: py3.10
- os: windows-latest
env: tests/envs/310-env.yaml
py: py3.10
```
|
1.0
|
Create config file for CI tests with python 3.9 and 3.10 - Up to now, the coverage test using the [CI workflow](https://github.com/Peruz/icsd/blob/main/.github/workflows/tests_package.yml) is **only done for python 3.8 with os ubuntu**. We need to **extend it to 3.9 and 3.10 and to macos and windows os**.
Adding the following two files in the test folder:
- tests/envs/39-env.yaml
```
name: test
channels:
- conda-forge
dependencies:
- python=3.9
- ...
```
- tests/envs/310-env.yaml
```
name: test
channels:
- conda-forge
dependencies:
- python=3.10
- ...
```
And the following strategy command in the [CI workflow](https://github.com/Peruz/icsd/blob/main/.github/workflows/tests_package.yml)
```
strategy:
fail-fast: false
matrix:
include:
- os: ubuntu-latest
env: tests/envs/38-env.yaml
py: py3.8
- os: ubuntu-latest
env: tests/envs/39-env.yaml
py: py3.9
- os: ubuntu-latest
env: tests/envs/310-env.yaml
py: py3.10
- os: macos-latest
env: tests/envs/310-env.yaml
py: py3.10
- os: windows-latest
env: tests/envs/310-env.yaml
py: py3.10
```
|
non_test
|
create config file for ci tests with python and up to now the coverage test using the is only done for python with os ubuntu we need to extend it to and and to macos and windows os adding the following two files in the test folder tests envs env yaml name test channels conda forge dependencies python tests envs env yaml name test channels conda forge dependencies python and the following strategy command in the strategy fail fast false matrix include os ubuntu latest env tests envs env yaml py os ubuntu latest env tests envs env yaml py os ubuntu latest env tests envs env yaml py os macos latest env tests envs env yaml py os windows latest env tests envs env yaml py
| 0
|
156,396
| 19,849,669,405
|
IssuesEvent
|
2022-01-21 10:50:13
|
billmcchesney1/jazz
|
https://api.github.com/repos/billmcchesney1/jazz
|
opened
|
CVE-2022-21704 (Medium) detected in log4js-0.6.38.tgz, log4js-4.5.1.tgz
|
security vulnerability
|
## CVE-2022-21704 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>log4js-0.6.38.tgz</b>, <b>log4js-4.5.1.tgz</b></p></summary>
<p>
<details><summary><b>log4js-0.6.38.tgz</b></p></summary>
<p>Port of Log4js to work with node.</p>
<p>Library home page: <a href="https://registry.npmjs.org/log4js/-/log4js-0.6.38.tgz">https://registry.npmjs.org/log4js/-/log4js-0.6.38.tgz</a></p>
<p>Path to dependency file: /core/jazz_ui/src/app/primary-components/daterange-picker/ng2-datepicker/package.json</p>
<p>Path to vulnerable library: /core/jazz_ui/src/app/primary-components/daterange-picker/ng2-datepicker/node_modules/log4js/package.json,/core/jazz_ui/node_modules/log4js/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.4.1.tgz (Root Library)
- :x: **log4js-0.6.38.tgz** (Vulnerable Library)
</details>
<details><summary><b>log4js-4.5.1.tgz</b></p></summary>
<p>Port of Log4js to work with node.</p>
<p>Library home page: <a href="https://registry.npmjs.org/log4js/-/log4js-4.5.1.tgz">https://registry.npmjs.org/log4js/-/log4js-4.5.1.tgz</a></p>
<p>Path to dependency file: /templates/angular-website-template/app/package.json</p>
<p>Path to vulnerable library: /templates/angular-website-template/app/node_modules/log4js/package.json</p>
<p>
Dependency Hierarchy:
- karma-4.0.1.tgz (Root Library)
- :x: **log4js-4.5.1.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
log4js-node is a port of log4js to node.js. In affected versions default file permissions for log files created by the file, fileSync and dateFile appenders are world-readable (in unix). This could cause problems if log files contain sensitive information. This would affect any users that have not supplied their own permissions for the files via the mode parameter in the config. Users are advised to update.
<p>Publish Date: 2022-01-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21704>CVE-2022-21704</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/log4js-node/log4js-node/security/advisories/GHSA-82v2-mx6x-wq7q">https://github.com/log4js-node/log4js-node/security/advisories/GHSA-82v2-mx6x-wq7q</a></p>
<p>Release Date: 2022-01-19</p>
<p>Fix Resolution: log4js - 6.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"log4js","packageVersion":"0.6.38","packageFilePaths":["/core/jazz_ui/src/app/primary-components/daterange-picker/ng2-datepicker/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma:1.4.1;log4js:0.6.38","isMinimumFixVersionAvailable":true,"minimumFixVersion":"log4js - 6.4.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"log4js","packageVersion":"4.5.1","packageFilePaths":["/templates/angular-website-template/app/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma:4.0.1;log4js:4.5.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"log4js - 6.4.0","isBinary":false}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2022-21704","vulnerabilityDetails":"log4js-node is a port of log4js to node.js. In affected versions default file permissions for log files created by the file, fileSync and dateFile appenders are world-readable (in unix). This could cause problems if log files contain sensitive information. This would affect any users that have not supplied their own permissions for the files via the mode parameter in the config. Users are advised to update.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21704","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2022-21704 (Medium) detected in log4js-0.6.38.tgz, log4js-4.5.1.tgz - ## CVE-2022-21704 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>log4js-0.6.38.tgz</b>, <b>log4js-4.5.1.tgz</b></p></summary>
<p>
<details><summary><b>log4js-0.6.38.tgz</b></p></summary>
<p>Port of Log4js to work with node.</p>
<p>Library home page: <a href="https://registry.npmjs.org/log4js/-/log4js-0.6.38.tgz">https://registry.npmjs.org/log4js/-/log4js-0.6.38.tgz</a></p>
<p>Path to dependency file: /core/jazz_ui/src/app/primary-components/daterange-picker/ng2-datepicker/package.json</p>
<p>Path to vulnerable library: /core/jazz_ui/src/app/primary-components/daterange-picker/ng2-datepicker/node_modules/log4js/package.json,/core/jazz_ui/node_modules/log4js/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.4.1.tgz (Root Library)
- :x: **log4js-0.6.38.tgz** (Vulnerable Library)
</details>
<details><summary><b>log4js-4.5.1.tgz</b></p></summary>
<p>Port of Log4js to work with node.</p>
<p>Library home page: <a href="https://registry.npmjs.org/log4js/-/log4js-4.5.1.tgz">https://registry.npmjs.org/log4js/-/log4js-4.5.1.tgz</a></p>
<p>Path to dependency file: /templates/angular-website-template/app/package.json</p>
<p>Path to vulnerable library: /templates/angular-website-template/app/node_modules/log4js/package.json</p>
<p>
Dependency Hierarchy:
- karma-4.0.1.tgz (Root Library)
- :x: **log4js-4.5.1.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
log4js-node is a port of log4js to node.js. In affected versions default file permissions for log files created by the file, fileSync and dateFile appenders are world-readable (in unix). This could cause problems if log files contain sensitive information. This would affect any users that have not supplied their own permissions for the files via the mode parameter in the config. Users are advised to update.
<p>Publish Date: 2022-01-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21704>CVE-2022-21704</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/log4js-node/log4js-node/security/advisories/GHSA-82v2-mx6x-wq7q">https://github.com/log4js-node/log4js-node/security/advisories/GHSA-82v2-mx6x-wq7q</a></p>
<p>Release Date: 2022-01-19</p>
<p>Fix Resolution: log4js - 6.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"log4js","packageVersion":"0.6.38","packageFilePaths":["/core/jazz_ui/src/app/primary-components/daterange-picker/ng2-datepicker/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma:1.4.1;log4js:0.6.38","isMinimumFixVersionAvailable":true,"minimumFixVersion":"log4js - 6.4.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"log4js","packageVersion":"4.5.1","packageFilePaths":["/templates/angular-website-template/app/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma:4.0.1;log4js:4.5.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"log4js - 6.4.0","isBinary":false}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2022-21704","vulnerabilityDetails":"log4js-node is a port of log4js to node.js. In affected versions default file permissions for log files created by the file, fileSync and dateFile appenders are world-readable (in unix). This could cause problems if log files contain sensitive information. This would affect any users that have not supplied their own permissions for the files via the mode parameter in the config. Users are advised to update.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21704","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve medium detected in tgz tgz cve medium severity vulnerability vulnerable libraries tgz tgz tgz port of to work with node library home page a href path to dependency file core jazz ui src app primary components daterange picker datepicker package json path to vulnerable library core jazz ui src app primary components daterange picker datepicker node modules package json core jazz ui node modules package json dependency hierarchy karma tgz root library x tgz vulnerable library tgz port of to work with node library home page a href path to dependency file templates angular website template app package json path to vulnerable library templates angular website template app node modules package json dependency hierarchy karma tgz root library x tgz vulnerable library found in base branch develop vulnerability details node is a port of to node js in affected versions default file permissions for log files created by the file filesync and datefile appenders are world readable in unix this could cause problems if log files contain sensitive information this would affect any users that have not supplied their own permissions for the files via the mode parameter in the config users are advised to update publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree karma isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename packageversion packagefilepaths istransitivedependency true dependencytree karma isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails node is a port of to node js in affected versions default file permissions for log files created by the file filesync and datefile appenders are world readable in unix this could cause problems if log files contain sensitive information this would affect any users that have not supplied their own permissions for the files via the mode parameter in the config users are advised to update vulnerabilityurl
| 0
|
599,349
| 18,271,618,090
|
IssuesEvent
|
2021-10-04 14:21:12
|
ita-social-projects/TeachUA
|
https://api.github.com/repos/ita-social-projects/TeachUA
|
closed
|
[Advanced Search] "Вік дитини" field can accept unlimited letters via 'paste' function
|
bug Priority: Medium
|
**Environment:** Windows 10 21H1 64bit, Chrome Version 93.0.4577.63 (Official Build) (64-bit)
**Reproducible:** always
**Build found:** latest on dev and prod
**Preconditions**
https://speak-ukrainian.org.ua/dev/
**Steps to reproduce**
1. Go to https://speak-ukrainian.org.ua/dev/
2. Click on advanced search button
3. Paste 300+ letters into "Вік дитини" field
**Actual result**
Unlimited letters can be paste into the "Вік дитини" field
**Expected result**
"Вік дитини" field should not allow to paste letters
**Labels to be added**
"Bug", Priority ("pri: mid"), Severity ("severity:low"), Type ("Functional"), "API" (for back-end bugs).
|
1.0
|
[Advanced Search] "Вік дитини" field can accept unlimited letters via 'paste' function - **Environment:** Windows 10 21H1 64bit, Chrome Version 93.0.4577.63 (Official Build) (64-bit)
**Reproducible:** always
**Build found:** latest on dev and prod
**Preconditions**
https://speak-ukrainian.org.ua/dev/
**Steps to reproduce**
1. Go to https://speak-ukrainian.org.ua/dev/
2. Click on advanced search button
3. Paste 300+ letters into "Вік дитини" field
**Actual result**
Unlimited letters can be paste into the "Вік дитини" field
**Expected result**
"Вік дитини" field should not allow to paste letters
**Labels to be added**
"Bug", Priority ("pri: mid"), Severity ("severity:low"), Type ("Functional"), "API" (for back-end bugs).
|
non_test
|
вік дитини field can accept unlimited letters via paste function environment windows chrome version official build bit reproducible always build found latest on dev and prod preconditions steps to reproduce go to click on advanced search button paste letters into вік дитини field actual result unlimited letters can be paste into the вік дитини field expected result вік дитини field should not allow to paste letters labels to be added bug priority pri mid severity severity low type functional api for back end bugs
| 0
|
163,648
| 12,739,860,418
|
IssuesEvent
|
2020-06-26 00:29:25
|
jensengroup/propka
|
https://api.github.com/repos/jensengroup/propka
|
closed
|
report coverage from CI
|
continuous integration testing
|
With PR #62 merged, we should use coverage reporting and make it part of the CI reporting.
|
1.0
|
report coverage from CI - With PR #62 merged, we should use coverage reporting and make it part of the CI reporting.
|
test
|
report coverage from ci with pr merged we should use coverage reporting and make it part of the ci reporting
| 1
|
2,834
| 2,641,060,798
|
IssuesEvent
|
2015-03-11 15:49:51
|
bpmn-miwg/bpmn-miwg-test-suite
|
https://api.github.com/repos/bpmn-miwg/bpmn-miwg-test-suite
|
closed
|
Participant missing in collaboration
|
bug Test: A.4.0
|
The model A.4.0 has two processes, obviously collaborating. The collaboration element has only one participant. The specification has 3 relevant statements to be considered for this issue:
- “One, and only one, Pool in a diagram MAY be presented without a boundary. If there is more than one Pool in the diagram, then the remaining Pools MUST have a boundary”
- “A Collaboration can contain two (2) or more Pools (i.e., Participants). However, a Process that represents the work performed from the point of view of the modeler or the modeler’s organization can be considered “internal” and is NOT REQUIRED to be surrounded by the boundary of the Pool, while the other Pools in the Diagram MUST have their boundary (see Figure 9.5).”
- "A Pool is the graphical representation of a Participant in a Collaboration."
According to the aforementioned statements, the collaboration element should have two participants, but no BPMNShape representing the second participant as the process is considered “internal”.
|
1.0
|
Participant missing in collaboration - The model A.4.0 has two processes, obviously collaborating. The collaboration element has only one participant. The specification has 3 relevant statements to be considered for this issue:
- “One, and only one, Pool in a diagram MAY be presented without a boundary. If there is more than one Pool in the diagram, then the remaining Pools MUST have a boundary”
- “A Collaboration can contain two (2) or more Pools (i.e., Participants). However, a Process that represents the work performed from the point of view of the modeler or the modeler’s organization can be considered “internal” and is NOT REQUIRED to be surrounded by the boundary of the Pool, while the other Pools in the Diagram MUST have their boundary (see Figure 9.5).”
- "A Pool is the graphical representation of a Participant in a Collaboration."
According to the aforementioned statements, the collaboration element should have two participants, but no BPMNShape representing the second participant as the process is considered “internal”.
|
test
|
participant missing in collaboration the model a has two processes obviously collaborating the collaboration element has only one participant the specification has relevant statements to be considered for this issue “one and only one pool in a diagram may be presented without a boundary if there is more than one pool in the diagram then the remaining pools must have a boundary” “a collaboration can contain two or more pools i e participants however a process that represents the work performed from the point of view of the modeler or the modeler’s organization can be considered “internal” and is not required to be surrounded by the boundary of the pool while the other pools in the diagram must have their boundary see figure ” a pool is the graphical representation of a participant in a collaboration according to the aforementioned statements the collaboration element should have two participants but no bpmnshape representing the second participant as the process is considered “internal”
| 1
|
138,318
| 11,198,884,803
|
IssuesEvent
|
2020-01-03 17:06:28
|
Princeton-CDH/ppa-django
|
https://api.github.com/repos/Princeton-CDH/ppa-django
|
closed
|
As a content editor, I want to create linkable anchors in documents so that I can reference specific sections of my content on other pages.
|
awaiting testing
|
## testing notes
This feature was adapted from mep-django, where it has already been tested and accepted. See testing notes and comments on https://github.com/Princeton-CDH/mep-django/issues/464
Probably useful to run through similar testing steps to make sure you're comfortable with it, and also check that the display for PPA looks ok (link styles are slightly different).
Feel free to compare with the mep-django version of the feature in production; it's currently in use for the FAQ sections: https://shakespeareandco.princeton.edu/about/faq/
I thought this feature would be valuable to bring over to use with the most/least collected items listed in my essay, so that we can link directly to them and make them easier to refer to.
|
1.0
|
As a content editor, I want to create linkable anchors in documents so that I can reference specific sections of my content on other pages. - ## testing notes
This feature was adapted from mep-django, where it has already been tested and accepted. See testing notes and comments on https://github.com/Princeton-CDH/mep-django/issues/464
Probably useful to run through similar testing steps to make sure you're comfortable with it, and also check that the display for PPA looks ok (link styles are slightly different).
Feel free to compare with the mep-django version of the feature in production; it's currently in use for the FAQ sections: https://shakespeareandco.princeton.edu/about/faq/
I thought this feature would be valuable to bring over to use with the most/least collected items listed in my essay, so that we can link directly to them and make them easier to refer to.
|
test
|
as a content editor i want to create linkable anchors in documents so that i can reference specific sections of my content on other pages testing notes this feature was adapted from mep django where it has already been tested and accepted see testing notes and comments on probably useful to run through similar testing steps to make sure you re comfortable with it and also check that the display for ppa looks ok link styles are slightly different feel free to compare with the mep django version of the feature in production it s currently in use for the faq sections i thought this feature would be valuable to bring over to use with the most least collected items listed in my essay so that we can link directly to them and make them easier to refer to
| 1
|
42,783
| 22,839,710,948
|
IssuesEvent
|
2022-07-12 20:24:53
|
pulumi/pulumi
|
https://api.github.com/repos/pulumi/pulumi
|
closed
|
[Python SDK] slow performance with large CRD in helm chart.
|
kind/bug impact/performance language/python
|
<!--- Provide a general summary of the issue -->
## Expected Behavior
<!--- Tell us what should happen -->
`pulumi up` is fast just like in typescript ~5 seconds
[Typescript SDK TRACE out.txt](https://github.com/pulumi/pulumi/files/5795889/out.txt)
## Current Behavior
<!--- Tell us what happens instead of the expected behavior -->
`pulumi up` with python is very very slow ~1 hour !
[Python SDK TRACE out.txt.zip](https://github.com/pulumi/pulumi/files/5796026/out.txt.zip)
(I didn't wait and canceled the command `pulumi up`, very long wait)
## Steps to Reproduce
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1) Chart with a large CRD yaml size ~10k lines
Example:
https://github.com/VictoriaMetrics/helm-charts/blob/master/charts/victoria-metrics-operator/templates/crd.yaml
https://github.com/VictoriaMetrics/helm-charts/tree/master/charts/victoria-metrics-operator
2) use python sdk to install chart:
```
from pulumi_kubernetes.helm import v3
from pulumi_kubernetes.helm.v3.helm import ChartOpts
vm_chart = v3.Chart("vm", ChartOpts(chart="victoria-metrics-operator", repo="vm"))
```
## Context (Environment)
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
<!--- Provide a general summary of the issue in the Title above -->
```
macOS Big Sur 11.1
Python 3.9.1
pulumi v2.17.0
pulumi-kubernetes==2.7.4
```
|
True
|
[Python SDK] slow performance with large CRD in helm chart. - <!--- Provide a general summary of the issue -->
## Expected Behavior
<!--- Tell us what should happen -->
`pulumi up` is fast just like in typescript ~5 seconds
[Typescript SDK TRACE out.txt](https://github.com/pulumi/pulumi/files/5795889/out.txt)
## Current Behavior
<!--- Tell us what happens instead of the expected behavior -->
`pulumi up` with python is very very slow ~1 hour !
[Python SDK TRACE out.txt.zip](https://github.com/pulumi/pulumi/files/5796026/out.txt.zip)
(I didn't wait and canceled the command `pulumi up`, very long wait)
## Steps to Reproduce
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1) Chart with a large CRD yaml size ~10k lines
Example:
https://github.com/VictoriaMetrics/helm-charts/blob/master/charts/victoria-metrics-operator/templates/crd.yaml
https://github.com/VictoriaMetrics/helm-charts/tree/master/charts/victoria-metrics-operator
2) use python sdk to install chart:
```
from pulumi_kubernetes.helm import v3
from pulumi_kubernetes.helm.v3.helm import ChartOpts
vm_chart = v3.Chart("vm", ChartOpts(chart="victoria-metrics-operator", repo="vm"))
```
## Context (Environment)
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
<!--- Provide a general summary of the issue in the Title above -->
```
macOS Big Sur 11.1
Python 3.9.1
pulumi v2.17.0
pulumi-kubernetes==2.7.4
```
|
non_test
|
slow performance with large crd in helm chart expected behavior pulumi up is fast just like in typescript seconds current behavior pulumi up with python is very very slow hour i didn t wait and canceled the command pulumi up very long wait steps to reproduce chart with a large crd yaml size lines example use python sdk to install chart from pulumi kubernetes helm import from pulumi kubernetes helm helm import chartopts vm chart chart vm chartopts chart victoria metrics operator repo vm context environment macos big sur python pulumi pulumi kubernetes
| 0
|
773,516
| 27,160,398,636
|
IssuesEvent
|
2023-02-17 11:21:21
|
status-im/status-mobile
|
https://api.github.com/repos/status-im/status-mobile
|
closed
|
Some components disappear from the list of pinned messages after chat reopening
|
bug low-priority pinned messages
|
**Steps to reproduce:**
1. Go to the 1-1 chat
2. Send a message
3. Pin the message
4. Reopen the chat
**Actual result:**
The following info is dissapaered:
- Owner and his chatkey
- Timemark
- Pin
- Style for pinned messages
https://user-images.githubusercontent.com/52490791/217331367-b65df057-1dd0-43ce-baeb-5cdddedd8dd8.mp4
**Expected result:**

**ENV:**
- Nightly 7 Feb 2023 [2619]
|
1.0
|
Some components disappear from the list of pinned messages after chat reopening - **Steps to reproduce:**
1. Go to the 1-1 chat
2. Send a message
3. Pin the message
4. Reopen the chat
**Actual result:**
The following info is dissapaered:
- Owner and his chatkey
- Timemark
- Pin
- Style for pinned messages
https://user-images.githubusercontent.com/52490791/217331367-b65df057-1dd0-43ce-baeb-5cdddedd8dd8.mp4
**Expected result:**

**ENV:**
- Nightly 7 Feb 2023 [2619]
|
non_test
|
some components disappear from the list of pinned messages after chat reopening steps to reproduce go to the chat send a message pin the message reopen the chat actual result the following info is dissapaered owner and his chatkey timemark pin style for pinned messages expected result env nightly feb
| 0
|
797,730
| 28,153,593,236
|
IssuesEvent
|
2023-04-03 05:04:43
|
PrefectHQ/prefect
|
https://api.github.com/repos/PrefectHQ/prefect
|
closed
|
Subpaths and remote subdirectory paths can't be used on deployments with GCS storage block
|
bug status:accepted v2 priority:medium
|
### First check
- [X] I added a descriptive title to this issue.
- [X] I used the GitHub search to find a similar issue and didn't find it.
- [X] I searched the Prefect documentation for this issue.
- [X] I checked that this issue is related to Prefect and not one of its dependencies.
### Bug summary
If you create a deployment with a storage block that has a directory as part of its "base path" _and_ a `path`, the files are not downloaded correctly from the storage and the flow run fails.
(GCS) Storage Blocks can have a directory added [as part of their bucket path](https://docs.prefect.io/tutorials/storage/#create-a-storage-block).
[As described in the tutorial](https://docs.prefect.io/tutorials/storage/#sharing-storage), you can also share Storage Blocks between deployments and put each deployment's file in its own subdirectory by making use of `path` on a deployment.
---
I also tested by removing the directory from the "base path". Then it runs fine. When I add a `path`, it still runs fine.
When I add another folder level, it fails (see Reproduction method 2).
### Reproduction
There's two ways to get multiple dir levels as part of the flow storage.
#### Method 1:
I followed the tutorial (https://docs.prefect.io/tutorials/deployments/#setting-up).
Differences:
- I'm running inside a Poetry venv.
- My GCS block has an additional directory as part of its base path:

Everything works fine if the deployment does not have a `path` specified:
```
poetry run prefect deployment build ./log_flow.py:log_flow -n log-simple -q test
-sb gcs/poc-gcs-bucket-dev --apply
```
When adding a `path`, then running a flow from that deployment, fails.
```
poetry run prefect deployment build ./log_flow.py:log_flow -n log-simple -q test
-sb gcs/poc-gcs-bucket-dev --path log-flow-directory --apply
```
Errors posted in error section.
---
#### Method 2:
Remove the base path from the storage block.
Add more than one level to the path here:
```
poetry run prefect deployment build ./log_flow.py:log_flow -n log-simple -q test
-sb gcs/poc-gcs-bucket-dev --path testing/log-flow-directory --apply
```
### Error
<details>
<summary>On Windows</summary>
```
09:37:13.657 | INFO | prefect.agent - Submitting flow run '926654d1-b5fc-4679-a0e6-01acf2bb70c2'
09:37:14.538 | INFO | prefect.infrastructure.process - Opening process 'omicron3-firiendfeld-point'...
09:37:14.573 | INFO | prefect.agent - Completed submission of flow run '926654d1-b5fc-4679-a0e6-01acf2bb70c2'
09:37:19.042 | ERROR | Flow run 'omicron3-firiendfeld-point' - Flow could not be retrieved from deployment.
Traceback (most recent call last):
File "<frozen importlib._bootstrap_external>", line 879, in exec_module
File "<frozen importlib._bootstrap_external>", line 1016, in get_code
File "<frozen importlib._bootstrap_external>", line 1073, in get_data
FileNotFoundError: [Errno 2] No such file or directory: 'C:\\Users\\STALJA~1\\AppData\\Local\\Temp\\tmpal0n0yjbprefect\\log_flow.py'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\engine.py", line 256, in retrieve_flow_then_begin_flow_run
flow = await load_flow_from_flow_run(flow_run, client=client)
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\client.py", line 103, in with_injected_client
return await fn(*args, **kwargs)
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\deployments.py", line 69, in load_flow_from_flow_run
flow = await run_sync_in_worker_thread(import_object, str(import_path))
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\utilities\asyncutils.py", line 57, in run_sync_in_worker_thread
return await anyio.to_thread.run_sync(call, cancellable=True)
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\anyio\to_thread.py", line 31, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\anyio\_backends\_asyncio.py", line 937, in run_sync_in_worker_thread
return await future
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\anyio\_backends\_asyncio.py", line 867, in run
result = context.run(func, *args)
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\utilities\importtools.py", line 193, in import_object
module = load_script_as_module(script_path)
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\utilities\importtools.py", line 156, in load_script_as_module
raise ScriptError(user_exc=exc, path=path) from exc
prefect.exceptions.ScriptError: Script at 'log_flow.py' encountered an exception
09:37:20.008 | INFO | prefect.infrastructure.process - Process 'omicron3-firiendfeld-point' exited cleanly.
```
</details>
<details>
<summary>On Ubuntu</summary>
```
08:46:12.139 | INFO | prefect.infrastructure.process - Opening process 'eta2-nivoch-nebula'...
08:46:12.143 | INFO | prefect.agent - Completed submission of flow run '8116942b-acfb-495b-8386-975d0aa6980e'
08:46:16.120 | ERROR | Flow run 'eta2-nivoch-nebula' - Flow could not be retrieved from deployment.
Traceback (most recent call last):
File "<frozen importlib._bootstrap_external>", line 879, in exec_module
File "<frozen importlib._bootstrap_external>", line 1016, in get_code
File "<frozen importlib._bootstrap_external>", line 1073, in get_data
FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmp8lc3zuayprefect/log_flow.py'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/engine.py", line 256, in retrieve_flow_then_begin_flow_run
flow = await load_flow_from_flow_run(flow_run, client=client)
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/client.py", line 103, in with_injected_client
return await fn(*args, **kwargs)
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/deployments.py", line 69, in load_flow_from_flow_run
flow = await run_sync_in_worker_thread(import_object, str(import_path))
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/utilities/asyncutils.py", line 57, in run_sync_in_worker_thread
return await anyio.to_thread.run_sync(call, cancellable=True)
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/anyio/to_thread.py", line 31, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
return await future
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 867, in run
result = context.run(func, *args)
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/utilities/importtools.py", line 193, in import_object
module = load_script_as_module(script_path)
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/utilities/importtools.py", line 156, in load_script_as_module
raise ScriptError(user_exc=exc, path=path) from exc
prefect.exceptions.ScriptError: Script at 'log_flow.py' encountered an exception
08:46:17.028 | INFO | prefect.infrastructure.process - Process 'eta2-nivoch-nebula' exited cleanly.
```
</details>
### Versions
<details>
<summary>On Windows</summary>
```
Version: 2.4.0
API version: 0.8.0
Python version: 3.10.7
Git commit: 513639e8
Built: Tue, Sep 13, 2022 2:15 PM
OS/Arch: win32/AMD64
Profile: eta-cloud
Server type: cloud
```
</details>
<details>
<summary>On Ubuntu</summary>
```
Version: 2.4.0
API version: 0.8.0
Python version: 3.10.4
Git commit: 513639e8
Built: Tue, Sep 13, 2022 2:15 PM
OS/Arch: linux/x86_64
Profile: cloud
Server type: cloud
```
</details>
### Additional context
I was only able to test using the GCS block.
I tested on Windows and Ubuntu via WSL.
I'm running everything inside the same Poetry venv.
|
1.0
|
Subpaths and remote subdirectory paths can't be used on deployments with GCS storage block - ### First check
- [X] I added a descriptive title to this issue.
- [X] I used the GitHub search to find a similar issue and didn't find it.
- [X] I searched the Prefect documentation for this issue.
- [X] I checked that this issue is related to Prefect and not one of its dependencies.
### Bug summary
If you create a deployment with a storage block that has a directory as part of its "base path" _and_ a `path`, the files are not downloaded correctly from the storage and the flow run fails.
(GCS) Storage Blocks can have a directory added [as part of their bucket path](https://docs.prefect.io/tutorials/storage/#create-a-storage-block).
[As described in the tutorial](https://docs.prefect.io/tutorials/storage/#sharing-storage), you can also share Storage Blocks between deployments and put each deployment's file in its own subdirectory by making use of `path` on a deployment.
---
I also tested by removing the directory from the "base path". Then it runs fine. When I add a `path`, it still runs fine.
When I add another folder level, it fails (see Reproduction method 2).
### Reproduction
There's two ways to get multiple dir levels as part of the flow storage.
#### Method 1:
I followed the tutorial (https://docs.prefect.io/tutorials/deployments/#setting-up).
Differences:
- I'm running inside a Poetry venv.
- My GCS block has an additional directory as part of its base path:

Everything works fine if the deployment does not have a `path` specified:
```
poetry run prefect deployment build ./log_flow.py:log_flow -n log-simple -q test
-sb gcs/poc-gcs-bucket-dev --apply
```
When adding a `path`, then running a flow from that deployment, fails.
```
poetry run prefect deployment build ./log_flow.py:log_flow -n log-simple -q test
-sb gcs/poc-gcs-bucket-dev --path log-flow-directory --apply
```
Errors posted in error section.
---
#### Method 2:
Remove the base path from the storage block.
Add more than one level to the path here:
```
poetry run prefect deployment build ./log_flow.py:log_flow -n log-simple -q test
-sb gcs/poc-gcs-bucket-dev --path testing/log-flow-directory --apply
```
### Error
<details>
<summary>On Windows</summary>
```
09:37:13.657 | INFO | prefect.agent - Submitting flow run '926654d1-b5fc-4679-a0e6-01acf2bb70c2'
09:37:14.538 | INFO | prefect.infrastructure.process - Opening process 'omicron3-firiendfeld-point'...
09:37:14.573 | INFO | prefect.agent - Completed submission of flow run '926654d1-b5fc-4679-a0e6-01acf2bb70c2'
09:37:19.042 | ERROR | Flow run 'omicron3-firiendfeld-point' - Flow could not be retrieved from deployment.
Traceback (most recent call last):
File "<frozen importlib._bootstrap_external>", line 879, in exec_module
File "<frozen importlib._bootstrap_external>", line 1016, in get_code
File "<frozen importlib._bootstrap_external>", line 1073, in get_data
FileNotFoundError: [Errno 2] No such file or directory: 'C:\\Users\\STALJA~1\\AppData\\Local\\Temp\\tmpal0n0yjbprefect\\log_flow.py'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\engine.py", line 256, in retrieve_flow_then_begin_flow_run
flow = await load_flow_from_flow_run(flow_run, client=client)
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\client.py", line 103, in with_injected_client
return await fn(*args, **kwargs)
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\deployments.py", line 69, in load_flow_from_flow_run
flow = await run_sync_in_worker_thread(import_object, str(import_path))
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\utilities\asyncutils.py", line 57, in run_sync_in_worker_thread
return await anyio.to_thread.run_sync(call, cancellable=True)
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\anyio\to_thread.py", line 31, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\anyio\_backends\_asyncio.py", line 937, in run_sync_in_worker_thread
return await future
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\anyio\_backends\_asyncio.py", line 867, in run
result = context.run(func, *args)
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\utilities\importtools.py", line 193, in import_object
module = load_script_as_module(script_path)
File "C:\Users\staljaard\Source\data-integration-poc\.venv\lib\site-packages\prefect\utilities\importtools.py", line 156, in load_script_as_module
raise ScriptError(user_exc=exc, path=path) from exc
prefect.exceptions.ScriptError: Script at 'log_flow.py' encountered an exception
09:37:20.008 | INFO | prefect.infrastructure.process - Process 'omicron3-firiendfeld-point' exited cleanly.
```
</details>
<details>
<summary>On Ubuntu</summary>
```
08:46:12.139 | INFO | prefect.infrastructure.process - Opening process 'eta2-nivoch-nebula'...
08:46:12.143 | INFO | prefect.agent - Completed submission of flow run '8116942b-acfb-495b-8386-975d0aa6980e'
08:46:16.120 | ERROR | Flow run 'eta2-nivoch-nebula' - Flow could not be retrieved from deployment.
Traceback (most recent call last):
File "<frozen importlib._bootstrap_external>", line 879, in exec_module
File "<frozen importlib._bootstrap_external>", line 1016, in get_code
File "<frozen importlib._bootstrap_external>", line 1073, in get_data
FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmp8lc3zuayprefect/log_flow.py'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/engine.py", line 256, in retrieve_flow_then_begin_flow_run
flow = await load_flow_from_flow_run(flow_run, client=client)
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/client.py", line 103, in with_injected_client
return await fn(*args, **kwargs)
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/deployments.py", line 69, in load_flow_from_flow_run
flow = await run_sync_in_worker_thread(import_object, str(import_path))
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/utilities/asyncutils.py", line 57, in run_sync_in_worker_thread
return await anyio.to_thread.run_sync(call, cancellable=True)
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/anyio/to_thread.py", line 31, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
return await future
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 867, in run
result = context.run(func, *args)
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/utilities/importtools.py", line 193, in import_object
module = load_script_as_module(script_path)
File "/home/stephan/di_poc/.venv/lib/python3.10/site-packages/prefect/utilities/importtools.py", line 156, in load_script_as_module
raise ScriptError(user_exc=exc, path=path) from exc
prefect.exceptions.ScriptError: Script at 'log_flow.py' encountered an exception
08:46:17.028 | INFO | prefect.infrastructure.process - Process 'eta2-nivoch-nebula' exited cleanly.
```
</details>
### Versions
<details>
<summary>On Windows</summary>
```
Version: 2.4.0
API version: 0.8.0
Python version: 3.10.7
Git commit: 513639e8
Built: Tue, Sep 13, 2022 2:15 PM
OS/Arch: win32/AMD64
Profile: eta-cloud
Server type: cloud
```
</details>
<details>
<summary>On Ubuntu</summary>
```
Version: 2.4.0
API version: 0.8.0
Python version: 3.10.4
Git commit: 513639e8
Built: Tue, Sep 13, 2022 2:15 PM
OS/Arch: linux/x86_64
Profile: cloud
Server type: cloud
```
</details>
### Additional context
I was only able to test using the GCS block.
I tested on Windows and Ubuntu via WSL.
I'm running everything inside the same Poetry venv.
|
non_test
|
subpaths and remote subdirectory paths can t be used on deployments with gcs storage block first check i added a descriptive title to this issue i used the github search to find a similar issue and didn t find it i searched the prefect documentation for this issue i checked that this issue is related to prefect and not one of its dependencies bug summary if you create a deployment with a storage block that has a directory as part of its base path and a path the files are not downloaded correctly from the storage and the flow run fails gcs storage blocks can have a directory added you can also share storage blocks between deployments and put each deployment s file in its own subdirectory by making use of path on a deployment i also tested by removing the directory from the base path then it runs fine when i add a path it still runs fine when i add another folder level it fails see reproduction method reproduction there s two ways to get multiple dir levels as part of the flow storage method i followed the tutorial differences i m running inside a poetry venv my gcs block has an additional directory as part of its base path everything works fine if the deployment does not have a path specified poetry run prefect deployment build log flow py log flow n log simple q test sb gcs poc gcs bucket dev apply when adding a path then running a flow from that deployment fails poetry run prefect deployment build log flow py log flow n log simple q test sb gcs poc gcs bucket dev path log flow directory apply errors posted in error section method remove the base path from the storage block add more than one level to the path here poetry run prefect deployment build log flow py log flow n log simple q test sb gcs poc gcs bucket dev path testing log flow directory apply error on windows info prefect agent submitting flow run info prefect infrastructure process opening process firiendfeld point info prefect agent completed submission of flow run error flow run firiendfeld point flow could not be retrieved from deployment traceback most recent call last file line in exec module file line in get code file line in get data filenotfounderror no such file or directory c users stalja appdata local temp log flow py the above exception was the direct cause of the following exception traceback most recent call last file c users staljaard source data integration poc venv lib site packages prefect engine py line in retrieve flow then begin flow run flow await load flow from flow run flow run client client file c users staljaard source data integration poc venv lib site packages prefect client py line in with injected client return await fn args kwargs file c users staljaard source data integration poc venv lib site packages prefect deployments py line in load flow from flow run flow await run sync in worker thread import object str import path file c users staljaard source data integration poc venv lib site packages prefect utilities asyncutils py line in run sync in worker thread return await anyio to thread run sync call cancellable true file c users staljaard source data integration poc venv lib site packages anyio to thread py line in run sync return await get asynclib run sync in worker thread file c users staljaard source data integration poc venv lib site packages anyio backends asyncio py line in run sync in worker thread return await future file c users staljaard source data integration poc venv lib site packages anyio backends asyncio py line in run result context run func args file c users staljaard source data integration poc venv lib site packages prefect utilities importtools py line in import object module load script as module script path file c users staljaard source data integration poc venv lib site packages prefect utilities importtools py line in load script as module raise scripterror user exc exc path path from exc prefect exceptions scripterror script at log flow py encountered an exception info prefect infrastructure process process firiendfeld point exited cleanly on ubuntu info prefect infrastructure process opening process nivoch nebula info prefect agent completed submission of flow run acfb error flow run nivoch nebula flow could not be retrieved from deployment traceback most recent call last file line in exec module file line in get code file line in get data filenotfounderror no such file or directory tmp log flow py the above exception was the direct cause of the following exception traceback most recent call last file home stephan di poc venv lib site packages prefect engine py line in retrieve flow then begin flow run flow await load flow from flow run flow run client client file home stephan di poc venv lib site packages prefect client py line in with injected client return await fn args kwargs file home stephan di poc venv lib site packages prefect deployments py line in load flow from flow run flow await run sync in worker thread import object str import path file home stephan di poc venv lib site packages prefect utilities asyncutils py line in run sync in worker thread return await anyio to thread run sync call cancellable true file home stephan di poc venv lib site packages anyio to thread py line in run sync return await get asynclib run sync in worker thread file home stephan di poc venv lib site packages anyio backends asyncio py line in run sync in worker thread return await future file home stephan di poc venv lib site packages anyio backends asyncio py line in run result context run func args file home stephan di poc venv lib site packages prefect utilities importtools py line in import object module load script as module script path file home stephan di poc venv lib site packages prefect utilities importtools py line in load script as module raise scripterror user exc exc path path from exc prefect exceptions scripterror script at log flow py encountered an exception info prefect infrastructure process process nivoch nebula exited cleanly versions on windows version api version python version git commit built tue sep pm os arch profile eta cloud server type cloud on ubuntu version api version python version git commit built tue sep pm os arch linux profile cloud server type cloud additional context i was only able to test using the gcs block i tested on windows and ubuntu via wsl i m running everything inside the same poetry venv
| 0
|
350,994
| 31,932,887,205
|
IssuesEvent
|
2023-09-19 08:37:08
|
bazelbuild/intellij
|
https://api.github.com/repos/bazelbuild/intellij
|
closed
|
"Illegal Character" trying to debug go_test, and passing tests don't show up when run.
|
type: bug P3 lang: go product: GoLand topic: testing
|
The problem occurs when run a test in Goland from the "Run/Debug configurations" using bazel test.
If I just run, then the tests run and I can see the results on the console, but the test navigator says "No tests were found" (if they all pass)

Note that I have "show passing tests" selected, so the passing tests are expected to show up.
Interestingly, if I change one of the tests to fail, I get...

so the failing test shows up and the passing ones do not.
But the original issue I had is that I'm not able to debug. If I click the debug button I get:

I tried upgrading the bazel plugin (and Goland) since I saw that fixing debugging with `go_test` was on the list of recent fixes, but the issue persists after all of that, plus doing `bazel clean` and even nuking my output-user-root directory to force everything to rebuild.
Also note that the illegal character remains at index 116 even if I change the test to have everything commented out, leaving only one empty test method.
Throughout all of this, `bazel build` and `bazel test` continue to work from the terminal as expected.
Versions:
Goland: 2022.1
Bazelisk: 1.11.0
Bazel: 5.1.1
Bazel Plugin: 2022.04.28.0.0-api-version-221
|
1.0
|
"Illegal Character" trying to debug go_test, and passing tests don't show up when run. - The problem occurs when run a test in Goland from the "Run/Debug configurations" using bazel test.
If I just run, then the tests run and I can see the results on the console, but the test navigator says "No tests were found" (if they all pass)

Note that I have "show passing tests" selected, so the passing tests are expected to show up.
Interestingly, if I change one of the tests to fail, I get...

so the failing test shows up and the passing ones do not.
But the original issue I had is that I'm not able to debug. If I click the debug button I get:

I tried upgrading the bazel plugin (and Goland) since I saw that fixing debugging with `go_test` was on the list of recent fixes, but the issue persists after all of that, plus doing `bazel clean` and even nuking my output-user-root directory to force everything to rebuild.
Also note that the illegal character remains at index 116 even if I change the test to have everything commented out, leaving only one empty test method.
Throughout all of this, `bazel build` and `bazel test` continue to work from the terminal as expected.
Versions:
Goland: 2022.1
Bazelisk: 1.11.0
Bazel: 5.1.1
Bazel Plugin: 2022.04.28.0.0-api-version-221
|
test
|
illegal character trying to debug go test and passing tests don t show up when run the problem occurs when run a test in goland from the run debug configurations using bazel test if i just run then the tests run and i can see the results on the console but the test navigator says no tests were found if they all pass note that i have show passing tests selected so the passing tests are expected to show up interestingly if i change one of the tests to fail i get so the failing test shows up and the passing ones do not but the original issue i had is that i m not able to debug if i click the debug button i get i tried upgrading the bazel plugin and goland since i saw that fixing debugging with go test was on the list of recent fixes but the issue persists after all of that plus doing bazel clean and even nuking my output user root directory to force everything to rebuild also note that the illegal character remains at index even if i change the test to have everything commented out leaving only one empty test method throughout all of this bazel build and bazel test continue to work from the terminal as expected versions goland bazelisk bazel bazel plugin api version
| 1
|
326,962
| 24,110,497,290
|
IssuesEvent
|
2022-09-20 10:56:05
|
CryptoBlades/cryptoblades
|
https://api.github.com/repos/CryptoBlades/cryptoblades
|
closed
|
Special (pickable) quests
|
documentation type-frontend type-solidity
|
Special quests are event-like offers we present to players when we want to create traffic for certain NFTs and resources.
This extension involves a new pool of quests that can be freely picked to be done by any character regardless of reputation, as well as do quests that are tied only to wallets, without character involvement.
The idea for wallet quests is to allow low barrier of entry options for users migrating from other projects (but possibly other use cases too). All wallet quests can be progressed simultaneously on a single wallet, it doesn't have to be one at a time.
For quests hand picked for characters, it's the usual routine but no reputation requirement, and the user can choose the quest whenever the character is not on a quest already. This serves a similar purpose to wallet quests but we can tie it to character features (such as stamina, xp and rewarding reputation)
Development is already underway (by me), character-pickable quests are done on the "pickableQuests" branch, wallet quests remain to be implemented still.
Expected completion time (before testnet) is a few productive days away (but definitely before June 7th, ideally deploying to testnet by June 3rd to be tested on the weekend.
This will need additional frontend work soon by someone with that skill set, to display the special quests on offer and hook in the buttons to pick them.
|
1.0
|
Special (pickable) quests - Special quests are event-like offers we present to players when we want to create traffic for certain NFTs and resources.
This extension involves a new pool of quests that can be freely picked to be done by any character regardless of reputation, as well as do quests that are tied only to wallets, without character involvement.
The idea for wallet quests is to allow low barrier of entry options for users migrating from other projects (but possibly other use cases too). All wallet quests can be progressed simultaneously on a single wallet, it doesn't have to be one at a time.
For quests hand picked for characters, it's the usual routine but no reputation requirement, and the user can choose the quest whenever the character is not on a quest already. This serves a similar purpose to wallet quests but we can tie it to character features (such as stamina, xp and rewarding reputation)
Development is already underway (by me), character-pickable quests are done on the "pickableQuests" branch, wallet quests remain to be implemented still.
Expected completion time (before testnet) is a few productive days away (but definitely before June 7th, ideally deploying to testnet by June 3rd to be tested on the weekend.
This will need additional frontend work soon by someone with that skill set, to display the special quests on offer and hook in the buttons to pick them.
|
non_test
|
special pickable quests special quests are event like offers we present to players when we want to create traffic for certain nfts and resources this extension involves a new pool of quests that can be freely picked to be done by any character regardless of reputation as well as do quests that are tied only to wallets without character involvement the idea for wallet quests is to allow low barrier of entry options for users migrating from other projects but possibly other use cases too all wallet quests can be progressed simultaneously on a single wallet it doesn t have to be one at a time for quests hand picked for characters it s the usual routine but no reputation requirement and the user can choose the quest whenever the character is not on a quest already this serves a similar purpose to wallet quests but we can tie it to character features such as stamina xp and rewarding reputation development is already underway by me character pickable quests are done on the pickablequests branch wallet quests remain to be implemented still expected completion time before testnet is a few productive days away but definitely before june ideally deploying to testnet by june to be tested on the weekend this will need additional frontend work soon by someone with that skill set to display the special quests on offer and hook in the buttons to pick them
| 0
|
14,033
| 3,372,449,058
|
IssuesEvent
|
2015-11-23 23:38:33
|
18F/doi-extractives-data
|
https://api.github.com/repos/18F/doi-extractives-data
|
closed
|
BUG: Header nav item not active for all subpages
|
type:bug workflow:testing
|
The header nav items are 'active' for direct subpages, but anything deeper (e.g. `/how-it-works/goverance`) does not activate the header nav item.
|
1.0
|
BUG: Header nav item not active for all subpages - The header nav items are 'active' for direct subpages, but anything deeper (e.g. `/how-it-works/goverance`) does not activate the header nav item.
|
test
|
bug header nav item not active for all subpages the header nav items are active for direct subpages but anything deeper e g how it works goverance does not activate the header nav item
| 1
|
53,321
| 6,310,091,615
|
IssuesEvent
|
2017-07-23 06:36:39
|
MajkiIT/polish-ads-filter
|
https://api.github.com/repos/MajkiIT/polish-ads-filter
|
closed
|
spidersweb.pl
|
reguły gotowe/testowanie reklama
|
szare boxy resztki
http://www.spidersweb.pl/2017/07/komputer-do-cs-go.html
http://www.spidersweb.pl/2015/06/apple-music-ceny-w-polsce.html

od czasu do czasu sa i na main page

|
1.0
|
spidersweb.pl - szare boxy resztki
http://www.spidersweb.pl/2017/07/komputer-do-cs-go.html
http://www.spidersweb.pl/2015/06/apple-music-ceny-w-polsce.html

od czasu do czasu sa i na main page

|
test
|
spidersweb pl szare boxy resztki od czasu do czasu sa i na main page
| 1
|
94,537
| 3,927,185,876
|
IssuesEvent
|
2016-04-23 11:47:07
|
MinetestForFun/server-minetestforfun-skyblock
|
https://api.github.com/repos/MinetestForFun/server-minetestforfun-skyblock
|
opened
|
Add PVP features
|
Priority: High
|
This days, many players ask us for a PVP system, we will include it but i think, in a way it will be acceptable for everyone, let me explain our roadmap :
1) Add swords, with linear increase damage with the ore you use to craft them.
We will add them with modified textures from our "MFF_subgame", this default textures are 32px, well done and beautiful
2) Add the "PVP_button" mod
Tweak/recode the mod to be compatible with "u_skins" inventory.
Also we will add a button in your inventory to activate or deactivate PVP.
We need to set a countdown for this activate/deactivate feature, well let's try 5 minutes for the beginning.
3) At last, activate damage on the server
|
1.0
|
Add PVP features - This days, many players ask us for a PVP system, we will include it but i think, in a way it will be acceptable for everyone, let me explain our roadmap :
1) Add swords, with linear increase damage with the ore you use to craft them.
We will add them with modified textures from our "MFF_subgame", this default textures are 32px, well done and beautiful
2) Add the "PVP_button" mod
Tweak/recode the mod to be compatible with "u_skins" inventory.
Also we will add a button in your inventory to activate or deactivate PVP.
We need to set a countdown for this activate/deactivate feature, well let's try 5 minutes for the beginning.
3) At last, activate damage on the server
|
non_test
|
add pvp features this days many players ask us for a pvp system we will include it but i think in a way it will be acceptable for everyone let me explain our roadmap add swords with linear increase damage with the ore you use to craft them we will add them with modified textures from our mff subgame this default textures are well done and beautiful add the pvp button mod tweak recode the mod to be compatible with u skins inventory also we will add a button in your inventory to activate or deactivate pvp we need to set a countdown for this activate deactivate feature well let s try minutes for the beginning at last activate damage on the server
| 0
|
328,967
| 28,143,202,572
|
IssuesEvent
|
2023-04-02 07:00:51
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
opened
|
Fix general.test_explicit_ivy_framework_handles
|
Sub Task Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4558171043/jobs/8040729339" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4586003022/jobs/8098525241" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4575494898/jobs/8078405705" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4586062504/jobs/8098629047" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_general.py::test_explicit_ivy_framework_handles[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-04-02T00:50:01.0748115Z E AttributeError: module 'ivy' has no attribute 'get_backend'
</details>
<details>
<summary>Not found</summary>
Not found
</details>
|
1.0
|
Fix general.test_explicit_ivy_framework_handles - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4558171043/jobs/8040729339" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4586003022/jobs/8098525241" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4575494898/jobs/8078405705" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4586062504/jobs/8098629047" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_general.py::test_explicit_ivy_framework_handles[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-04-02T00:50:01.0748115Z E AttributeError: module 'ivy' has no attribute 'get_backend'
</details>
<details>
<summary>Not found</summary>
Not found
</details>
|
test
|
fix general test explicit ivy framework handles tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test functional test core test general py test explicit ivy framework handles e attributeerror module ivy has no attribute get backend not found not found
| 1
|
141,209
| 11,403,846,043
|
IssuesEvent
|
2020-01-31 08:23:36
|
ifb-git/ProteoRE
|
https://api.github.com/repos/ifb-git/ProteoRE
|
closed
|
New tool: Add_Protein_Features (Rattus) - uniprot
|
ToTest enhancement
|
Same specifications as "Add protein features (Mouse) [UniProt]"
With Taxonomy ID: 10116 for Rattus norvegicus
|
1.0
|
New tool: Add_Protein_Features (Rattus) - uniprot - Same specifications as "Add protein features (Mouse) [UniProt]"
With Taxonomy ID: 10116 for Rattus norvegicus
|
test
|
new tool add protein features rattus uniprot same specifications as add protein features mouse with taxonomy id for rattus norvegicus
| 1
|
283,531
| 8,719,830,902
|
IssuesEvent
|
2018-12-08 05:02:56
|
aowen87/BAR
|
https://api.github.com/repos/aowen87/BAR
|
closed
|
INSTALL_NOTES mentions installing from the dmg file and the tar.gz file, the tar.gz file is not on web page
|
bug likelihood medium priority reviewed severity low
|
The INSTALL_NOTES mentions that there are two ways to install on the Mac. One is with the dmg file and the other is with the tar.gz file. THe problem is that the tar.gz file is not on the download page (it is in the releases directory). We should either change the INSTALL_NOTES or add links for the tar.gz files. I think adding the tar.gz file to the download page is probably the way to go. I also noticed that the INSTALL_NOTES still mention 32 bit linux distributions, which we no longer provide.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1281
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: INSTALL_NOTES mentions installing from the dmg file and the tar.gz file, the tar.gz file is not on web page
Assigned to: Eric Brugger
Category:
Target version: 2.6.1
Author: Eric Brugger
Start: 12/18/2012
Due date:
% Done: 100
Estimated time: 1.0
Created: 12/18/2012 12:10 pm
Updated: 12/19/2012 05:15 pm
Likelihood: 3 - Occasional
Severity: 2 - Minor Irritation
Found in version: 2.6.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
The INSTALL_NOTES mentions that there are two ways to install on the Mac. One is with the dmg file and the other is with the tar.gz file. THe problem is that the tar.gz file is not on the download page (it is in the releases directory). We should either change the INSTALL_NOTES or add links for the tar.gz files. I think adding the tar.gz file to the download page is probably the way to go. I also noticed that the INSTALL_NOTES still mention 32 bit linux distributions, which we no longer provide.
Comments:
After discussing this with Brad, Cyrus and Kathleen, we decided that we should remove it from the install notes rather than put the "tar.gz" file on the download page.
I committed revisions 19872 and 19874 to the 2.6 RC and trunk with thefollowing change:1) I updated the INSTALL_NOTES so that only describes how to install the application bundle and no longer mentions installing it using the visit-install script. This resolves #1281.M INSTALL_NOTES
|
1.0
|
INSTALL_NOTES mentions installing from the dmg file and the tar.gz file, the tar.gz file is not on web page - The INSTALL_NOTES mentions that there are two ways to install on the Mac. One is with the dmg file and the other is with the tar.gz file. THe problem is that the tar.gz file is not on the download page (it is in the releases directory). We should either change the INSTALL_NOTES or add links for the tar.gz files. I think adding the tar.gz file to the download page is probably the way to go. I also noticed that the INSTALL_NOTES still mention 32 bit linux distributions, which we no longer provide.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1281
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: INSTALL_NOTES mentions installing from the dmg file and the tar.gz file, the tar.gz file is not on web page
Assigned to: Eric Brugger
Category:
Target version: 2.6.1
Author: Eric Brugger
Start: 12/18/2012
Due date:
% Done: 100
Estimated time: 1.0
Created: 12/18/2012 12:10 pm
Updated: 12/19/2012 05:15 pm
Likelihood: 3 - Occasional
Severity: 2 - Minor Irritation
Found in version: 2.6.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
The INSTALL_NOTES mentions that there are two ways to install on the Mac. One is with the dmg file and the other is with the tar.gz file. THe problem is that the tar.gz file is not on the download page (it is in the releases directory). We should either change the INSTALL_NOTES or add links for the tar.gz files. I think adding the tar.gz file to the download page is probably the way to go. I also noticed that the INSTALL_NOTES still mention 32 bit linux distributions, which we no longer provide.
Comments:
After discussing this with Brad, Cyrus and Kathleen, we decided that we should remove it from the install notes rather than put the "tar.gz" file on the download page.
I committed revisions 19872 and 19874 to the 2.6 RC and trunk with thefollowing change:1) I updated the INSTALL_NOTES so that only describes how to install the application bundle and no longer mentions installing it using the visit-install script. This resolves #1281.M INSTALL_NOTES
|
non_test
|
install notes mentions installing from the dmg file and the tar gz file the tar gz file is not on web page the install notes mentions that there are two ways to install on the mac one is with the dmg file and the other is with the tar gz file the problem is that the tar gz file is not on the download page it is in the releases directory we should either change the install notes or add links for the tar gz files i think adding the tar gz file to the download page is probably the way to go i also noticed that the install notes still mention bit linux distributions which we no longer provide redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject install notes mentions installing from the dmg file and the tar gz file the tar gz file is not on web page assigned to eric brugger category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity minor irritation found in version impact expected use os all support group any description the install notes mentions that there are two ways to install on the mac one is with the dmg file and the other is with the tar gz file the problem is that the tar gz file is not on the download page it is in the releases directory we should either change the install notes or add links for the tar gz files i think adding the tar gz file to the download page is probably the way to go i also noticed that the install notes still mention bit linux distributions which we no longer provide comments after discussing this with brad cyrus and kathleen we decided that we should remove it from the install notes rather than put the tar gz file on the download page i committed revisions and to the rc and trunk with thefollowing change i updated the install notes so that only describes how to install the application bundle and no longer mentions installing it using the visit install script this resolves m install notes
| 0
|
140,035
| 11,301,413,854
|
IssuesEvent
|
2020-01-17 15:32:49
|
stevenschader/kabanero-foundation
|
https://api.github.com/repos/stevenschader/kabanero-foundation
|
closed
|
SVT: TER: Kabanero Automation test execution master branch - Setup: svtcrc-91747-1
|
SVT Kabanero Test Execution Test Execution Record
|
Original logs: kabanero-svt01.fyre.ibm.com:/home/nest/kabanero-crc-logs/kabanero-auto-crc.sh.2019-10-24-05:45:33.setup.log
PLAY [Create a one host Fyre stack] ***********************************************************************************************
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 05:45:37 -0700 (0:00:00.207) 0:00:00.207 ******
ok: [localhost]
TASK [check status of the named cluster] ******************************************************************************************
Thursday 24 October 2019 05:45:37 -0700 (0:00:00.047) 0:00:00.255 ******
changed: [localhost]
TASK [debug] **********************************************************************************************************************
Thursday 24 October 2019 05:45:38 -0700 (0:00:01.565) 0:00:01.821 ******
skipping: [localhost]
TASK [get the public ssh id] ******************************************************************************************************
Thursday 24 October 2019 05:45:38 -0700 (0:00:00.057) 0:00:01.878 ******
changed: [localhost -> localhost]
TASK [create input json file] *****************************************************************************************************
Thursday 24 October 2019 05:45:39 -0700 (0:00:00.306) 0:00:02.184 ******
changed: [localhost]
TASK [Create Fyre stack] **********************************************************************************************************
Thursday 24 October 2019 05:45:40 -0700 (0:00:01.016) 0:00:03.201 ******
changed: [localhost]
TASK [remove temp json] ***********************************************************************************************************
Thursday 24 October 2019 05:45:41 -0700 (0:00:01.866) 0:00:05.068 ******
changed: [localhost -> localhost]
TASK [check Fyre request status] **************************************************************************************************
Thursday 24 October 2019 05:45:42 -0700 (0:00:00.472) 0:00:05.540 ******
changed: [localhost]
TASK [debug] **********************************************************************************************************************
Thursday 24 October 2019 05:45:43 -0700 (0:00:00.905) 0:00:06.445 ******
ok: [localhost] => {
"fyreStatus.stdout": {
"request": [
{
"cluster_name": "",
"created": "2019-10-24 05:45:41",
"displayname": "steven r. schader",
"error_details": "0",
"id": "2420698",
"status": "building",
"type": "build",
"updated": "2019-10-24 05:45:42",
"username": "schader"
}
]
}
}
TASK [pause] **********************************************************************************************************************
Thursday 24 October 2019 05:45:43 -0700 (0:00:00.041) 0:00:06.487 ******
Pausing for 120 seconds
(ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort)
ok: [localhost]
TASK [get Fyre request status] ****************************************************************************************************
Thursday 24 October 2019 05:47:43 -0700 (0:02:00.047) 0:02:06.535 ******
changed: [localhost]
TASK [check Fyre request status] **************************************************************************************************
Thursday 24 October 2019 05:47:44 -0700 (0:00:01.028) 0:02:07.564 ******
skipping: [localhost]
TASK [check the host for a active ssh] ********************************************************************************************
Thursday 24 October 2019 05:47:44 -0700 (0:00:00.041) 0:02:07.605 ******
changed: [localhost -> localhost]
TASK [remove new host from localhost known_hosts] *********************************************************************************
Thursday 24 October 2019 05:47:45 -0700 (0:00:00.582) 0:02:08.188 ******
changed: [localhost -> localhost]
TASK [add host to known_hosts] ****************************************************************************************************
Thursday 24 October 2019 05:47:45 -0700 (0:00:00.315) 0:02:08.503 ******
changed: [localhost -> localhost]
TASK [create host inventory file] *************************************************************************************************
Thursday 24 October 2019 05:47:45 -0700 (0:00:00.378) 0:02:08.881 ******
changed: [localhost -> localhost]
TASK [create plain file with host] ************************************************************************************************
Thursday 24 October 2019 05:47:46 -0700 (0:00:00.595) 0:02:09.476 ******
changed: [localhost -> localhost]
TASK [create plain file with cluster name] ****************************************************************************************
Thursday 24 October 2019 05:47:46 -0700 (0:00:00.563) 0:02:10.040 ******
changed: [localhost -> localhost]
TASK [remove linked hosts.crc] ****************************************************************************************************
Thursday 24 October 2019 05:47:47 -0700 (0:00:00.574) 0:02:10.615 ******
changed: [localhost -> localhost]
TASK [link new host inventory file to hosts.crc] **********************************************************************************
Thursday 24 October 2019 05:47:47 -0700 (0:00:00.285) 0:02:10.901 ******
changed: [localhost -> localhost]
PLAY RECAP ************************************************************************************************************************
localhost : ok=18 changed=15 unreachable=0 failed=0 skipped=2 rescued=0 ignored=0
Thursday 24 October 2019 05:47:48 -0700 (0:00:00.288) 0:02:11.189 ******
===============================================================================
pause -------------------------------------------------------------------------------------------------------------------- 120.05s
Create Fyre stack ---------------------------------------------------------------------------------------------------------- 1.87s
check status of the named cluster ------------------------------------------------------------------------------------------ 1.57s
get Fyre request status ---------------------------------------------------------------------------------------------------- 1.03s
create input json file ----------------------------------------------------------------------------------------------------- 1.02s
check Fyre request status -------------------------------------------------------------------------------------------------- 0.91s
create host inventory file ------------------------------------------------------------------------------------------------- 0.60s
check the host for a active ssh -------------------------------------------------------------------------------------------- 0.58s
create plain file with cluster name ---------------------------------------------------------------------------------------- 0.57s
create plain file with host ------------------------------------------------------------------------------------------------ 0.56s
remove temp json ----------------------------------------------------------------------------------------------------------- 0.47s
add host to known_hosts ---------------------------------------------------------------------------------------------------- 0.38s
remove new host from localhost known_hosts --------------------------------------------------------------------------------- 0.32s
get the public ssh id ------------------------------------------------------------------------------------------------------ 0.31s
link new host inventory file to hosts.crc ---------------------------------------------------------------------------------- 0.29s
remove linked hosts.crc ---------------------------------------------------------------------------------------------------- 0.29s
debug ---------------------------------------------------------------------------------------------------------------------- 0.06s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s
debug ---------------------------------------------------------------------------------------------------------------------- 0.04s
check Fyre request status -------------------------------------------------------------------------------------------------- 0.04s
fyrecrcvm complete. runtime 2 minutes
PLAY [check RHEL 7 or greater x64 only] *******************************************************************************************
TASK [Gathering Facts] ************************************************************************************************************
Thursday 24 October 2019 05:47:50 -0700 (0:00:00.170) 0:00:00.170 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [Target host is RHEL7 ( or greater ) x86_64] *********************************************************************************
Thursday 24 October 2019 05:47:52 -0700 (0:00:01.803) 0:00:01.974 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
PLAY RECAP ************************************************************************************************************************
svtcrc-91747-1.fyre.ibm.com : ok=2 changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0
Thursday 24 October 2019 05:47:52 -0700 (0:00:00.039) 0:00:02.014 ******
===============================================================================
Gathering Facts ------------------------------------------------------------------------------------------------------------ 1.80s
Target host is RHEL7 ( or greater ) x86_64 --------------------------------------------------------------------------------- 0.04s
PLAY [Create a User] **************************************************************************************************************
TASK [Generate User Password file] ************************************************************************************************
Thursday 24 October 2019 05:47:55 -0700 (0:00:00.183) 0:00:00.183 ******
ok: [svtcrc-91747-1.fyre.ibm.com -> localhost]
TASK [Ensure User is Present] *****************************************************************************************************
Thursday 24 October 2019 05:47:56 -0700 (0:00:00.495) 0:00:00.678 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [authorized_key] *************************************************************************************************************
Thursday 24 October 2019 05:47:57 -0700 (0:00:01.163) 0:00:01.842 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [kevin sudoers] **************************************************************************************************************
Thursday 24 October 2019 05:47:58 -0700 (0:00:00.727) 0:00:02.569 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [kevin hard ulimit nofile] ***************************************************************************************************
Thursday 24 October 2019 05:47:58 -0700 (0:00:00.750) 0:00:03.320 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [kevin soft ulimit nofile] ***************************************************************************************************
Thursday 24 October 2019 05:47:59 -0700 (0:00:00.434) 0:00:03.754 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [add bin to path] ************************************************************************************************************
Thursday 24 October 2019 05:47:59 -0700 (0:00:00.437) 0:00:04.192 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
PLAY RECAP ************************************************************************************************************************
svtcrc-91747-1.fyre.ibm.com : ok=7 changed=6 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0
Thursday 24 October 2019 05:48:00 -0700 (0:00:00.428) 0:00:04.620 ******
===============================================================================
Ensure User is Present ----------------------------------------------------------------------------------------------------- 1.16s
kevin sudoers -------------------------------------------------------------------------------------------------------------- 0.75s
authorized_key ------------------------------------------------------------------------------------------------------------- 0.73s
Generate User Password file ------------------------------------------------------------------------------------------------ 0.50s
kevin soft ulimit nofile --------------------------------------------------------------------------------------------------- 0.44s
kevin hard ulimit nofile --------------------------------------------------------------------------------------------------- 0.43s
add bin to path ------------------------------------------------------------------------------------------------------------ 0.43s
PLAY [switch hosts.crc ansible_user] **********************************************************************************************
TASK [remove kevin ansible_user] **************************************************************************************************
Thursday 24 October 2019 05:48:02 -0700 (0:00:00.178) 0:00:00.178 ******
ok: [svtcrc-91747-1.fyre.ibm.com -> localhost]
TASK [set kevin ansible_user] *****************************************************************************************************
Thursday 24 October 2019 05:48:03 -0700 (0:00:00.476) 0:00:00.654 ******
changed: [svtcrc-91747-1.fyre.ibm.com -> localhost]
PLAY RECAP ************************************************************************************************************************
svtcrc-91747-1.fyre.ibm.com : ok=2 changed=1 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0
Thursday 24 October 2019 05:48:03 -0700 (0:00:00.266) 0:00:00.921 ******
===============================================================================
remove kevin ansible_user -------------------------------------------------------------------------------------------------- 0.48s
set kevin ansible_user ----------------------------------------------------------------------------------------------------- 0.27s
PLAY [SVT - install OCP CodeReadyContainer binary] ********************************************************************************
TASK [Gathering Facts] ************************************************************************************************************
Thursday 24 October 2019 05:48:06 -0700 (0:00:00.171) 0:00:00.171 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [check if crc is installed] **************************************************************************************************
Thursday 24 October 2019 05:48:08 -0700 (0:00:01.862) 0:00:02.034 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [archive dir structure] ******************************************************************************************************
Thursday 24 October 2019 05:48:09 -0700 (0:00:00.653) 0:00:02.688 ******
changed: [svtcrc-91747-1.fyre.ibm.com -> localhost]
TASK [debug] **********************************************************************************************************************
Thursday 24 October 2019 05:48:17 -0700 (0:00:08.477) 0:00:11.165 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"tarlist.stdout_lines": [
"crc-linux-1.0.0-amd64/",
"crc-linux-1.0.0-amd64/LICENSE",
"crc-linux-1.0.0-amd64/doc.pdf",
"crc-linux-1.0.0-amd64/crc"
]
}
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 05:48:17 -0700 (0:00:00.055) 0:00:11.221 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [copy crc archive] ***********************************************************************************************************
Thursday 24 October 2019 05:48:17 -0700 (0:00:00.062) 0:00:11.284 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [install crc] ****************************************************************************************************************
Thursday 24 October 2019 05:50:03 -0700 (0:01:46.017) 0:01:57.301 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [home bin] *******************************************************************************************************************
Thursday 24 October 2019 05:50:25 -0700 (0:00:22.091) 0:02:19.392 ******
changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/bin)
TASK [link crc] *******************************************************************************************************************
Thursday 24 October 2019 05:50:26 -0700 (0:00:00.761) 0:02:20.154 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
PLAY RECAP ************************************************************************************************************************
svtcrc-91747-1.fyre.ibm.com : ok=9 changed=5 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0
Thursday 24 October 2019 05:50:27 -0700 (0:00:00.483) 0:02:20.637 ******
===============================================================================
copy crc archive --------------------------------------------------------------------------------------------------------- 106.02s
install crc --------------------------------------------------------------------------------------------------------------- 22.09s
archive dir structure ------------------------------------------------------------------------------------------------------ 8.48s
Gathering Facts ------------------------------------------------------------------------------------------------------------ 1.86s
home bin ------------------------------------------------------------------------------------------------------------------- 0.76s
check if crc is installed -------------------------------------------------------------------------------------------------- 0.65s
link crc ------------------------------------------------------------------------------------------------------------------- 0.48s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.06s
debug ---------------------------------------------------------------------------------------------------------------------- 0.06s
crc-install complete. runtime 2 minutes
PLAY [SVT - Kabanero on crc] ******************************************************************************************************
TASK [Gathering Facts] ************************************************************************************************************
Thursday 24 October 2019 05:50:31 -0700 (0:00:00.175) 0:00:00.175 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 05:50:33 -0700 (0:00:01.727) 0:00:01.902 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [Install NetworkManager FireFox] *********************************************************************************************
Thursday 24 October 2019 05:50:33 -0700 (0:00:00.060) 0:00:01.963 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [enable and start NetworkManager] ********************************************************************************************
Thursday 24 October 2019 05:51:47 -0700 (0:01:13.989) 0:01:15.952 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [check if crc is installed] **************************************************************************************************
Thursday 24 October 2019 05:51:48 -0700 (0:00:01.324) 0:01:17.276 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [fail] ***********************************************************************************************************************
Thursday 24 October 2019 05:51:49 -0700 (0:00:00.713) 0:01:17.990 ******
skipping: [svtcrc-91747-1.fyre.ibm.com]
TASK [prereq directories] *********************************************************************************************************
Thursday 24 October 2019 05:51:49 -0700 (0:00:00.059) 0:01:18.049 ******
changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/git)
changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/testDir)
changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/kabanero/logs)
TASK [retrieve Kabanero Foundation master zip] ************************************************************************************
Thursday 24 October 2019 05:51:51 -0700 (0:00:01.507) 0:01:19.557 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [unzip Kabanerio Foundation master zip] **************************************************************************************
Thursday 24 October 2019 05:51:53 -0700 (0:00:02.402) 0:01:21.959 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [whoami] *********************************************************************************************************************
Thursday 24 October 2019 05:51:54 -0700 (0:00:01.073) 0:01:23.032 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [whoami stdout] **************************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.689) 0:01:23.722 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"kevin"
]
}
TASK [check if .crc is created] ***************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.047) 0:01:23.770 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc stop] *******************************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.481) 0:01:24.251 ******
skipping: [svtcrc-91747-1.fyre.ibm.com]
TASK [crcstop stdout] *************************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.053) 0:01:24.305 ******
skipping: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc delete] *****************************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.058) 0:01:24.364 ******
skipping: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc delete stdout] **********************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.059) 0:01:24.423 ******
skipping: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc setup] ******************************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.050) 0:01:24.474 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc setup stdout] ***********************************************************************************************************
Thursday 24 October 2019 05:52:34 -0700 (0:00:38.733) 0:02:03.207 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"level=info msg=\"Checking if running as non-root\"",
"level=info msg=\"Caching oc binary\"",
"level=info msg=\"Setting up virtualization\"",
"level=info msg=\"Setting up KVM\"",
"level=info msg=\"Installing libvirt service and dependencies\"",
"level=info msg=\"Will use root access: install virtualization related packages\"",
"level=info msg=\"Adding user to libvirt group\"",
"level=info msg=\"Will use root access: add user to libvirt group\"",
"level=info msg=\"Enabling libvirt\"",
"level=info msg=\"Starting libvirt service\"",
"level=info msg=\"Will use root access: start libvirtd service\"",
"level=info msg=\"Checking if a supported libvirt version is installed\"",
"level=info msg=\"Installing crc-driver-libvirt\"",
"level=info msg=\"Removing older system-wide crc-driver-libvirt\"",
"level=info msg=\"Setting up libvirt 'crc' network\"",
"level=info msg=\"Starting libvirt 'crc' network\"",
"level=info msg=\"Checking if NetworkManager is installed\"",
"level=info msg=\"Checking if NetworkManager service is running\"",
"level=info msg=\"Writing Network Manager config for crc\"",
"level=info msg=\"Will use root access: write NetworkManager config in /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf\"",
"level=info msg=\"Will use root access: execute systemctl daemon-reload command\"",
"level=info msg=\"Will use root access: execute systemctl stop/start command\"",
"level=info msg=\"Writing dnsmasq config for crc\"",
"level=info msg=\"Will use root access: write dnsmasq configuration in /etc/NetworkManager/dnsmasq.d/crc.conf\"",
"level=info msg=\"Will use root access: execute systemctl daemon-reload command\"",
"level=info msg=\"Will use root access: execute systemctl stop/start command\"",
"level=info msg=\"Unpacking bundle from the CRC binary\"",
"Setup is complete, you can now run 'crc start' to start the OpenShift cluster"
]
}
TASK [crc pull secret] ************************************************************************************************************
Thursday 24 October 2019 05:52:34 -0700 (0:00:00.074) 0:02:03.281 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc start] ******************************************************************************************************************
Thursday 24 October 2019 05:52:35 -0700 (0:00:01.108) 0:02:04.390 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc start stdout] ***********************************************************************************************************
Thursday 24 October 2019 06:03:32 -0700 (0:10:56.959) 0:13:01.349 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"level=info msg=\"Checking if running as non-root\"",
"level=info msg=\"Checking if oc binary is cached\"",
"level=info msg=\"Checking if Virtualization is enabled\"",
"level=info msg=\"Checking if KVM is enabled\"",
"level=info msg=\"Checking if libvirt is installed\"",
"level=info msg=\"Checking if user is part of libvirt group\"",
"level=info msg=\"Checking if libvirt is enabled\"",
"level=info msg=\"Checking if libvirt daemon is running\"",
"level=info msg=\"Checking if a supported libvirt version is installed\"",
"level=info msg=\"Checking if crc-driver-libvirt is installed\"",
"level=info msg=\"Checking if libvirt 'crc' network is available\"",
"level=info msg=\"Checking if libvirt 'crc' network is active\"",
"level=info msg=\"Checking if NetworkManager is installed\"",
"level=info msg=\"Checking if NetworkManager service is running\"",
"level=info msg=\"Checking if /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf exists\"",
"level=info msg=\"Checking if /etc/NetworkManager/dnsmasq.d/crc.conf exists\"",
"level=info msg=\"Extracting bundle: crc_libvirt_4.2.0.crcbundle ...\"",
"level=info msg=\"Creating CodeReady Containers VM for OpenShift 4.2.0...\"",
"level=info msg=\"Verifying validity of the cluster certificates ...\"",
"level=info msg=\"Check internal and public DNS query ...\"",
"level=info msg=\"Copying kubeconfig file to instance dir ...\"",
"level=info msg=\"Adding user's pull secret and cluster ID ...\"",
"level=info msg=\"Starting OpenShift cluster ... [waiting 3m]\"",
"level=info",
"level=info msg=\"To access the cluster, first set up your environment by following 'crc oc-env' instructions\"",
"level=info msg=\"Then you can access it by running 'oc login -u developer -p developer https://api.crc.testing:6443'\"",
"level=info msg=\"To login as an admin, username is 'kubeadmin' and password is xx-password-xx\"",
"level=info",
"level=info msg=\"You can now run 'crc console' and use these credentials to access the OpenShift web console\"",
"The OpenShift cluster is running",
"level=warning msg=\"The cluster might report a degraded or error state. This is expected since several operators have been disable to lower the resource usage. For more information, please consult the documentation\""
]
}
TASK [oc in user path] ************************************************************************************************************
Thursday 24 October 2019 06:03:32 -0700 (0:00:00.066) 0:13:01.416 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [link oc kubectl] ************************************************************************************************************
Thursday 24 October 2019 06:03:33 -0700 (0:00:00.663) 0:13:02.079 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc creds] ******************************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.438) 0:13:02.518 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc creds stdout] ***********************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.498) 0:13:03.016 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"To login as a regular user, username is 'developer' and password is 'developer'.",
"To login as an admin, username is 'kubeadmin' and password is 'xx-password-xx'."
]
}
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.064) 0:13:03.081 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.052) 0:13:03.134 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.051) 0:13:03.185 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.051) 0:13:03.236 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [password masking process] ***************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.076) 0:13:03.312 ******
changed: [svtcrc-91747-1.fyre.ibm.com -> localhost]
TASK [set admin passwd] ***********************************************************************************************************
Thursday 24 October 2019 06:03:35 -0700 (0:00:00.638) 0:13:03.951 ******
changed: [svtcrc-91747-1.fyre.ibm.com -> localhost]
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 06:03:35 -0700 (0:00:00.356) 0:13:04.307 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [oc login] *******************************************************************************************************************
Thursday 24 October 2019 06:03:35 -0700 (0:00:00.064) 0:13:04.371 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [oc login stdout] ************************************************************************************************************
Thursday 24 October 2019 06:03:37 -0700 (0:00:01.414) 0:13:05.786 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"Login successful.",
"",
"You have access to 51 projects, the list has been suppressed. You can list all projects with 'oc projects'",
"",
"Using project \"default\".",
"Welcome! See 'oc help' to get started."
]
}
TASK [install kabanero-foundation] ************************************************************************************************
Thursday 24 October 2019 06:03:37 -0700 (0:00:00.062) 0:13:05.849 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [install kabanero-foundation stdout] *****************************************************************************************
Thursday 24 October 2019 06:14:36 -0700 (0:10:59.331) 0:24:05.180 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-ingress-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:default\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:prometheus\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-egressgateway-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-citadel-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-ingressgateway-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-cleanup-old-ca-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-mixer-post-install-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-mixer-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-pilot-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-sidecar-injector-service-account\"]",
"clusterrole.rbac.authorization.k8s.io/cluster-admin added: \"istio-galley-service-account\"",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:cluster-local-gateway-service-account\"]",
"customresourcedefinition.apiextensions.k8s.io/virtualservices.networking.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/destinationrules.networking.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/serviceentries.networking.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/gateways.networking.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/envoyfilters.networking.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/clusterrbacconfigs.rbac.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/policies.authentication.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/meshpolicies.authentication.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/httpapispecbindings.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/httpapispecs.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/quotaspecbindings.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/quotaspecs.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/rules.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/attributemanifests.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/bypasses.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/circonuses.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/deniers.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/fluentds.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/kubernetesenvs.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/listcheckers.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/memquotas.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/noops.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/opas.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/prometheuses.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/rbacs.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/redisquotas.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/signalfxs.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/solarwindses.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/stackdrivers.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/statsds.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/stdios.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/apikeys.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/authorizations.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/checknothings.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/kuberneteses.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/listentries.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/logentries.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/edges.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/metrics.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/quotas.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/reportnothings.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/tracespans.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/rbacconfigs.rbac.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/serviceroles.rbac.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/servicerolebindings.rbac.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/adapters.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/instances.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/templates.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/handlers.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/cloudwatches.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/dogstatsds.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/sidecars.networking.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/zipkins.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/clusterissuers.certmanager.k8s.io created",
"customresourcedefinition.apiextensions.k8s.io/issuers.certmanager.k8s.io created",
"customresourcedefinition.apiextensions.k8s.io/certificates.certmanager.k8s.io created",
"customresourcedefinition.apiextensions.k8s.io/orders.certmanager.k8s.io created",
"customresourcedefinition.apiextensions.k8s.io/challenges.certmanager.k8s.io created",
"namespace/istio-system created",
"customresourcedefinition.apiextensions.k8s.io/virtualservices.networking.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/destinationrules.networking.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/serviceentries.networking.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/gateways.networking.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/envoyfilters.networking.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/clusterrbacconfigs.rbac.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/policies.authentication.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/meshpolicies.authentication.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/httpapispecbindings.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/httpapispecs.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/quotaspecbindings.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/quotaspecs.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/rules.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/attributemanifests.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/bypasses.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/circonuses.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/deniers.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/fluentds.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/kubernetesenvs.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/listcheckers.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/memquotas.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/noops.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/opas.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/prometheuses.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/rbacs.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/redisquotas.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/signalfxs.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/solarwindses.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/stackdrivers.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/statsds.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/stdios.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/apikeys.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/authorizations.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/checknothings.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/kuberneteses.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/listentries.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/logentries.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/edges.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/metrics.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/quotas.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/reportnothings.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/tracespans.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/rbacconfigs.rbac.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/serviceroles.rbac.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/servicerolebindings.rbac.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/adapters.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/instances.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/templates.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/handlers.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/cloudwatches.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/dogstatsds.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/sidecars.networking.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/zipkins.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/clusterissuers.certmanager.k8s.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/issuers.certmanager.k8s.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/orders.certmanager.k8s.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/challenges.certmanager.k8s.io unchanged",
"secret/kiali created",
"configmap/istio-galley-configuration created",
"configmap/istio-grafana-custom-resources created",
"configmap/istio-grafana-configuration-dashboards-galley-dashboard created",
"configmap/istio-grafana-configuration-dashboards-istio-mesh-dashboard created",
"configmap/istio-grafana-configuration-dashboards-istio-performance-dashboard created",
"configmap/istio-grafana-configuration-dashboards-istio-service-dashboard created",
"configmap/istio-grafana-configuration-dashboards-istio-workload-dashboard created",
"configmap/istio-grafana-configuration-dashboards-mixer-dashboard created",
"configmap/istio-grafana-configuration-dashboards-pilot-dashboard created",
"configmap/istio-grafana created",
"configmap/kiali created",
"configmap/prometheus created",
"configmap/istio-security-custom-resources created",
"configmap/istio created",
"configmap/istio-sidecar-injector created",
"serviceaccount/istio-galley-service-account created",
"serviceaccount/istio-egressgateway-service-account created",
"serviceaccount/istio-ingressgateway-service-account created",
"serviceaccount/istio-grafana-post-install-account created",
"clusterrole.rbac.authorization.k8s.io/istio-grafana-post-install-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-grafana-post-install-role-binding-istio-system created",
"job.batch/istio-grafana-post-install-1.1.7 created",
"serviceaccount/kiali-service-account created",
"serviceaccount/istio-mixer-service-account created",
"serviceaccount/istio-pilot-service-account created",
"serviceaccount/prometheus created",
"serviceaccount/istio-cleanup-secrets-service-account created",
"clusterrole.rbac.authorization.k8s.io/istio-cleanup-secrets-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-cleanup-secrets-istio-system created",
"job.batch/istio-cleanup-secrets-1.1.7 created",
"serviceaccount/istio-security-post-install-account created",
"clusterrole.rbac.authorization.k8s.io/istio-security-post-install-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-security-post-install-role-binding-istio-system created",
"job.batch/istio-security-post-install-1.1.7 created",
"serviceaccount/istio-citadel-service-account created",
"serviceaccount/istio-sidecar-injector-service-account created",
"serviceaccount/istio-multi created",
"clusterrole.rbac.authorization.k8s.io/istio-galley-istio-system created",
"clusterrole.rbac.authorization.k8s.io/istio-egressgateway-istio-system created",
"clusterrole.rbac.authorization.k8s.io/istio-ingressgateway-istio-system created",
"clusterrole.rbac.authorization.k8s.io/kiali created",
"clusterrole.rbac.authorization.k8s.io/kiali-viewer created",
"clusterrole.rbac.authorization.k8s.io/istio-mixer-istio-system created",
"clusterrole.rbac.authorization.k8s.io/istio-pilot-istio-system created",
"clusterrole.rbac.authorization.k8s.io/prometheus-istio-system created",
"clusterrole.rbac.authorization.k8s.io/istio-citadel-istio-system created",
"clusterrole.rbac.authorization.k8s.io/istio-sidecar-injector-istio-system created",
"clusterrole.rbac.authorization.k8s.io/istio-reader created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-galley-admin-role-binding-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-egressgateway-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-ingressgateway-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-kiali-admin-role-binding-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-mixer-admin-role-binding-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-pilot-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/prometheus-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-citadel-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-sidecar-injector-admin-role-binding-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-multi created",
"role.rbac.authorization.k8s.io/istio-ingressgateway-sds created",
"rolebinding.rbac.authorization.k8s.io/istio-ingressgateway-sds created",
"service/istio-galley created",
"service/istio-egressgateway created",
"service/istio-ingressgateway created",
"service/grafana created",
"service/kiali created",
"service/istio-policy created",
"service/istio-telemetry created",
"service/istio-pilot created",
"service/prometheus created",
"service/istio-citadel created",
"service/istio-sidecar-injector created",
"deployment.extensions/istio-galley created",
"deployment.extensions/istio-egressgateway created",
"deployment.extensions/istio-ingressgateway created",
"deployment.extensions/grafana created",
"deployment.extensions/kiali created",
"deployment.extensions/istio-policy created",
"deployment.extensions/istio-telemetry created",
"deployment.extensions/istio-pilot created",
"deployment.extensions/prometheus created",
"deployment.extensions/istio-citadel created",
"deployment.extensions/istio-sidecar-injector created",
"deployment.extensions/istio-tracing created",
"horizontalpodautoscaler.autoscaling/istio-egressgateway created",
"horizontalpodautoscaler.autoscaling/istio-ingressgateway created",
"horizontalpodautoscaler.autoscaling/istio-policy created",
"horizontalpodautoscaler.autoscaling/istio-telemetry created",
"horizontalpodautoscaler.autoscaling/istio-pilot created",
"service/jaeger-query created",
"service/jaeger-collector created",
"service/jaeger-agent created",
"service/zipkin created",
"service/tracing created",
"mutatingwebhookconfiguration.admissionregistration.k8s.io/istio-sidecar-injector created",
"poddisruptionbudget.policy/istio-galley created",
"poddisruptionbudget.policy/istio-egressgateway created",
"poddisruptionbudget.policy/istio-ingressgateway created",
"poddisruptionbudget.policy/istio-policy created",
"poddisruptionbudget.policy/istio-telemetry created",
"poddisruptionbudget.policy/istio-pilot created",
"attributemanifest.config.istio.io/istioproxy created",
"attributemanifest.config.istio.io/kubernetes created",
"handler.config.istio.io/stdio created",
"logentry.config.istio.io/accesslog created",
"logentry.config.istio.io/tcpaccesslog created",
"rule.config.istio.io/stdio created",
"rule.config.istio.io/stdiotcp created",
"metric.config.istio.io/requestcount created",
"metric.config.istio.io/requestduration created",
"metric.config.istio.io/requestsize created",
"metric.config.istio.io/responsesize created",
"metric.config.istio.io/tcpbytesent created",
"metric.config.istio.io/tcpbytereceived created",
"metric.config.istio.io/tcpconnectionsopened created",
"metric.config.istio.io/tcpconnectionsclosed created",
"handler.config.istio.io/prometheus created",
"rule.config.istio.io/promhttp created",
"rule.config.istio.io/promtcp created",
"rule.config.istio.io/promtcpconnectionopen created",
"rule.config.istio.io/promtcpconnectionclosed created",
"handler.config.istio.io/kubernetesenv created",
"rule.config.istio.io/kubeattrgenrulerule created",
"rule.config.istio.io/tcpkubeattrgenrulerule created",
"kubernetes.config.istio.io/attributes created",
"destinationrule.networking.istio.io/istio-policy created",
"destinationrule.networking.istio.io/istio-telemetry created",
"Now using project \"kabanero\" on server \"https://api.crc.testing:6443\".",
"",
"You can add applications to this project with the 'new-app' command. For example, try:",
"",
" oc new-app django-psql-example",
"",
"to build a new example application in Python. Or use kubectl to deploy a simple Kubernetes application:",
"",
" kubectl create deployment hello-node --image=gcr.io/hello-minikube-zero-install/hello-node",
"",
"namespace/kabanero configured",
"deployment.apps/knative-eventing-operator created",
"role.rbac.authorization.k8s.io/knative-eventing-operator created",
"clusterrole.rbac.authorization.k8s.io/knative-eventing-operator created",
"rolebinding.rbac.authorization.k8s.io/knative-eventing-operator created",
"clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-operator created",
"serviceaccount/knative-eventing-operator created",
"customresourcedefinition.apiextensions.k8s.io/knativeeventings.eventing.knative.dev created",
"deployment.apps/knative-serving-operator created",
"role.rbac.authorization.k8s.io/knative-serving-operator created",
"clusterrole.rbac.authorization.k8s.io/knative-serving-operator created",
"rolebinding.rbac.authorization.k8s.io/knative-serving-operator created",
"clusterrolebinding.rbac.authorization.k8s.io/knative-serving-operator created",
"serviceaccount/knative-serving-operator created",
"customresourcedefinition.apiextensions.k8s.io/knativeservings.serving.knative.dev created",
"namespace/knative-sources created",
"serviceaccount/controller-manager created",
"clusterrole.rbac.authorization.k8s.io/eventing-sources-controller created",
"clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-controller created",
"clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-addressable-resolver created",
"customresourcedefinition.apiextensions.k8s.io/githubsources.sources.eventing.knative.dev created",
"service/controller created",
"statefulset.apps/controller-manager created",
"deployment.apps/openshift-pipelines-operator created",
"clusterrole.rbac.authorization.k8s.io/openshift-pipelines-operator created",
"clusterrolebinding.rbac.authorization.k8s.io/openshift-pipelines-operator created",
"serviceaccount/openshift-pipelines-operator created",
"namespace/openshift-pipelines created",
"customresourcedefinition.apiextensions.k8s.io/config.operator.tekton.dev created",
"clusterrole.rbac.authorization.k8s.io/knative-serving-operator configured",
"clusterrole.rbac.authorization.k8s.io/knative-eventing-operator configured",
"customresourcedefinition.apiextensions.k8s.io/kabaneros.kabanero.io created",
"customresourcedefinition.apiextensions.k8s.io/collections.kabanero.io created",
"serviceaccount/kabanero-operator created",
"deployment.apps/kabanero-operator created",
"clusterrole.rbac.authorization.k8s.io/kabanero-operator created",
"clusterrolebinding.rbac.authorization.k8s.io/kabanero-operator created",
"namespace/kabanero unchanged",
"deployment.apps/knative-eventing-operator configured",
"role.rbac.authorization.k8s.io/knative-eventing-operator configured",
"clusterrole.rbac.authorization.k8s.io/knative-eventing-operator configured",
"rolebinding.rbac.authorization.k8s.io/knative-eventing-operator unchanged",
"clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-operator unchanged",
"serviceaccount/knative-eventing-operator unchanged",
"customresourcedefinition.apiextensions.k8s.io/knativeeventings.eventing.knative.dev unchanged",
"deployment.apps/knative-serving-operator configured",
"role.rbac.authorization.k8s.io/knative-serving-operator configured",
"clusterrole.rbac.authorization.k8s.io/knative-serving-operator configured",
"rolebinding.rbac.authorization.k8s.io/knative-serving-operator unchanged",
"clusterrolebinding.rbac.authorization.k8s.io/knative-serving-operator unchanged",
"serviceaccount/knative-serving-operator unchanged",
"customresourcedefinition.apiextensions.k8s.io/knativeservings.serving.knative.dev unchanged",
"namespace/knative-sources unchanged",
"serviceaccount/controller-manager unchanged",
"clusterrole.rbac.authorization.k8s.io/eventing-sources-controller unchanged",
"clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-controller unchanged",
"clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-addressable-resolver unchanged",
"customresourcedefinition.apiextensions.k8s.io/githubsources.sources.eventing.knative.dev unchanged",
"service/controller unchanged",
"statefulset.apps/controller-manager configured",
"deployment.apps/openshift-pipelines-operator configured",
"clusterrole.rbac.authorization.k8s.io/openshift-pipelines-operator configured",
"clusterrolebinding.rbac.authorization.k8s.io/openshift-pipelines-operator unchanged",
"serviceaccount/openshift-pipelines-operator unchanged",
"namespace/openshift-pipelines unchanged",
"customresourcedefinition.apiextensions.k8s.io/config.operator.tekton.dev unchanged",
"config.operator.tekton.dev/cluster created",
"clusterrole.rbac.authorization.k8s.io/knative-serving-operator configured",
"clusterrole.rbac.authorization.k8s.io/knative-eventing-operator configured",
"customresourcedefinition.apiextensions.k8s.io/kabaneros.kabanero.io configured",
"customresourcedefinition.apiextensions.k8s.io/collections.kabanero.io configured",
"serviceaccount/kabanero-operator unchanged",
"deployment.apps/kabanero-operator configured",
"clusterrole.rbac.authorization.k8s.io/kabanero-operator configured",
"clusterrolebinding.rbac.authorization.k8s.io/kabanero-operator unchanged",
"clusterrole.rbac.authorization.k8s.io/cluster-admin added: \"kabanero-operator\"",
"NAME CREATED AT",
"services.serving.knative.dev 2019-10-24T13:12:56Z",
"NAME CREATED AT",
"clustertasks.tekton.dev 2019-10-24T13:12:23Z",
"config.operator.tekton.dev 2019-10-24T13:04:17Z",
"pipelineresources.tekton.dev 2019-10-24T13:12:30Z",
"pipelineruns.tekton.dev 2019-10-24T13:12:37Z",
"pipelines.tekton.dev 2019-10-24T13:12:23Z",
"taskruns.tekton.dev 2019-10-24T13:12:31Z",
"tasks.tekton.dev 2019-10-24T13:12:37Z",
"serviceaccount/tekton-webhooks-extension created",
"clusterrole.rbac.authorization.k8s.io/tekton-webhooks-extension-minimal created",
"clusterrolebinding.rbac.authorization.k8s.io/tekton-webhooks-extension-minimal created",
"deployment.apps/webhooks-extension created",
"service/webhooks-extension created",
"service.serving.knative.dev/webhooks-extension-sink created",
"task.tekton.dev/monitor-result-task created",
"serviceaccount/tekton-dashboard created",
"customresourcedefinition.apiextensions.k8s.io/extensions.dashboard.tekton.dev created",
"clusterrole.rbac.authorization.k8s.io/tekton-dashboard-minimal created",
"clusterrolebinding.rbac.authorization.k8s.io/tekton-dashboard-minimal created",
"deployment.apps/tekton-dashboard created",
"route.route.openshift.io/tekton-dashboard created",
"service/tekton-dashboard created",
"pipeline.tekton.dev/pipeline0 created",
"task.tekton.dev/pipeline0-task created",
"NAME AGE",
"tekton-dashboard-minimal 1s",
"clusterrole.rbac.authorization.k8s.io/tekton-dashboard-minimal patched",
"deployment.extensions/tekton-dashboard scaled",
"deployment.extensions/tekton-dashboard scaled",
"configmap/config-domain patched",
"NAME CREATED AT",
"extensions.dashboard.tekton.dev 2019-10-24T13:14:28Z",
"The installation script is complete. You can now create an instance",
"of the Kabanero CR. If you have cloned and curated a collection set,",
"apply the Kabanero CR that you created. Or, to create the default ",
"instance:",
"oc apply -n kabanero -f https://raw.githubusercontent.com/kabanero-io/kabanero-operator/0.2.0/config/samples/default.yaml"
]
}
TASK [install kabanero-foundation return code] ************************************************************************************
Thursday 24 October 2019 06:14:36 -0700 (0:00:00.085) 0:24:05.265 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": "0"
}
TASK [install kabanero foundation errors] *****************************************************************************************
Thursday 24 October 2019 06:14:36 -0700 (0:00:00.056) 0:24:05.321 ******
skipping: [svtcrc-91747-1.fyre.ibm.com]
TASK [enable monitoring, alerting, and telemetry services] ************************************************************************
Thursday 24 October 2019 06:14:36 -0700 (0:00:00.058) 0:24:05.380 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [enable monitoring, alerting, and telemetry services stdout] *****************************************************************
Thursday 24 October 2019 06:14:38 -0700 (0:00:01.407) 0:24:06.787 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"statefulset.apps/alertmanager-main scaled",
"statefulset.apps/prometheus-k8s scaled",
"deployment.extensions/cluster-monitoring-operator scaled",
"deployment.extensions/grafana scaled",
"deployment.extensions/kube-state-metrics scaled",
"deployment.extensions/openshift-state-metrics scaled",
"deployment.extensions/prometheus-adapter scaled",
"deployment.extensions/prometheus-operator scaled",
"deployment.extensions/telemeter-client scaled"
]
}
TASK [oc version] *****************************************************************************************************************
Thursday 24 October 2019 06:14:38 -0700 (0:00:00.061) 0:24:06.848 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [oc version stdout] **********************************************************************************************************
Thursday 24 October 2019 06:14:39 -0700 (0:00:01.037) 0:24:07.886 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"Client Version: v4.3.0",
"Server Version: 4.2.0",
"Kubernetes Version: v1.14.6+2e5ed54"
]
}
PLAY RECAP ************************************************************************************************************************
svtcrc-91747-1.fyre.ibm.com : ok=36 changed=18 unreachable=0 failed=0 skipped=6 rescued=0 ignored=0
Thursday 24 October 2019 06:14:39 -0700 (0:00:00.029) 0:24:07.916 ******
===============================================================================
install kabanero-foundation ---------------------------------------------------------------------------------------------- 659.33s
crc start ---------------------------------------------------------------------------------------------------------------- 656.96s
Install NetworkManager FireFox -------------------------------------------------------------------------------------------- 73.99s
crc setup ----------------------------------------------------------------------------------------------------------------- 38.73s
retrieve Kabanero Foundation master zip ------------------------------------------------------------------------------------ 2.40s
Gathering Facts ------------------------------------------------------------------------------------------------------------ 1.73s
prereq directories --------------------------------------------------------------------------------------------------------- 1.51s
oc login ------------------------------------------------------------------------------------------------------------------- 1.41s
enable monitoring, alerting, and telemetry services ------------------------------------------------------------------------ 1.41s
enable and start NetworkManager -------------------------------------------------------------------------------------------- 1.32s
crc pull secret ------------------------------------------------------------------------------------------------------------ 1.11s
unzip Kabanerio Foundation master zip -------------------------------------------------------------------------------------- 1.07s
oc version ----------------------------------------------------------------------------------------------------------------- 1.04s
check if crc is installed -------------------------------------------------------------------------------------------------- 0.71s
whoami --------------------------------------------------------------------------------------------------------------------- 0.69s
oc in user path ------------------------------------------------------------------------------------------------------------ 0.66s
password masking process --------------------------------------------------------------------------------------------------- 0.64s
crc creds ------------------------------------------------------------------------------------------------------------------ 0.50s
check if .crc is created --------------------------------------------------------------------------------------------------- 0.48s
link oc kubectl ------------------------------------------------------------------------------------------------------------ 0.44s
set admin passwd ----------------------------------------------------------------------------------------------------------- 0.36s
install kabanero-foundation stdout ----------------------------------------------------------------------------------------- 0.09s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.08s
crc setup stdout ----------------------------------------------------------------------------------------------------------- 0.07s
crc start stdout ----------------------------------------------------------------------------------------------------------- 0.07s
crc creds stdout ----------------------------------------------------------------------------------------------------------- 0.06s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.06s
oc login stdout ------------------------------------------------------------------------------------------------------------ 0.06s
enable monitoring, alerting, and telemetry services stdout ----------------------------------------------------------------- 0.06s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.06s
fail ----------------------------------------------------------------------------------------------------------------------- 0.06s
crc delete ----------------------------------------------------------------------------------------------------------------- 0.06s
crcstop stdout ------------------------------------------------------------------------------------------------------------- 0.06s
install kabanero foundation errors ----------------------------------------------------------------------------------------- 0.06s
install kabanero-foundation return code ------------------------------------------------------------------------------------ 0.06s
crc stop ------------------------------------------------------------------------------------------------------------------- 0.05s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s
crc delete stdout ---------------------------------------------------------------------------------------------------------- 0.05s
whoami stdout -------------------------------------------------------------------------------------------------------------- 0.05s
oc version stdout ---------------------------------------------------------------------------------------------------------- 0.03s
kabanero-crc complete. runtime 24 minutes
Setup complete. overall runtime 29 minutes
|
2.0
|
SVT: TER: Kabanero Automation test execution master branch - Setup: svtcrc-91747-1 - Original logs: kabanero-svt01.fyre.ibm.com:/home/nest/kabanero-crc-logs/kabanero-auto-crc.sh.2019-10-24-05:45:33.setup.log
PLAY [Create a one host Fyre stack] ***********************************************************************************************
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 05:45:37 -0700 (0:00:00.207) 0:00:00.207 ******
ok: [localhost]
TASK [check status of the named cluster] ******************************************************************************************
Thursday 24 October 2019 05:45:37 -0700 (0:00:00.047) 0:00:00.255 ******
changed: [localhost]
TASK [debug] **********************************************************************************************************************
Thursday 24 October 2019 05:45:38 -0700 (0:00:01.565) 0:00:01.821 ******
skipping: [localhost]
TASK [get the public ssh id] ******************************************************************************************************
Thursday 24 October 2019 05:45:38 -0700 (0:00:00.057) 0:00:01.878 ******
changed: [localhost -> localhost]
TASK [create input json file] *****************************************************************************************************
Thursday 24 October 2019 05:45:39 -0700 (0:00:00.306) 0:00:02.184 ******
changed: [localhost]
TASK [Create Fyre stack] **********************************************************************************************************
Thursday 24 October 2019 05:45:40 -0700 (0:00:01.016) 0:00:03.201 ******
changed: [localhost]
TASK [remove temp json] ***********************************************************************************************************
Thursday 24 October 2019 05:45:41 -0700 (0:00:01.866) 0:00:05.068 ******
changed: [localhost -> localhost]
TASK [check Fyre request status] **************************************************************************************************
Thursday 24 October 2019 05:45:42 -0700 (0:00:00.472) 0:00:05.540 ******
changed: [localhost]
TASK [debug] **********************************************************************************************************************
Thursday 24 October 2019 05:45:43 -0700 (0:00:00.905) 0:00:06.445 ******
ok: [localhost] => {
"fyreStatus.stdout": {
"request": [
{
"cluster_name": "",
"created": "2019-10-24 05:45:41",
"displayname": "steven r. schader",
"error_details": "0",
"id": "2420698",
"status": "building",
"type": "build",
"updated": "2019-10-24 05:45:42",
"username": "schader"
}
]
}
}
TASK [pause] **********************************************************************************************************************
Thursday 24 October 2019 05:45:43 -0700 (0:00:00.041) 0:00:06.487 ******
Pausing for 120 seconds
(ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort)
ok: [localhost]
TASK [get Fyre request status] ****************************************************************************************************
Thursday 24 October 2019 05:47:43 -0700 (0:02:00.047) 0:02:06.535 ******
changed: [localhost]
TASK [check Fyre request status] **************************************************************************************************
Thursday 24 October 2019 05:47:44 -0700 (0:00:01.028) 0:02:07.564 ******
skipping: [localhost]
TASK [check the host for a active ssh] ********************************************************************************************
Thursday 24 October 2019 05:47:44 -0700 (0:00:00.041) 0:02:07.605 ******
changed: [localhost -> localhost]
TASK [remove new host from localhost known_hosts] *********************************************************************************
Thursday 24 October 2019 05:47:45 -0700 (0:00:00.582) 0:02:08.188 ******
changed: [localhost -> localhost]
TASK [add host to known_hosts] ****************************************************************************************************
Thursday 24 October 2019 05:47:45 -0700 (0:00:00.315) 0:02:08.503 ******
changed: [localhost -> localhost]
TASK [create host inventory file] *************************************************************************************************
Thursday 24 October 2019 05:47:45 -0700 (0:00:00.378) 0:02:08.881 ******
changed: [localhost -> localhost]
TASK [create plain file with host] ************************************************************************************************
Thursday 24 October 2019 05:47:46 -0700 (0:00:00.595) 0:02:09.476 ******
changed: [localhost -> localhost]
TASK [create plain file with cluster name] ****************************************************************************************
Thursday 24 October 2019 05:47:46 -0700 (0:00:00.563) 0:02:10.040 ******
changed: [localhost -> localhost]
TASK [remove linked hosts.crc] ****************************************************************************************************
Thursday 24 October 2019 05:47:47 -0700 (0:00:00.574) 0:02:10.615 ******
changed: [localhost -> localhost]
TASK [link new host inventory file to hosts.crc] **********************************************************************************
Thursday 24 October 2019 05:47:47 -0700 (0:00:00.285) 0:02:10.901 ******
changed: [localhost -> localhost]
PLAY RECAP ************************************************************************************************************************
localhost : ok=18 changed=15 unreachable=0 failed=0 skipped=2 rescued=0 ignored=0
Thursday 24 October 2019 05:47:48 -0700 (0:00:00.288) 0:02:11.189 ******
===============================================================================
pause -------------------------------------------------------------------------------------------------------------------- 120.05s
Create Fyre stack ---------------------------------------------------------------------------------------------------------- 1.87s
check status of the named cluster ------------------------------------------------------------------------------------------ 1.57s
get Fyre request status ---------------------------------------------------------------------------------------------------- 1.03s
create input json file ----------------------------------------------------------------------------------------------------- 1.02s
check Fyre request status -------------------------------------------------------------------------------------------------- 0.91s
create host inventory file ------------------------------------------------------------------------------------------------- 0.60s
check the host for a active ssh -------------------------------------------------------------------------------------------- 0.58s
create plain file with cluster name ---------------------------------------------------------------------------------------- 0.57s
create plain file with host ------------------------------------------------------------------------------------------------ 0.56s
remove temp json ----------------------------------------------------------------------------------------------------------- 0.47s
add host to known_hosts ---------------------------------------------------------------------------------------------------- 0.38s
remove new host from localhost known_hosts --------------------------------------------------------------------------------- 0.32s
get the public ssh id ------------------------------------------------------------------------------------------------------ 0.31s
link new host inventory file to hosts.crc ---------------------------------------------------------------------------------- 0.29s
remove linked hosts.crc ---------------------------------------------------------------------------------------------------- 0.29s
debug ---------------------------------------------------------------------------------------------------------------------- 0.06s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s
debug ---------------------------------------------------------------------------------------------------------------------- 0.04s
check Fyre request status -------------------------------------------------------------------------------------------------- 0.04s
fyrecrcvm complete. runtime 2 minutes
PLAY [check RHEL 7 or greater x64 only] *******************************************************************************************
TASK [Gathering Facts] ************************************************************************************************************
Thursday 24 October 2019 05:47:50 -0700 (0:00:00.170) 0:00:00.170 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [Target host is RHEL7 ( or greater ) x86_64] *********************************************************************************
Thursday 24 October 2019 05:47:52 -0700 (0:00:01.803) 0:00:01.974 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
PLAY RECAP ************************************************************************************************************************
svtcrc-91747-1.fyre.ibm.com : ok=2 changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0
Thursday 24 October 2019 05:47:52 -0700 (0:00:00.039) 0:00:02.014 ******
===============================================================================
Gathering Facts ------------------------------------------------------------------------------------------------------------ 1.80s
Target host is RHEL7 ( or greater ) x86_64 --------------------------------------------------------------------------------- 0.04s
PLAY [Create a User] **************************************************************************************************************
TASK [Generate User Password file] ************************************************************************************************
Thursday 24 October 2019 05:47:55 -0700 (0:00:00.183) 0:00:00.183 ******
ok: [svtcrc-91747-1.fyre.ibm.com -> localhost]
TASK [Ensure User is Present] *****************************************************************************************************
Thursday 24 October 2019 05:47:56 -0700 (0:00:00.495) 0:00:00.678 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [authorized_key] *************************************************************************************************************
Thursday 24 October 2019 05:47:57 -0700 (0:00:01.163) 0:00:01.842 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [kevin sudoers] **************************************************************************************************************
Thursday 24 October 2019 05:47:58 -0700 (0:00:00.727) 0:00:02.569 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [kevin hard ulimit nofile] ***************************************************************************************************
Thursday 24 October 2019 05:47:58 -0700 (0:00:00.750) 0:00:03.320 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [kevin soft ulimit nofile] ***************************************************************************************************
Thursday 24 October 2019 05:47:59 -0700 (0:00:00.434) 0:00:03.754 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [add bin to path] ************************************************************************************************************
Thursday 24 October 2019 05:47:59 -0700 (0:00:00.437) 0:00:04.192 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
PLAY RECAP ************************************************************************************************************************
svtcrc-91747-1.fyre.ibm.com : ok=7 changed=6 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0
Thursday 24 October 2019 05:48:00 -0700 (0:00:00.428) 0:00:04.620 ******
===============================================================================
Ensure User is Present ----------------------------------------------------------------------------------------------------- 1.16s
kevin sudoers -------------------------------------------------------------------------------------------------------------- 0.75s
authorized_key ------------------------------------------------------------------------------------------------------------- 0.73s
Generate User Password file ------------------------------------------------------------------------------------------------ 0.50s
kevin soft ulimit nofile --------------------------------------------------------------------------------------------------- 0.44s
kevin hard ulimit nofile --------------------------------------------------------------------------------------------------- 0.43s
add bin to path ------------------------------------------------------------------------------------------------------------ 0.43s
PLAY [switch hosts.crc ansible_user] **********************************************************************************************
TASK [remove kevin ansible_user] **************************************************************************************************
Thursday 24 October 2019 05:48:02 -0700 (0:00:00.178) 0:00:00.178 ******
ok: [svtcrc-91747-1.fyre.ibm.com -> localhost]
TASK [set kevin ansible_user] *****************************************************************************************************
Thursday 24 October 2019 05:48:03 -0700 (0:00:00.476) 0:00:00.654 ******
changed: [svtcrc-91747-1.fyre.ibm.com -> localhost]
PLAY RECAP ************************************************************************************************************************
svtcrc-91747-1.fyre.ibm.com : ok=2 changed=1 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0
Thursday 24 October 2019 05:48:03 -0700 (0:00:00.266) 0:00:00.921 ******
===============================================================================
remove kevin ansible_user -------------------------------------------------------------------------------------------------- 0.48s
set kevin ansible_user ----------------------------------------------------------------------------------------------------- 0.27s
PLAY [SVT - install OCP CodeReadyContainer binary] ********************************************************************************
TASK [Gathering Facts] ************************************************************************************************************
Thursday 24 October 2019 05:48:06 -0700 (0:00:00.171) 0:00:00.171 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [check if crc is installed] **************************************************************************************************
Thursday 24 October 2019 05:48:08 -0700 (0:00:01.862) 0:00:02.034 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [archive dir structure] ******************************************************************************************************
Thursday 24 October 2019 05:48:09 -0700 (0:00:00.653) 0:00:02.688 ******
changed: [svtcrc-91747-1.fyre.ibm.com -> localhost]
TASK [debug] **********************************************************************************************************************
Thursday 24 October 2019 05:48:17 -0700 (0:00:08.477) 0:00:11.165 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"tarlist.stdout_lines": [
"crc-linux-1.0.0-amd64/",
"crc-linux-1.0.0-amd64/LICENSE",
"crc-linux-1.0.0-amd64/doc.pdf",
"crc-linux-1.0.0-amd64/crc"
]
}
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 05:48:17 -0700 (0:00:00.055) 0:00:11.221 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [copy crc archive] ***********************************************************************************************************
Thursday 24 October 2019 05:48:17 -0700 (0:00:00.062) 0:00:11.284 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [install crc] ****************************************************************************************************************
Thursday 24 October 2019 05:50:03 -0700 (0:01:46.017) 0:01:57.301 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [home bin] *******************************************************************************************************************
Thursday 24 October 2019 05:50:25 -0700 (0:00:22.091) 0:02:19.392 ******
changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/bin)
TASK [link crc] *******************************************************************************************************************
Thursday 24 October 2019 05:50:26 -0700 (0:00:00.761) 0:02:20.154 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
PLAY RECAP ************************************************************************************************************************
svtcrc-91747-1.fyre.ibm.com : ok=9 changed=5 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0
Thursday 24 October 2019 05:50:27 -0700 (0:00:00.483) 0:02:20.637 ******
===============================================================================
copy crc archive --------------------------------------------------------------------------------------------------------- 106.02s
install crc --------------------------------------------------------------------------------------------------------------- 22.09s
archive dir structure ------------------------------------------------------------------------------------------------------ 8.48s
Gathering Facts ------------------------------------------------------------------------------------------------------------ 1.86s
home bin ------------------------------------------------------------------------------------------------------------------- 0.76s
check if crc is installed -------------------------------------------------------------------------------------------------- 0.65s
link crc ------------------------------------------------------------------------------------------------------------------- 0.48s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.06s
debug ---------------------------------------------------------------------------------------------------------------------- 0.06s
crc-install complete. runtime 2 minutes
PLAY [SVT - Kabanero on crc] ******************************************************************************************************
TASK [Gathering Facts] ************************************************************************************************************
Thursday 24 October 2019 05:50:31 -0700 (0:00:00.175) 0:00:00.175 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 05:50:33 -0700 (0:00:01.727) 0:00:01.902 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [Install NetworkManager FireFox] *********************************************************************************************
Thursday 24 October 2019 05:50:33 -0700 (0:00:00.060) 0:00:01.963 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [enable and start NetworkManager] ********************************************************************************************
Thursday 24 October 2019 05:51:47 -0700 (0:01:13.989) 0:01:15.952 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [check if crc is installed] **************************************************************************************************
Thursday 24 October 2019 05:51:48 -0700 (0:00:01.324) 0:01:17.276 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [fail] ***********************************************************************************************************************
Thursday 24 October 2019 05:51:49 -0700 (0:00:00.713) 0:01:17.990 ******
skipping: [svtcrc-91747-1.fyre.ibm.com]
TASK [prereq directories] *********************************************************************************************************
Thursday 24 October 2019 05:51:49 -0700 (0:00:00.059) 0:01:18.049 ******
changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/git)
changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/testDir)
changed: [svtcrc-91747-1.fyre.ibm.com] => (item=~/kabanero/logs)
TASK [retrieve Kabanero Foundation master zip] ************************************************************************************
Thursday 24 October 2019 05:51:51 -0700 (0:00:01.507) 0:01:19.557 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [unzip Kabanerio Foundation master zip] **************************************************************************************
Thursday 24 October 2019 05:51:53 -0700 (0:00:02.402) 0:01:21.959 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [whoami] *********************************************************************************************************************
Thursday 24 October 2019 05:51:54 -0700 (0:00:01.073) 0:01:23.032 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [whoami stdout] **************************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.689) 0:01:23.722 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"kevin"
]
}
TASK [check if .crc is created] ***************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.047) 0:01:23.770 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc stop] *******************************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.481) 0:01:24.251 ******
skipping: [svtcrc-91747-1.fyre.ibm.com]
TASK [crcstop stdout] *************************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.053) 0:01:24.305 ******
skipping: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc delete] *****************************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.058) 0:01:24.364 ******
skipping: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc delete stdout] **********************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.059) 0:01:24.423 ******
skipping: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc setup] ******************************************************************************************************************
Thursday 24 October 2019 05:51:55 -0700 (0:00:00.050) 0:01:24.474 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc setup stdout] ***********************************************************************************************************
Thursday 24 October 2019 05:52:34 -0700 (0:00:38.733) 0:02:03.207 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"level=info msg=\"Checking if running as non-root\"",
"level=info msg=\"Caching oc binary\"",
"level=info msg=\"Setting up virtualization\"",
"level=info msg=\"Setting up KVM\"",
"level=info msg=\"Installing libvirt service and dependencies\"",
"level=info msg=\"Will use root access: install virtualization related packages\"",
"level=info msg=\"Adding user to libvirt group\"",
"level=info msg=\"Will use root access: add user to libvirt group\"",
"level=info msg=\"Enabling libvirt\"",
"level=info msg=\"Starting libvirt service\"",
"level=info msg=\"Will use root access: start libvirtd service\"",
"level=info msg=\"Checking if a supported libvirt version is installed\"",
"level=info msg=\"Installing crc-driver-libvirt\"",
"level=info msg=\"Removing older system-wide crc-driver-libvirt\"",
"level=info msg=\"Setting up libvirt 'crc' network\"",
"level=info msg=\"Starting libvirt 'crc' network\"",
"level=info msg=\"Checking if NetworkManager is installed\"",
"level=info msg=\"Checking if NetworkManager service is running\"",
"level=info msg=\"Writing Network Manager config for crc\"",
"level=info msg=\"Will use root access: write NetworkManager config in /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf\"",
"level=info msg=\"Will use root access: execute systemctl daemon-reload command\"",
"level=info msg=\"Will use root access: execute systemctl stop/start command\"",
"level=info msg=\"Writing dnsmasq config for crc\"",
"level=info msg=\"Will use root access: write dnsmasq configuration in /etc/NetworkManager/dnsmasq.d/crc.conf\"",
"level=info msg=\"Will use root access: execute systemctl daemon-reload command\"",
"level=info msg=\"Will use root access: execute systemctl stop/start command\"",
"level=info msg=\"Unpacking bundle from the CRC binary\"",
"Setup is complete, you can now run 'crc start' to start the OpenShift cluster"
]
}
TASK [crc pull secret] ************************************************************************************************************
Thursday 24 October 2019 05:52:34 -0700 (0:00:00.074) 0:02:03.281 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc start] ******************************************************************************************************************
Thursday 24 October 2019 05:52:35 -0700 (0:00:01.108) 0:02:04.390 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc start stdout] ***********************************************************************************************************
Thursday 24 October 2019 06:03:32 -0700 (0:10:56.959) 0:13:01.349 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"level=info msg=\"Checking if running as non-root\"",
"level=info msg=\"Checking if oc binary is cached\"",
"level=info msg=\"Checking if Virtualization is enabled\"",
"level=info msg=\"Checking if KVM is enabled\"",
"level=info msg=\"Checking if libvirt is installed\"",
"level=info msg=\"Checking if user is part of libvirt group\"",
"level=info msg=\"Checking if libvirt is enabled\"",
"level=info msg=\"Checking if libvirt daemon is running\"",
"level=info msg=\"Checking if a supported libvirt version is installed\"",
"level=info msg=\"Checking if crc-driver-libvirt is installed\"",
"level=info msg=\"Checking if libvirt 'crc' network is available\"",
"level=info msg=\"Checking if libvirt 'crc' network is active\"",
"level=info msg=\"Checking if NetworkManager is installed\"",
"level=info msg=\"Checking if NetworkManager service is running\"",
"level=info msg=\"Checking if /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf exists\"",
"level=info msg=\"Checking if /etc/NetworkManager/dnsmasq.d/crc.conf exists\"",
"level=info msg=\"Extracting bundle: crc_libvirt_4.2.0.crcbundle ...\"",
"level=info msg=\"Creating CodeReady Containers VM for OpenShift 4.2.0...\"",
"level=info msg=\"Verifying validity of the cluster certificates ...\"",
"level=info msg=\"Check internal and public DNS query ...\"",
"level=info msg=\"Copying kubeconfig file to instance dir ...\"",
"level=info msg=\"Adding user's pull secret and cluster ID ...\"",
"level=info msg=\"Starting OpenShift cluster ... [waiting 3m]\"",
"level=info",
"level=info msg=\"To access the cluster, first set up your environment by following 'crc oc-env' instructions\"",
"level=info msg=\"Then you can access it by running 'oc login -u developer -p developer https://api.crc.testing:6443'\"",
"level=info msg=\"To login as an admin, username is 'kubeadmin' and password is xx-password-xx\"",
"level=info",
"level=info msg=\"You can now run 'crc console' and use these credentials to access the OpenShift web console\"",
"The OpenShift cluster is running",
"level=warning msg=\"The cluster might report a degraded or error state. This is expected since several operators have been disable to lower the resource usage. For more information, please consult the documentation\""
]
}
TASK [oc in user path] ************************************************************************************************************
Thursday 24 October 2019 06:03:32 -0700 (0:00:00.066) 0:13:01.416 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [link oc kubectl] ************************************************************************************************************
Thursday 24 October 2019 06:03:33 -0700 (0:00:00.663) 0:13:02.079 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc creds] ******************************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.438) 0:13:02.518 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [crc creds stdout] ***********************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.498) 0:13:03.016 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"To login as a regular user, username is 'developer' and password is 'developer'.",
"To login as an admin, username is 'kubeadmin' and password is 'xx-password-xx'."
]
}
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.064) 0:13:03.081 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.052) 0:13:03.134 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.051) 0:13:03.185 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.051) 0:13:03.236 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [password masking process] ***************************************************************************************************
Thursday 24 October 2019 06:03:34 -0700 (0:00:00.076) 0:13:03.312 ******
changed: [svtcrc-91747-1.fyre.ibm.com -> localhost]
TASK [set admin passwd] ***********************************************************************************************************
Thursday 24 October 2019 06:03:35 -0700 (0:00:00.638) 0:13:03.951 ******
changed: [svtcrc-91747-1.fyre.ibm.com -> localhost]
TASK [set_fact] *******************************************************************************************************************
Thursday 24 October 2019 06:03:35 -0700 (0:00:00.356) 0:13:04.307 ******
ok: [svtcrc-91747-1.fyre.ibm.com]
TASK [oc login] *******************************************************************************************************************
Thursday 24 October 2019 06:03:35 -0700 (0:00:00.064) 0:13:04.371 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [oc login stdout] ************************************************************************************************************
Thursday 24 October 2019 06:03:37 -0700 (0:00:01.414) 0:13:05.786 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"Login successful.",
"",
"You have access to 51 projects, the list has been suppressed. You can list all projects with 'oc projects'",
"",
"Using project \"default\".",
"Welcome! See 'oc help' to get started."
]
}
TASK [install kabanero-foundation] ************************************************************************************************
Thursday 24 October 2019 06:03:37 -0700 (0:00:00.062) 0:13:05.849 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [install kabanero-foundation stdout] *****************************************************************************************
Thursday 24 October 2019 06:14:36 -0700 (0:10:59.331) 0:24:05.180 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-ingress-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:default\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:prometheus\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-egressgateway-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-citadel-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-ingressgateway-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-cleanup-old-ca-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-mixer-post-install-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-mixer-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-pilot-service-account\"]",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:istio-sidecar-injector-service-account\"]",
"clusterrole.rbac.authorization.k8s.io/cluster-admin added: \"istio-galley-service-account\"",
"securitycontextconstraints.security.openshift.io/anyuid added to: [\"system:serviceaccount:istio-system:cluster-local-gateway-service-account\"]",
"customresourcedefinition.apiextensions.k8s.io/virtualservices.networking.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/destinationrules.networking.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/serviceentries.networking.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/gateways.networking.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/envoyfilters.networking.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/clusterrbacconfigs.rbac.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/policies.authentication.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/meshpolicies.authentication.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/httpapispecbindings.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/httpapispecs.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/quotaspecbindings.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/quotaspecs.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/rules.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/attributemanifests.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/bypasses.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/circonuses.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/deniers.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/fluentds.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/kubernetesenvs.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/listcheckers.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/memquotas.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/noops.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/opas.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/prometheuses.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/rbacs.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/redisquotas.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/signalfxs.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/solarwindses.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/stackdrivers.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/statsds.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/stdios.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/apikeys.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/authorizations.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/checknothings.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/kuberneteses.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/listentries.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/logentries.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/edges.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/metrics.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/quotas.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/reportnothings.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/tracespans.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/rbacconfigs.rbac.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/serviceroles.rbac.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/servicerolebindings.rbac.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/adapters.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/instances.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/templates.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/handlers.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/cloudwatches.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/dogstatsds.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/sidecars.networking.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/zipkins.config.istio.io created",
"customresourcedefinition.apiextensions.k8s.io/clusterissuers.certmanager.k8s.io created",
"customresourcedefinition.apiextensions.k8s.io/issuers.certmanager.k8s.io created",
"customresourcedefinition.apiextensions.k8s.io/certificates.certmanager.k8s.io created",
"customresourcedefinition.apiextensions.k8s.io/orders.certmanager.k8s.io created",
"customresourcedefinition.apiextensions.k8s.io/challenges.certmanager.k8s.io created",
"namespace/istio-system created",
"customresourcedefinition.apiextensions.k8s.io/virtualservices.networking.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/destinationrules.networking.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/serviceentries.networking.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/gateways.networking.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/envoyfilters.networking.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/clusterrbacconfigs.rbac.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/policies.authentication.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/meshpolicies.authentication.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/httpapispecbindings.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/httpapispecs.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/quotaspecbindings.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/quotaspecs.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/rules.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/attributemanifests.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/bypasses.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/circonuses.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/deniers.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/fluentds.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/kubernetesenvs.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/listcheckers.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/memquotas.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/noops.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/opas.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/prometheuses.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/rbacs.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/redisquotas.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/signalfxs.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/solarwindses.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/stackdrivers.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/statsds.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/stdios.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/apikeys.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/authorizations.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/checknothings.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/kuberneteses.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/listentries.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/logentries.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/edges.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/metrics.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/quotas.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/reportnothings.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/tracespans.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/rbacconfigs.rbac.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/serviceroles.rbac.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/servicerolebindings.rbac.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/adapters.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/instances.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/templates.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/handlers.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/cloudwatches.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/dogstatsds.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/sidecars.networking.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/zipkins.config.istio.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/clusterissuers.certmanager.k8s.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/issuers.certmanager.k8s.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/orders.certmanager.k8s.io unchanged",
"customresourcedefinition.apiextensions.k8s.io/challenges.certmanager.k8s.io unchanged",
"secret/kiali created",
"configmap/istio-galley-configuration created",
"configmap/istio-grafana-custom-resources created",
"configmap/istio-grafana-configuration-dashboards-galley-dashboard created",
"configmap/istio-grafana-configuration-dashboards-istio-mesh-dashboard created",
"configmap/istio-grafana-configuration-dashboards-istio-performance-dashboard created",
"configmap/istio-grafana-configuration-dashboards-istio-service-dashboard created",
"configmap/istio-grafana-configuration-dashboards-istio-workload-dashboard created",
"configmap/istio-grafana-configuration-dashboards-mixer-dashboard created",
"configmap/istio-grafana-configuration-dashboards-pilot-dashboard created",
"configmap/istio-grafana created",
"configmap/kiali created",
"configmap/prometheus created",
"configmap/istio-security-custom-resources created",
"configmap/istio created",
"configmap/istio-sidecar-injector created",
"serviceaccount/istio-galley-service-account created",
"serviceaccount/istio-egressgateway-service-account created",
"serviceaccount/istio-ingressgateway-service-account created",
"serviceaccount/istio-grafana-post-install-account created",
"clusterrole.rbac.authorization.k8s.io/istio-grafana-post-install-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-grafana-post-install-role-binding-istio-system created",
"job.batch/istio-grafana-post-install-1.1.7 created",
"serviceaccount/kiali-service-account created",
"serviceaccount/istio-mixer-service-account created",
"serviceaccount/istio-pilot-service-account created",
"serviceaccount/prometheus created",
"serviceaccount/istio-cleanup-secrets-service-account created",
"clusterrole.rbac.authorization.k8s.io/istio-cleanup-secrets-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-cleanup-secrets-istio-system created",
"job.batch/istio-cleanup-secrets-1.1.7 created",
"serviceaccount/istio-security-post-install-account created",
"clusterrole.rbac.authorization.k8s.io/istio-security-post-install-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-security-post-install-role-binding-istio-system created",
"job.batch/istio-security-post-install-1.1.7 created",
"serviceaccount/istio-citadel-service-account created",
"serviceaccount/istio-sidecar-injector-service-account created",
"serviceaccount/istio-multi created",
"clusterrole.rbac.authorization.k8s.io/istio-galley-istio-system created",
"clusterrole.rbac.authorization.k8s.io/istio-egressgateway-istio-system created",
"clusterrole.rbac.authorization.k8s.io/istio-ingressgateway-istio-system created",
"clusterrole.rbac.authorization.k8s.io/kiali created",
"clusterrole.rbac.authorization.k8s.io/kiali-viewer created",
"clusterrole.rbac.authorization.k8s.io/istio-mixer-istio-system created",
"clusterrole.rbac.authorization.k8s.io/istio-pilot-istio-system created",
"clusterrole.rbac.authorization.k8s.io/prometheus-istio-system created",
"clusterrole.rbac.authorization.k8s.io/istio-citadel-istio-system created",
"clusterrole.rbac.authorization.k8s.io/istio-sidecar-injector-istio-system created",
"clusterrole.rbac.authorization.k8s.io/istio-reader created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-galley-admin-role-binding-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-egressgateway-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-ingressgateway-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-kiali-admin-role-binding-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-mixer-admin-role-binding-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-pilot-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/prometheus-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-citadel-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-sidecar-injector-admin-role-binding-istio-system created",
"clusterrolebinding.rbac.authorization.k8s.io/istio-multi created",
"role.rbac.authorization.k8s.io/istio-ingressgateway-sds created",
"rolebinding.rbac.authorization.k8s.io/istio-ingressgateway-sds created",
"service/istio-galley created",
"service/istio-egressgateway created",
"service/istio-ingressgateway created",
"service/grafana created",
"service/kiali created",
"service/istio-policy created",
"service/istio-telemetry created",
"service/istio-pilot created",
"service/prometheus created",
"service/istio-citadel created",
"service/istio-sidecar-injector created",
"deployment.extensions/istio-galley created",
"deployment.extensions/istio-egressgateway created",
"deployment.extensions/istio-ingressgateway created",
"deployment.extensions/grafana created",
"deployment.extensions/kiali created",
"deployment.extensions/istio-policy created",
"deployment.extensions/istio-telemetry created",
"deployment.extensions/istio-pilot created",
"deployment.extensions/prometheus created",
"deployment.extensions/istio-citadel created",
"deployment.extensions/istio-sidecar-injector created",
"deployment.extensions/istio-tracing created",
"horizontalpodautoscaler.autoscaling/istio-egressgateway created",
"horizontalpodautoscaler.autoscaling/istio-ingressgateway created",
"horizontalpodautoscaler.autoscaling/istio-policy created",
"horizontalpodautoscaler.autoscaling/istio-telemetry created",
"horizontalpodautoscaler.autoscaling/istio-pilot created",
"service/jaeger-query created",
"service/jaeger-collector created",
"service/jaeger-agent created",
"service/zipkin created",
"service/tracing created",
"mutatingwebhookconfiguration.admissionregistration.k8s.io/istio-sidecar-injector created",
"poddisruptionbudget.policy/istio-galley created",
"poddisruptionbudget.policy/istio-egressgateway created",
"poddisruptionbudget.policy/istio-ingressgateway created",
"poddisruptionbudget.policy/istio-policy created",
"poddisruptionbudget.policy/istio-telemetry created",
"poddisruptionbudget.policy/istio-pilot created",
"attributemanifest.config.istio.io/istioproxy created",
"attributemanifest.config.istio.io/kubernetes created",
"handler.config.istio.io/stdio created",
"logentry.config.istio.io/accesslog created",
"logentry.config.istio.io/tcpaccesslog created",
"rule.config.istio.io/stdio created",
"rule.config.istio.io/stdiotcp created",
"metric.config.istio.io/requestcount created",
"metric.config.istio.io/requestduration created",
"metric.config.istio.io/requestsize created",
"metric.config.istio.io/responsesize created",
"metric.config.istio.io/tcpbytesent created",
"metric.config.istio.io/tcpbytereceived created",
"metric.config.istio.io/tcpconnectionsopened created",
"metric.config.istio.io/tcpconnectionsclosed created",
"handler.config.istio.io/prometheus created",
"rule.config.istio.io/promhttp created",
"rule.config.istio.io/promtcp created",
"rule.config.istio.io/promtcpconnectionopen created",
"rule.config.istio.io/promtcpconnectionclosed created",
"handler.config.istio.io/kubernetesenv created",
"rule.config.istio.io/kubeattrgenrulerule created",
"rule.config.istio.io/tcpkubeattrgenrulerule created",
"kubernetes.config.istio.io/attributes created",
"destinationrule.networking.istio.io/istio-policy created",
"destinationrule.networking.istio.io/istio-telemetry created",
"Now using project \"kabanero\" on server \"https://api.crc.testing:6443\".",
"",
"You can add applications to this project with the 'new-app' command. For example, try:",
"",
" oc new-app django-psql-example",
"",
"to build a new example application in Python. Or use kubectl to deploy a simple Kubernetes application:",
"",
" kubectl create deployment hello-node --image=gcr.io/hello-minikube-zero-install/hello-node",
"",
"namespace/kabanero configured",
"deployment.apps/knative-eventing-operator created",
"role.rbac.authorization.k8s.io/knative-eventing-operator created",
"clusterrole.rbac.authorization.k8s.io/knative-eventing-operator created",
"rolebinding.rbac.authorization.k8s.io/knative-eventing-operator created",
"clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-operator created",
"serviceaccount/knative-eventing-operator created",
"customresourcedefinition.apiextensions.k8s.io/knativeeventings.eventing.knative.dev created",
"deployment.apps/knative-serving-operator created",
"role.rbac.authorization.k8s.io/knative-serving-operator created",
"clusterrole.rbac.authorization.k8s.io/knative-serving-operator created",
"rolebinding.rbac.authorization.k8s.io/knative-serving-operator created",
"clusterrolebinding.rbac.authorization.k8s.io/knative-serving-operator created",
"serviceaccount/knative-serving-operator created",
"customresourcedefinition.apiextensions.k8s.io/knativeservings.serving.knative.dev created",
"namespace/knative-sources created",
"serviceaccount/controller-manager created",
"clusterrole.rbac.authorization.k8s.io/eventing-sources-controller created",
"clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-controller created",
"clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-addressable-resolver created",
"customresourcedefinition.apiextensions.k8s.io/githubsources.sources.eventing.knative.dev created",
"service/controller created",
"statefulset.apps/controller-manager created",
"deployment.apps/openshift-pipelines-operator created",
"clusterrole.rbac.authorization.k8s.io/openshift-pipelines-operator created",
"clusterrolebinding.rbac.authorization.k8s.io/openshift-pipelines-operator created",
"serviceaccount/openshift-pipelines-operator created",
"namespace/openshift-pipelines created",
"customresourcedefinition.apiextensions.k8s.io/config.operator.tekton.dev created",
"clusterrole.rbac.authorization.k8s.io/knative-serving-operator configured",
"clusterrole.rbac.authorization.k8s.io/knative-eventing-operator configured",
"customresourcedefinition.apiextensions.k8s.io/kabaneros.kabanero.io created",
"customresourcedefinition.apiextensions.k8s.io/collections.kabanero.io created",
"serviceaccount/kabanero-operator created",
"deployment.apps/kabanero-operator created",
"clusterrole.rbac.authorization.k8s.io/kabanero-operator created",
"clusterrolebinding.rbac.authorization.k8s.io/kabanero-operator created",
"namespace/kabanero unchanged",
"deployment.apps/knative-eventing-operator configured",
"role.rbac.authorization.k8s.io/knative-eventing-operator configured",
"clusterrole.rbac.authorization.k8s.io/knative-eventing-operator configured",
"rolebinding.rbac.authorization.k8s.io/knative-eventing-operator unchanged",
"clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-operator unchanged",
"serviceaccount/knative-eventing-operator unchanged",
"customresourcedefinition.apiextensions.k8s.io/knativeeventings.eventing.knative.dev unchanged",
"deployment.apps/knative-serving-operator configured",
"role.rbac.authorization.k8s.io/knative-serving-operator configured",
"clusterrole.rbac.authorization.k8s.io/knative-serving-operator configured",
"rolebinding.rbac.authorization.k8s.io/knative-serving-operator unchanged",
"clusterrolebinding.rbac.authorization.k8s.io/knative-serving-operator unchanged",
"serviceaccount/knative-serving-operator unchanged",
"customresourcedefinition.apiextensions.k8s.io/knativeservings.serving.knative.dev unchanged",
"namespace/knative-sources unchanged",
"serviceaccount/controller-manager unchanged",
"clusterrole.rbac.authorization.k8s.io/eventing-sources-controller unchanged",
"clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-controller unchanged",
"clusterrolebinding.rbac.authorization.k8s.io/eventing-sources-addressable-resolver unchanged",
"customresourcedefinition.apiextensions.k8s.io/githubsources.sources.eventing.knative.dev unchanged",
"service/controller unchanged",
"statefulset.apps/controller-manager configured",
"deployment.apps/openshift-pipelines-operator configured",
"clusterrole.rbac.authorization.k8s.io/openshift-pipelines-operator configured",
"clusterrolebinding.rbac.authorization.k8s.io/openshift-pipelines-operator unchanged",
"serviceaccount/openshift-pipelines-operator unchanged",
"namespace/openshift-pipelines unchanged",
"customresourcedefinition.apiextensions.k8s.io/config.operator.tekton.dev unchanged",
"config.operator.tekton.dev/cluster created",
"clusterrole.rbac.authorization.k8s.io/knative-serving-operator configured",
"clusterrole.rbac.authorization.k8s.io/knative-eventing-operator configured",
"customresourcedefinition.apiextensions.k8s.io/kabaneros.kabanero.io configured",
"customresourcedefinition.apiextensions.k8s.io/collections.kabanero.io configured",
"serviceaccount/kabanero-operator unchanged",
"deployment.apps/kabanero-operator configured",
"clusterrole.rbac.authorization.k8s.io/kabanero-operator configured",
"clusterrolebinding.rbac.authorization.k8s.io/kabanero-operator unchanged",
"clusterrole.rbac.authorization.k8s.io/cluster-admin added: \"kabanero-operator\"",
"NAME CREATED AT",
"services.serving.knative.dev 2019-10-24T13:12:56Z",
"NAME CREATED AT",
"clustertasks.tekton.dev 2019-10-24T13:12:23Z",
"config.operator.tekton.dev 2019-10-24T13:04:17Z",
"pipelineresources.tekton.dev 2019-10-24T13:12:30Z",
"pipelineruns.tekton.dev 2019-10-24T13:12:37Z",
"pipelines.tekton.dev 2019-10-24T13:12:23Z",
"taskruns.tekton.dev 2019-10-24T13:12:31Z",
"tasks.tekton.dev 2019-10-24T13:12:37Z",
"serviceaccount/tekton-webhooks-extension created",
"clusterrole.rbac.authorization.k8s.io/tekton-webhooks-extension-minimal created",
"clusterrolebinding.rbac.authorization.k8s.io/tekton-webhooks-extension-minimal created",
"deployment.apps/webhooks-extension created",
"service/webhooks-extension created",
"service.serving.knative.dev/webhooks-extension-sink created",
"task.tekton.dev/monitor-result-task created",
"serviceaccount/tekton-dashboard created",
"customresourcedefinition.apiextensions.k8s.io/extensions.dashboard.tekton.dev created",
"clusterrole.rbac.authorization.k8s.io/tekton-dashboard-minimal created",
"clusterrolebinding.rbac.authorization.k8s.io/tekton-dashboard-minimal created",
"deployment.apps/tekton-dashboard created",
"route.route.openshift.io/tekton-dashboard created",
"service/tekton-dashboard created",
"pipeline.tekton.dev/pipeline0 created",
"task.tekton.dev/pipeline0-task created",
"NAME AGE",
"tekton-dashboard-minimal 1s",
"clusterrole.rbac.authorization.k8s.io/tekton-dashboard-minimal patched",
"deployment.extensions/tekton-dashboard scaled",
"deployment.extensions/tekton-dashboard scaled",
"configmap/config-domain patched",
"NAME CREATED AT",
"extensions.dashboard.tekton.dev 2019-10-24T13:14:28Z",
"The installation script is complete. You can now create an instance",
"of the Kabanero CR. If you have cloned and curated a collection set,",
"apply the Kabanero CR that you created. Or, to create the default ",
"instance:",
"oc apply -n kabanero -f https://raw.githubusercontent.com/kabanero-io/kabanero-operator/0.2.0/config/samples/default.yaml"
]
}
TASK [install kabanero-foundation return code] ************************************************************************************
Thursday 24 October 2019 06:14:36 -0700 (0:00:00.085) 0:24:05.265 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": "0"
}
TASK [install kabanero foundation errors] *****************************************************************************************
Thursday 24 October 2019 06:14:36 -0700 (0:00:00.056) 0:24:05.321 ******
skipping: [svtcrc-91747-1.fyre.ibm.com]
TASK [enable monitoring, alerting, and telemetry services] ************************************************************************
Thursday 24 October 2019 06:14:36 -0700 (0:00:00.058) 0:24:05.380 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [enable monitoring, alerting, and telemetry services stdout] *****************************************************************
Thursday 24 October 2019 06:14:38 -0700 (0:00:01.407) 0:24:06.787 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"statefulset.apps/alertmanager-main scaled",
"statefulset.apps/prometheus-k8s scaled",
"deployment.extensions/cluster-monitoring-operator scaled",
"deployment.extensions/grafana scaled",
"deployment.extensions/kube-state-metrics scaled",
"deployment.extensions/openshift-state-metrics scaled",
"deployment.extensions/prometheus-adapter scaled",
"deployment.extensions/prometheus-operator scaled",
"deployment.extensions/telemeter-client scaled"
]
}
TASK [oc version] *****************************************************************************************************************
Thursday 24 October 2019 06:14:38 -0700 (0:00:00.061) 0:24:06.848 ******
changed: [svtcrc-91747-1.fyre.ibm.com]
TASK [oc version stdout] **********************************************************************************************************
Thursday 24 October 2019 06:14:39 -0700 (0:00:01.037) 0:24:07.886 ******
ok: [svtcrc-91747-1.fyre.ibm.com] => {
"msg": [
"Client Version: v4.3.0",
"Server Version: 4.2.0",
"Kubernetes Version: v1.14.6+2e5ed54"
]
}
PLAY RECAP ************************************************************************************************************************
svtcrc-91747-1.fyre.ibm.com : ok=36 changed=18 unreachable=0 failed=0 skipped=6 rescued=0 ignored=0
Thursday 24 October 2019 06:14:39 -0700 (0:00:00.029) 0:24:07.916 ******
===============================================================================
install kabanero-foundation ---------------------------------------------------------------------------------------------- 659.33s
crc start ---------------------------------------------------------------------------------------------------------------- 656.96s
Install NetworkManager FireFox -------------------------------------------------------------------------------------------- 73.99s
crc setup ----------------------------------------------------------------------------------------------------------------- 38.73s
retrieve Kabanero Foundation master zip ------------------------------------------------------------------------------------ 2.40s
Gathering Facts ------------------------------------------------------------------------------------------------------------ 1.73s
prereq directories --------------------------------------------------------------------------------------------------------- 1.51s
oc login ------------------------------------------------------------------------------------------------------------------- 1.41s
enable monitoring, alerting, and telemetry services ------------------------------------------------------------------------ 1.41s
enable and start NetworkManager -------------------------------------------------------------------------------------------- 1.32s
crc pull secret ------------------------------------------------------------------------------------------------------------ 1.11s
unzip Kabanerio Foundation master zip -------------------------------------------------------------------------------------- 1.07s
oc version ----------------------------------------------------------------------------------------------------------------- 1.04s
check if crc is installed -------------------------------------------------------------------------------------------------- 0.71s
whoami --------------------------------------------------------------------------------------------------------------------- 0.69s
oc in user path ------------------------------------------------------------------------------------------------------------ 0.66s
password masking process --------------------------------------------------------------------------------------------------- 0.64s
crc creds ------------------------------------------------------------------------------------------------------------------ 0.50s
check if .crc is created --------------------------------------------------------------------------------------------------- 0.48s
link oc kubectl ------------------------------------------------------------------------------------------------------------ 0.44s
set admin passwd ----------------------------------------------------------------------------------------------------------- 0.36s
install kabanero-foundation stdout ----------------------------------------------------------------------------------------- 0.09s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.08s
crc setup stdout ----------------------------------------------------------------------------------------------------------- 0.07s
crc start stdout ----------------------------------------------------------------------------------------------------------- 0.07s
crc creds stdout ----------------------------------------------------------------------------------------------------------- 0.06s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.06s
oc login stdout ------------------------------------------------------------------------------------------------------------ 0.06s
enable monitoring, alerting, and telemetry services stdout ----------------------------------------------------------------- 0.06s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.06s
fail ----------------------------------------------------------------------------------------------------------------------- 0.06s
crc delete ----------------------------------------------------------------------------------------------------------------- 0.06s
crcstop stdout ------------------------------------------------------------------------------------------------------------- 0.06s
install kabanero foundation errors ----------------------------------------------------------------------------------------- 0.06s
install kabanero-foundation return code ------------------------------------------------------------------------------------ 0.06s
crc stop ------------------------------------------------------------------------------------------------------------------- 0.05s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s
set_fact ------------------------------------------------------------------------------------------------------------------- 0.05s
crc delete stdout ---------------------------------------------------------------------------------------------------------- 0.05s
whoami stdout -------------------------------------------------------------------------------------------------------------- 0.05s
oc version stdout ---------------------------------------------------------------------------------------------------------- 0.03s
kabanero-crc complete. runtime 24 minutes
Setup complete. overall runtime 29 minutes
|
test
|
svt ter kabanero automation test execution master branch setup svtcrc original logs kabanero fyre ibm com home nest kabanero crc logs kabanero auto crc sh setup log play task thursday october ok task thursday october changed task thursday october skipping task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october ok fyrestatus stdout request cluster name created displayname steven r schader error details id status building type build updated username schader task thursday october pausing for seconds ctrl c then c continue early ctrl c then a abort ok task thursday october changed task thursday october skipping task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed play recap localhost ok changed unreachable failed skipped rescued ignored thursday october pause create fyre stack check status of the named cluster get fyre request status create input json file check fyre request status create host inventory file check the host for a active ssh create plain file with cluster name create plain file with host remove temp json add host to known hosts remove new host from localhost known hosts get the public ssh id link new host inventory file to hosts crc remove linked hosts crc debug set fact debug check fyre request status fyrecrcvm complete runtime minutes play task thursday october ok task thursday october ok play recap svtcrc fyre ibm com ok changed unreachable failed skipped rescued ignored thursday october gathering facts target host is or greater play task thursday october ok task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed task thursday october changed play recap svtcrc fyre ibm com ok changed unreachable failed skipped rescued ignored thursday october ensure user is present kevin sudoers authorized key generate user password file kevin soft ulimit nofile kevin hard ulimit nofile add bin to path play task thursday october ok task thursday october changed play recap svtcrc fyre ibm com ok changed unreachable failed skipped rescued ignored thursday october remove kevin ansible user set kevin ansible user play task thursday october ok task thursday october ok task thursday october changed task thursday october ok tarlist stdout lines crc linux crc linux license crc linux doc pdf crc linux crc task thursday october ok task thursday october changed task thursday october changed task thursday october changed item bin task thursday october changed play recap svtcrc fyre ibm com ok changed unreachable failed skipped rescued ignored thursday october copy crc archive install crc archive dir structure gathering facts home bin check if crc is installed link crc set fact debug crc install complete runtime minutes play task thursday october ok task thursday october ok task thursday october changed task thursday october changed task thursday october ok task thursday october skipping task thursday october changed item git changed item testdir changed item kabanero logs task thursday october changed task thursday october changed task thursday october changed task thursday october ok msg kevin task thursday october ok task thursday october skipping task thursday october skipping task thursday october skipping task thursday october skipping task thursday october changed task thursday october ok msg level info msg checking if running as non root level info msg caching oc binary level info msg setting up virtualization level info msg setting up kvm level info msg installing libvirt service and dependencies level info msg will use root access install virtualization related packages level info msg adding user to libvirt group level info msg will use root access add user to libvirt group level info msg enabling libvirt level info msg starting libvirt service level info msg will use root access start libvirtd service level info msg checking if a supported libvirt version is installed level info msg installing crc driver libvirt level info msg removing older system wide crc driver libvirt level info msg setting up libvirt crc network level info msg starting libvirt crc network level info msg checking if networkmanager is installed level info msg checking if networkmanager service is running level info msg writing network manager config for crc level info msg will use root access write networkmanager config in etc networkmanager conf d crc nm dnsmasq conf level info msg will use root access execute systemctl daemon reload command level info msg will use root access execute systemctl stop start command level info msg writing dnsmasq config for crc level info msg will use root access write dnsmasq configuration in etc networkmanager dnsmasq d crc conf level info msg will use root access execute systemctl daemon reload command level info msg will use root access execute systemctl stop start command level info msg unpacking bundle from the crc binary setup is complete you can now run crc start to start the openshift cluster task thursday october changed task thursday october changed task thursday october ok msg level info msg checking if running as non root level info msg checking if oc binary is cached level info msg checking if virtualization is enabled level info msg checking if kvm is enabled level info msg checking if libvirt is installed level info msg checking if user is part of libvirt group level info msg checking if libvirt is enabled level info msg checking if libvirt daemon is running level info msg checking if a supported libvirt version is installed level info msg checking if crc driver libvirt is installed level info msg checking if libvirt crc network is available level info msg checking if libvirt crc network is active level info msg checking if networkmanager is installed level info msg checking if networkmanager service is running level info msg checking if etc networkmanager conf d crc nm dnsmasq conf exists level info msg checking if etc networkmanager dnsmasq d crc conf exists level info msg extracting bundle crc libvirt crcbundle level info msg creating codeready containers vm for openshift level info msg verifying validity of the cluster certificates level info msg check internal and public dns query level info msg copying kubeconfig file to instance dir level info msg adding user s pull secret and cluster id level info msg starting openshift cluster level info level info msg to access the cluster first set up your environment by following crc oc env instructions level info msg then you can access it by running oc login u developer p developer level info msg to login as an admin username is kubeadmin and password is xx password xx level info level info msg you can now run crc console and use these credentials to access the openshift web console the openshift cluster is running level warning msg the cluster might report a degraded or error state this is expected since several operators have been disable to lower the resource usage for more information please consult the documentation task thursday october changed task thursday october changed task thursday october changed task thursday october ok msg to login as a regular user username is developer and password is developer to login as an admin username is kubeadmin and password is xx password xx task thursday october ok task thursday october ok task thursday october ok task thursday october ok task thursday october changed task thursday october changed task thursday october ok task thursday october changed task thursday october ok msg login successful you have access to projects the list has been suppressed you can list all projects with oc projects using project default welcome see oc help to get started task thursday october changed task thursday october ok msg securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to securitycontextconstraints security openshift io anyuid added to clusterrole rbac authorization io cluster admin added istio galley service account securitycontextconstraints security openshift io anyuid added to customresourcedefinition apiextensions io virtualservices networking istio io created customresourcedefinition apiextensions io destinationrules networking istio io created customresourcedefinition apiextensions io serviceentries networking istio io created customresourcedefinition apiextensions io gateways networking istio io created customresourcedefinition apiextensions io envoyfilters networking istio io created customresourcedefinition apiextensions io clusterrbacconfigs rbac istio io created customresourcedefinition apiextensions io policies authentication istio io created customresourcedefinition apiextensions io meshpolicies authentication istio io created customresourcedefinition apiextensions io httpapispecbindings config istio io created customresourcedefinition apiextensions io httpapispecs config istio io created customresourcedefinition apiextensions io quotaspecbindings config istio io created customresourcedefinition apiextensions io quotaspecs config istio io created customresourcedefinition apiextensions io rules config istio io created customresourcedefinition apiextensions io attributemanifests config istio io created customresourcedefinition apiextensions io bypasses config istio io created customresourcedefinition apiextensions io circonuses config istio io created customresourcedefinition apiextensions io deniers config istio io created customresourcedefinition apiextensions io fluentds config istio io created customresourcedefinition apiextensions io kubernetesenvs config istio io created customresourcedefinition apiextensions io listcheckers config istio io created customresourcedefinition apiextensions io memquotas config istio io created customresourcedefinition apiextensions io noops config istio io created customresourcedefinition apiextensions io opas config istio io created customresourcedefinition apiextensions io prometheuses config istio io created customresourcedefinition apiextensions io rbacs config istio io created customresourcedefinition apiextensions io redisquotas config istio io created customresourcedefinition apiextensions io signalfxs config istio io created customresourcedefinition apiextensions io solarwindses config istio io created customresourcedefinition apiextensions io stackdrivers config istio io created customresourcedefinition apiextensions io statsds config istio io created customresourcedefinition apiextensions io stdios config istio io created customresourcedefinition apiextensions io apikeys config istio io created customresourcedefinition apiextensions io authorizations config istio io created customresourcedefinition apiextensions io checknothings config istio io created customresourcedefinition apiextensions io kuberneteses config istio io created customresourcedefinition apiextensions io listentries config istio io created customresourcedefinition apiextensions io logentries config istio io created customresourcedefinition apiextensions io edges config istio io created customresourcedefinition apiextensions io metrics config istio io created customresourcedefinition apiextensions io quotas config istio io created customresourcedefinition apiextensions io reportnothings config istio io created customresourcedefinition apiextensions io tracespans config istio io created customresourcedefinition apiextensions io rbacconfigs rbac istio io created customresourcedefinition apiextensions io serviceroles rbac istio io created customresourcedefinition apiextensions io servicerolebindings rbac istio io created customresourcedefinition apiextensions io adapters config istio io created customresourcedefinition apiextensions io instances config istio io created customresourcedefinition apiextensions io templates config istio io created customresourcedefinition apiextensions io handlers config istio io created customresourcedefinition apiextensions io cloudwatches config istio io created customresourcedefinition apiextensions io dogstatsds config istio io created customresourcedefinition apiextensions io sidecars networking istio io created customresourcedefinition apiextensions io zipkins config istio io created customresourcedefinition apiextensions io clusterissuers certmanager io created customresourcedefinition apiextensions io issuers certmanager io created customresourcedefinition apiextensions io certificates certmanager io created customresourcedefinition apiextensions io orders certmanager io created customresourcedefinition apiextensions io challenges certmanager io created namespace istio system created customresourcedefinition apiextensions io virtualservices networking istio io unchanged customresourcedefinition apiextensions io destinationrules networking istio io unchanged customresourcedefinition apiextensions io serviceentries networking istio io unchanged customresourcedefinition apiextensions io gateways networking istio io unchanged customresourcedefinition apiextensions io envoyfilters networking istio io unchanged customresourcedefinition apiextensions io clusterrbacconfigs rbac istio io unchanged customresourcedefinition apiextensions io policies authentication istio io unchanged customresourcedefinition apiextensions io meshpolicies authentication istio io unchanged customresourcedefinition apiextensions io httpapispecbindings config istio io unchanged customresourcedefinition apiextensions io httpapispecs config istio io unchanged customresourcedefinition apiextensions io quotaspecbindings config istio io unchanged customresourcedefinition apiextensions io quotaspecs config istio io unchanged customresourcedefinition apiextensions io rules config istio io unchanged customresourcedefinition apiextensions io attributemanifests config istio io unchanged customresourcedefinition apiextensions io bypasses config istio io unchanged customresourcedefinition apiextensions io circonuses config istio io unchanged customresourcedefinition apiextensions io deniers config istio io unchanged customresourcedefinition apiextensions io fluentds config istio io unchanged customresourcedefinition apiextensions io kubernetesenvs config istio io unchanged customresourcedefinition apiextensions io listcheckers config istio io unchanged customresourcedefinition apiextensions io memquotas config istio io unchanged customresourcedefinition apiextensions io noops config istio io unchanged customresourcedefinition apiextensions io opas config istio io unchanged customresourcedefinition apiextensions io prometheuses config istio io unchanged customresourcedefinition apiextensions io rbacs config istio io unchanged customresourcedefinition apiextensions io redisquotas config istio io unchanged customresourcedefinition apiextensions io signalfxs config istio io unchanged customresourcedefinition apiextensions io solarwindses config istio io unchanged customresourcedefinition apiextensions io stackdrivers config istio io unchanged customresourcedefinition apiextensions io statsds config istio io unchanged customresourcedefinition apiextensions io stdios config istio io unchanged customresourcedefinition apiextensions io apikeys config istio io unchanged customresourcedefinition apiextensions io authorizations config istio io unchanged customresourcedefinition apiextensions io checknothings config istio io unchanged customresourcedefinition apiextensions io kuberneteses config istio io unchanged customresourcedefinition apiextensions io listentries config istio io unchanged customresourcedefinition apiextensions io logentries config istio io unchanged customresourcedefinition apiextensions io edges config istio io unchanged customresourcedefinition apiextensions io metrics config istio io unchanged customresourcedefinition apiextensions io quotas config istio io unchanged customresourcedefinition apiextensions io reportnothings config istio io unchanged customresourcedefinition apiextensions io tracespans config istio io unchanged customresourcedefinition apiextensions io rbacconfigs rbac istio io unchanged customresourcedefinition apiextensions io serviceroles rbac istio io unchanged customresourcedefinition apiextensions io servicerolebindings rbac istio io unchanged customresourcedefinition apiextensions io adapters config istio io unchanged customresourcedefinition apiextensions io instances config istio io unchanged customresourcedefinition apiextensions io templates config istio io unchanged customresourcedefinition apiextensions io handlers config istio io unchanged customresourcedefinition apiextensions io cloudwatches config istio io unchanged customresourcedefinition apiextensions io dogstatsds config istio io unchanged customresourcedefinition apiextensions io sidecars networking istio io unchanged customresourcedefinition apiextensions io zipkins config istio io unchanged customresourcedefinition apiextensions io clusterissuers certmanager io unchanged customresourcedefinition apiextensions io issuers certmanager io unchanged customresourcedefinition apiextensions io orders certmanager io unchanged customresourcedefinition apiextensions io challenges certmanager io unchanged secret kiali created configmap istio galley configuration created configmap istio grafana custom resources created configmap istio grafana configuration dashboards galley dashboard created configmap istio grafana configuration dashboards istio mesh dashboard created configmap istio grafana configuration dashboards istio performance dashboard created configmap istio grafana configuration dashboards istio service dashboard created configmap istio grafana configuration dashboards istio workload dashboard created configmap istio grafana configuration dashboards mixer dashboard created configmap istio grafana configuration dashboards pilot dashboard created configmap istio grafana created configmap kiali created configmap prometheus created configmap istio security custom resources created configmap istio created configmap istio sidecar injector created serviceaccount istio galley service account created serviceaccount istio egressgateway service account created serviceaccount istio ingressgateway service account created serviceaccount istio grafana post install account created clusterrole rbac authorization io istio grafana post install istio system created clusterrolebinding rbac authorization io istio grafana post install role binding istio system created job batch istio grafana post install created serviceaccount kiali service account created serviceaccount istio mixer service account created serviceaccount istio pilot service account created serviceaccount prometheus created serviceaccount istio cleanup secrets service account created clusterrole rbac authorization io istio cleanup secrets istio system created clusterrolebinding rbac authorization io istio cleanup secrets istio system created job batch istio cleanup secrets created serviceaccount istio security post install account created clusterrole rbac authorization io istio security post install istio system created clusterrolebinding rbac authorization io istio security post install role binding istio system created job batch istio security post install created serviceaccount istio citadel service account created serviceaccount istio sidecar injector service account created serviceaccount istio multi created clusterrole rbac authorization io istio galley istio system created clusterrole rbac authorization io istio egressgateway istio system created clusterrole rbac authorization io istio ingressgateway istio system created clusterrole rbac authorization io kiali created clusterrole rbac authorization io kiali viewer created clusterrole rbac authorization io istio mixer istio system created clusterrole rbac authorization io istio pilot istio system created clusterrole rbac authorization io prometheus istio system created clusterrole rbac authorization io istio citadel istio system created clusterrole rbac authorization io istio sidecar injector istio system created clusterrole rbac authorization io istio reader created clusterrolebinding rbac authorization io istio galley admin role binding istio system created clusterrolebinding rbac authorization io istio egressgateway istio system created clusterrolebinding rbac authorization io istio ingressgateway istio system created clusterrolebinding rbac authorization io istio kiali admin role binding istio system created clusterrolebinding rbac authorization io istio mixer admin role binding istio system created clusterrolebinding rbac authorization io istio pilot istio system created clusterrolebinding rbac authorization io prometheus istio system created clusterrolebinding rbac authorization io istio citadel istio system created clusterrolebinding rbac authorization io istio sidecar injector admin role binding istio system created clusterrolebinding rbac authorization io istio multi created role rbac authorization io istio ingressgateway sds created rolebinding rbac authorization io istio ingressgateway sds created service istio galley created service istio egressgateway created service istio ingressgateway created service grafana created service kiali created service istio policy created service istio telemetry created service istio pilot created service prometheus created service istio citadel created service istio sidecar injector created deployment extensions istio galley created deployment extensions istio egressgateway created deployment extensions istio ingressgateway created deployment extensions grafana created deployment extensions kiali created deployment extensions istio policy created deployment extensions istio telemetry created deployment extensions istio pilot created deployment extensions prometheus created deployment extensions istio citadel created deployment extensions istio sidecar injector created deployment extensions istio tracing created horizontalpodautoscaler autoscaling istio egressgateway created horizontalpodautoscaler autoscaling istio ingressgateway created horizontalpodautoscaler autoscaling istio policy created horizontalpodautoscaler autoscaling istio telemetry created horizontalpodautoscaler autoscaling istio pilot created service jaeger query created service jaeger collector created service jaeger agent created service zipkin created service tracing created mutatingwebhookconfiguration admissionregistration io istio sidecar injector created poddisruptionbudget policy istio galley created poddisruptionbudget policy istio egressgateway created poddisruptionbudget policy istio ingressgateway created poddisruptionbudget policy istio policy created poddisruptionbudget policy istio telemetry created poddisruptionbudget policy istio pilot created attributemanifest config istio io istioproxy created attributemanifest config istio io kubernetes created handler config istio io stdio created logentry config istio io accesslog created logentry config istio io tcpaccesslog created rule config istio io stdio created rule config istio io stdiotcp created metric config istio io requestcount created metric config istio io requestduration created metric config istio io requestsize created metric config istio io responsesize created metric config istio io tcpbytesent created metric config istio io tcpbytereceived created metric config istio io tcpconnectionsopened created metric config istio io tcpconnectionsclosed created handler config istio io prometheus created rule config istio io promhttp created rule config istio io promtcp created rule config istio io promtcpconnectionopen created rule config istio io promtcpconnectionclosed created handler config istio io kubernetesenv created rule config istio io kubeattrgenrulerule created rule config istio io tcpkubeattrgenrulerule created kubernetes config istio io attributes created destinationrule networking istio io istio policy created destinationrule networking istio io istio telemetry created now using project kabanero on server you can add applications to this project with the new app command for example try oc new app django psql example to build a new example application in python or use kubectl to deploy a simple kubernetes application kubectl create deployment hello node image gcr io hello minikube zero install hello node namespace kabanero configured deployment apps knative eventing operator created role rbac authorization io knative eventing operator created clusterrole rbac authorization io knative eventing operator created rolebinding rbac authorization io knative eventing operator created clusterrolebinding rbac authorization io knative eventing operator created serviceaccount knative eventing operator created customresourcedefinition apiextensions io knativeeventings eventing knative dev created deployment apps knative serving operator created role rbac authorization io knative serving operator created clusterrole rbac authorization io knative serving operator created rolebinding rbac authorization io knative serving operator created clusterrolebinding rbac authorization io knative serving operator created serviceaccount knative serving operator created customresourcedefinition apiextensions io knativeservings serving knative dev created namespace knative sources created serviceaccount controller manager created clusterrole rbac authorization io eventing sources controller created clusterrolebinding rbac authorization io eventing sources controller created clusterrolebinding rbac authorization io eventing sources addressable resolver created customresourcedefinition apiextensions io githubsources sources eventing knative dev created service controller created statefulset apps controller manager created deployment apps openshift pipelines operator created clusterrole rbac authorization io openshift pipelines operator created clusterrolebinding rbac authorization io openshift pipelines operator created serviceaccount openshift pipelines operator created namespace openshift pipelines created customresourcedefinition apiextensions io config operator tekton dev created clusterrole rbac authorization io knative serving operator configured clusterrole rbac authorization io knative eventing operator configured customresourcedefinition apiextensions io kabaneros kabanero io created customresourcedefinition apiextensions io collections kabanero io created serviceaccount kabanero operator created deployment apps kabanero operator created clusterrole rbac authorization io kabanero operator created clusterrolebinding rbac authorization io kabanero operator created namespace kabanero unchanged deployment apps knative eventing operator configured role rbac authorization io knative eventing operator configured clusterrole rbac authorization io knative eventing operator configured rolebinding rbac authorization io knative eventing operator unchanged clusterrolebinding rbac authorization io knative eventing operator unchanged serviceaccount knative eventing operator unchanged customresourcedefinition apiextensions io knativeeventings eventing knative dev unchanged deployment apps knative serving operator configured role rbac authorization io knative serving operator configured clusterrole rbac authorization io knative serving operator configured rolebinding rbac authorization io knative serving operator unchanged clusterrolebinding rbac authorization io knative serving operator unchanged serviceaccount knative serving operator unchanged customresourcedefinition apiextensions io knativeservings serving knative dev unchanged namespace knative sources unchanged serviceaccount controller manager unchanged clusterrole rbac authorization io eventing sources controller unchanged clusterrolebinding rbac authorization io eventing sources controller unchanged clusterrolebinding rbac authorization io eventing sources addressable resolver unchanged customresourcedefinition apiextensions io githubsources sources eventing knative dev unchanged service controller unchanged statefulset apps controller manager configured deployment apps openshift pipelines operator configured clusterrole rbac authorization io openshift pipelines operator configured clusterrolebinding rbac authorization io openshift pipelines operator unchanged serviceaccount openshift pipelines operator unchanged namespace openshift pipelines unchanged customresourcedefinition apiextensions io config operator tekton dev unchanged config operator tekton dev cluster created clusterrole rbac authorization io knative serving operator configured clusterrole rbac authorization io knative eventing operator configured customresourcedefinition apiextensions io kabaneros kabanero io configured customresourcedefinition apiextensions io collections kabanero io configured serviceaccount kabanero operator unchanged deployment apps kabanero operator configured clusterrole rbac authorization io kabanero operator configured clusterrolebinding rbac authorization io kabanero operator unchanged clusterrole rbac authorization io cluster admin added kabanero operator name created at services serving knative dev name created at clustertasks tekton dev config operator tekton dev pipelineresources tekton dev pipelineruns tekton dev pipelines tekton dev taskruns tekton dev tasks tekton dev serviceaccount tekton webhooks extension created clusterrole rbac authorization io tekton webhooks extension minimal created clusterrolebinding rbac authorization io tekton webhooks extension minimal created deployment apps webhooks extension created service webhooks extension created service serving knative dev webhooks extension sink created task tekton dev monitor result task created serviceaccount tekton dashboard created customresourcedefinition apiextensions io extensions dashboard tekton dev created clusterrole rbac authorization io tekton dashboard minimal created clusterrolebinding rbac authorization io tekton dashboard minimal created deployment apps tekton dashboard created route route openshift io tekton dashboard created service tekton dashboard created pipeline tekton dev created task tekton dev task created name age tekton dashboard minimal clusterrole rbac authorization io tekton dashboard minimal patched deployment extensions tekton dashboard scaled deployment extensions tekton dashboard scaled configmap config domain patched name created at extensions dashboard tekton dev the installation script is complete you can now create an instance of the kabanero cr if you have cloned and curated a collection set apply the kabanero cr that you created or to create the default instance oc apply n kabanero f task thursday october ok msg task thursday october skipping task thursday october changed task thursday october ok msg statefulset apps alertmanager main scaled statefulset apps prometheus scaled deployment extensions cluster monitoring operator scaled deployment extensions grafana scaled deployment extensions kube state metrics scaled deployment extensions openshift state metrics scaled deployment extensions prometheus adapter scaled deployment extensions prometheus operator scaled deployment extensions telemeter client scaled task thursday october changed task thursday october ok msg client version server version kubernetes version play recap svtcrc fyre ibm com ok changed unreachable failed skipped rescued ignored thursday october install kabanero foundation crc start install networkmanager firefox crc setup retrieve kabanero foundation master zip gathering facts prereq directories oc login enable monitoring alerting and telemetry services enable and start networkmanager crc pull secret unzip kabanerio foundation master zip oc version check if crc is installed whoami oc in user path password masking process crc creds check if crc is created link oc kubectl set admin passwd install kabanero foundation stdout set fact crc setup stdout crc start stdout crc creds stdout set fact oc login stdout enable monitoring alerting and telemetry services stdout set fact fail crc delete crcstop stdout install kabanero foundation errors install kabanero foundation return code crc stop set fact set fact set fact crc delete stdout whoami stdout oc version stdout kabanero crc complete runtime minutes setup complete overall runtime minutes
| 1
|
270,534
| 23,516,313,901
|
IssuesEvent
|
2022-08-18 21:50:01
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
MonoTests.System.Runtime.Caching.MemoryCacheTestExpires4.TestCacheShrink test failure
|
area-System.Runtime.Caching test-bug
|
Configuration: `net6.0-windows-Release-arm64-CoreCLR_release-Windows.10.Arm64.Open`
```
MonoTests.System.Runtime.Caching.MemoryCacheTestExpires4.TestCacheShrink [FAIL]
Assert.Null() Failure
Expected: (null)
Actual: 0
Stack Trace:
/_/src/libraries/System.Runtime.Caching/tests/System.Runtime.Caching/MemoryCacheTest.cs(0,0): at MonoTests.System.Runtime.Caching.MemoryCacheTestExpires4.TestCacheShrink()
--- End of stack trace from previous location ---
```
Console: https://helixre8s23ayyeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-f2e3e082ea3b44a49c/System.Runtime.Caching.Tests/1/console.f9d74dab.log?sv=2019-07-07&se=2021-09-10T12%3A50%3A10Z&sr=c&sp=rl&sig=yohDg5MjQ17E738GWrsKtd5oBkc5WxKT8vvsgpAsogE%3D
Build: https://dev.azure.com/dnceng/public/_build/results?buildId=1309902&view=ms.vss-test-web.build-test-results-tab&runId=38562306&paneView=debug
|
1.0
|
MonoTests.System.Runtime.Caching.MemoryCacheTestExpires4.TestCacheShrink test failure - Configuration: `net6.0-windows-Release-arm64-CoreCLR_release-Windows.10.Arm64.Open`
```
MonoTests.System.Runtime.Caching.MemoryCacheTestExpires4.TestCacheShrink [FAIL]
Assert.Null() Failure
Expected: (null)
Actual: 0
Stack Trace:
/_/src/libraries/System.Runtime.Caching/tests/System.Runtime.Caching/MemoryCacheTest.cs(0,0): at MonoTests.System.Runtime.Caching.MemoryCacheTestExpires4.TestCacheShrink()
--- End of stack trace from previous location ---
```
Console: https://helixre8s23ayyeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-f2e3e082ea3b44a49c/System.Runtime.Caching.Tests/1/console.f9d74dab.log?sv=2019-07-07&se=2021-09-10T12%3A50%3A10Z&sr=c&sp=rl&sig=yohDg5MjQ17E738GWrsKtd5oBkc5WxKT8vvsgpAsogE%3D
Build: https://dev.azure.com/dnceng/public/_build/results?buildId=1309902&view=ms.vss-test-web.build-test-results-tab&runId=38562306&paneView=debug
|
test
|
monotests system runtime caching testcacheshrink test failure configuration windows release coreclr release windows open monotests system runtime caching testcacheshrink assert null failure expected null actual stack trace src libraries system runtime caching tests system runtime caching memorycachetest cs at monotests system runtime caching testcacheshrink end of stack trace from previous location console build
| 1
|
809,037
| 30,121,835,043
|
IssuesEvent
|
2023-06-30 15:46:56
|
conaticus/FileExplorer
|
https://api.github.com/repos/conaticus/FileExplorer
|
opened
|
Recaching interval not long enough on slower systems
|
bug medium priority
|
The interval works well on most systems, but slower systems or systems with a bigger file system have the chance of not being able to recache in time for this 30s interval.
The best solution is to have a check to ensure that the previous recache is completed, and skip if it is not. We could implement an algorithm to check for file changes first, but these usually happen so often that it would be redundant.
|
1.0
|
Recaching interval not long enough on slower systems - The interval works well on most systems, but slower systems or systems with a bigger file system have the chance of not being able to recache in time for this 30s interval.
The best solution is to have a check to ensure that the previous recache is completed, and skip if it is not. We could implement an algorithm to check for file changes first, but these usually happen so often that it would be redundant.
|
non_test
|
recaching interval not long enough on slower systems the interval works well on most systems but slower systems or systems with a bigger file system have the chance of not being able to recache in time for this interval the best solution is to have a check to ensure that the previous recache is completed and skip if it is not we could implement an algorithm to check for file changes first but these usually happen so often that it would be redundant
| 0
|
119,786
| 10,064,413,621
|
IssuesEvent
|
2019-07-23 08:37:02
|
microsoft/AzureStorageExplorer
|
https://api.github.com/repos/microsoft/AzureStorageExplorer
|
closed
|
The created new table doesn't show on the tree view for one Table Account
|
:gear: cosmosdb 🧪 testing
|
**Storage Explorer Version:** 1.9.0_**master**_20190709.5
**Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10
**Architecture:** ia32/x64
**Commit:** 8a04af7f
**Regression From:** Not a regression
**Steps to reproduce:**
1. Create a new **Table** account then navigate to it from Storage Explorer.
2. Right click the create Table account then select 'Create Table...'.
3. Type a valid table id then click 'OK'.
4. Check whether the created table shows well or not on the tree view.
**Expect Experience:**
The created table shows well on the tree view.
**Actual Experience:**
The created table doesn't show on the tree view.
**More info:**
The created table shows well after refreshing the Table account node.
|
1.0
|
The created new table doesn't show on the tree view for one Table Account - **Storage Explorer Version:** 1.9.0_**master**_20190709.5
**Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10
**Architecture:** ia32/x64
**Commit:** 8a04af7f
**Regression From:** Not a regression
**Steps to reproduce:**
1. Create a new **Table** account then navigate to it from Storage Explorer.
2. Right click the create Table account then select 'Create Table...'.
3. Type a valid table id then click 'OK'.
4. Check whether the created table shows well or not on the tree view.
**Expect Experience:**
The created table shows well on the tree view.
**Actual Experience:**
The created table doesn't show on the tree view.
**More info:**
The created table shows well after refreshing the Table account node.
|
test
|
the created new table doesn t show on the tree view for one table account storage explorer version master platform os linux ubuntu macos high sierra windows architecture commit regression from not a regression steps to reproduce create a new table account then navigate to it from storage explorer right click the create table account then select create table type a valid table id then click ok check whether the created table shows well or not on the tree view expect experience the created table shows well on the tree view actual experience the created table doesn t show on the tree view more info the created table shows well after refreshing the table account node
| 1
|
49,003
| 5,996,795,874
|
IssuesEvent
|
2017-06-03 17:33:57
|
ppazos/cabolabs-ehrserver
|
https://api.github.com/repos/ppazos/cabolabs-ehrserver
|
opened
|
Test spring security protection against brute force attacks on the login
|
mid_prio test
|
1. check if there is any built-in solution, like IP blocking.
2. check if invalid logins for a valid username generate DOS for that user.
3. check the two step login like the new google login to avoid attacks (analyze pros/cons and if that applies for this context).
|
1.0
|
Test spring security protection against brute force attacks on the login - 1. check if there is any built-in solution, like IP blocking.
2. check if invalid logins for a valid username generate DOS for that user.
3. check the two step login like the new google login to avoid attacks (analyze pros/cons and if that applies for this context).
|
test
|
test spring security protection against brute force attacks on the login check if there is any built in solution like ip blocking check if invalid logins for a valid username generate dos for that user check the two step login like the new google login to avoid attacks analyze pros cons and if that applies for this context
| 1
|
389
| 2,535,799,006
|
IssuesEvent
|
2015-01-26 07:46:45
|
localwiki/localwiki
|
https://api.github.com/repos/localwiki/localwiki
|
closed
|
First-time usability confusion points
|
defect general
|
Double click to end path when drawing -- was confusing. Wasn't obvious.
-----
Deleting things in the map -- delete button on kb should delete an area or line or point. deleting is just kinda weird.
------
Undo button in editor?
|
1.0
|
First-time usability confusion points - Double click to end path when drawing -- was confusing. Wasn't obvious.
-----
Deleting things in the map -- delete button on kb should delete an area or line or point. deleting is just kinda weird.
------
Undo button in editor?
|
non_test
|
first time usability confusion points double click to end path when drawing was confusing wasn t obvious deleting things in the map delete button on kb should delete an area or line or point deleting is just kinda weird undo button in editor
| 0
|
319,222
| 9,740,469,361
|
IssuesEvent
|
2019-06-01 20:48:54
|
Terrastories/terrastories
|
https://api.github.com/repos/Terrastories/terrastories
|
closed
|
[Video] Opening a video in fullscreen crashes the display
|
difficulty: intermediate priority: medium status: help wanted type: bug type: video
|
To reproduce:
1. Find a story with a video
2. Play the video
3. Click the fullscreen icon on the video player
4. Notice that the fullscreen display crashes after ~half a second
A good amount of discussion took place on this subject at Ruby by the Bay 2019: https://github.com/rubyforgood/terrastories/issues/24
|
1.0
|
[Video] Opening a video in fullscreen crashes the display - To reproduce:
1. Find a story with a video
2. Play the video
3. Click the fullscreen icon on the video player
4. Notice that the fullscreen display crashes after ~half a second
A good amount of discussion took place on this subject at Ruby by the Bay 2019: https://github.com/rubyforgood/terrastories/issues/24
|
non_test
|
opening a video in fullscreen crashes the display to reproduce find a story with a video play the video click the fullscreen icon on the video player notice that the fullscreen display crashes after half a second a good amount of discussion took place on this subject at ruby by the bay
| 0
|
156,210
| 12,301,091,581
|
IssuesEvent
|
2020-05-11 14:55:13
|
jorandmol/DP2-G7
|
https://api.github.com/repos/jorandmol/DP2-G7
|
closed
|
T18 - Actualizar Medicine
|
develop unit testing
|
Como auxiliar de clínica quiero poder editar los datos de un medicamento para solucionar posibles errores o actualizar dicho objeto.
|
1.0
|
T18 - Actualizar Medicine - Como auxiliar de clínica quiero poder editar los datos de un medicamento para solucionar posibles errores o actualizar dicho objeto.
|
test
|
actualizar medicine como auxiliar de clínica quiero poder editar los datos de un medicamento para solucionar posibles errores o actualizar dicho objeto
| 1
|
275,261
| 23,901,232,792
|
IssuesEvent
|
2022-09-08 18:59:01
|
PharmaLedger-IMI/eco-iot-pmed-workspace
|
https://api.github.com/repos/PharmaLedger-IMI/eco-iot-pmed-workspace
|
closed
|
[Clinical Site] Check the Notification "view questions"
|
business-testing
|
Questionnaire Notification LINK “view questions” to HCP does not show the results of the questions
|
1.0
|
[Clinical Site] Check the Notification "view questions" - Questionnaire Notification LINK “view questions” to HCP does not show the results of the questions
|
test
|
check the notification view questions questionnaire notification link “view questions” to hcp does not show the results of the questions
| 1
|
106,207
| 23,193,281,440
|
IssuesEvent
|
2022-08-01 14:20:06
|
open-contracting/standard
|
https://api.github.com/repos/open-contracting/standard
|
closed
|
Fix markup to avoid unwanted links
|
Schema Codelist: Open quick
|
linkify automatically hyperlinks URLs. We need to:
* Add backticks around release.id in the schema
* Remove the sentence about buyandsell.gc.ca from itemClassificationScheme.csv (no other code has a similar sentence)
Once that's done, we can remove the lines in `conf.py` that ignore the auto-created links.
|
1.0
|
Fix markup to avoid unwanted links - linkify automatically hyperlinks URLs. We need to:
* Add backticks around release.id in the schema
* Remove the sentence about buyandsell.gc.ca from itemClassificationScheme.csv (no other code has a similar sentence)
Once that's done, we can remove the lines in `conf.py` that ignore the auto-created links.
|
non_test
|
fix markup to avoid unwanted links linkify automatically hyperlinks urls we need to add backticks around release id in the schema remove the sentence about buyandsell gc ca from itemclassificationscheme csv no other code has a similar sentence once that s done we can remove the lines in conf py that ignore the auto created links
| 0
|
405,773
| 27,533,225,863
|
IssuesEvent
|
2023-03-07 00:18:09
|
sveltejs/kit
|
https://api.github.com/repos/sveltejs/kit
|
closed
|
Prerender tries to render outgoing link
|
documentation p3-edge-case
|
### Describe the bug
First off, sorry if this is the wrong place, I am fairly new to all the things going on here.
I have been trying to create a blog which uses SvelteKit, mdsvex and a static adapter.
Everything works smooth until I start adding links inside the markdown files.
If any .md file has any outgoing link, it seems the prerender tries to find those sites on my site.
This causes it to fail ofcourse.
The issue can be resolved by setting `onerror` for the prerenderer into `continue`
but I would like to avoid that since now any missing posts will also be continuned.
### Reproduction
Clone the repository and try running the build command
```bash
git clone git@github.com:percybolmer/pp-blog-svelte.git
cd pp-blog-svelte
git checkout bug/prerender-outgoinglink
npm install
npm run build
```
### Logs
```Shell
percy@pc038:~/private/blog/pp-blog-svelte$ npm run build
> build
> vite build
vite v3.1.8 building for production...
✓ 75 modules transformed.
vite v3.1.8 building SSR bundle for production...
✓ 81 modules transformed.
Generated an empty chunk: "hooks"
.svelte-kit/output/server/vite-manifest.json 4.99 KiB
.svelte-kit/output/server/index.js 69.60 KiB
.svelte-kit/output/server/entries/endpoints/rss/_server.js 1.42 KiB
.svelte-kit/output/server/entries/endpoints/api/posts/_server.js 0.43 KiB
.svelte-kit/output/server/entries/pages/_layout.svelte.js 1.58 KiB
.svelte-kit/output/server/entries/pages/_layout.js 0.15 KiB
.svelte-kit/output/server/entries/fallbacks/error.svelte.js 1.50 KiB
.svelte-kit/output/server/entries/pages/_page.svelte.js 0.65 KiB
.svelte-kit/output/server/entries/pages/about/_page.svelte.js 0.38 KiB
.svelte-kit/output/server/entries/pages/blog/_page.svelte.js 2.57 KiB
.svelte-kit/output/server/entries/pages/blog/_page.js 0.16 KiB
.svelte-kit/output/server/entries/pages/blog/_slug_/_page.svelte.js 0.81 KiB
.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js 0.87 KiB
.svelte-kit/output/server/entries/pages/contact/_page.svelte.js 0.38 KiB
.svelte-kit/output/server/entries/pages/uses/_page.md.js 0.45 KiB
.svelte-kit/output/server/chunks/paths.js 0.17 KiB
.svelte-kit/output/server/chunks/index.js 3.53 KiB
.svelte-kit/output/server/chunks/index2.js 0.97 KiB
.svelte-kit/output/server/chunks/index3.js 0.79 KiB
.svelte-kit/output/server/chunks/hooks.js 0.00 KiB
.svelte-kit/output/server/chunks/2.js 0.91 KiB
.svelte-kit/output/server/chunks/thumbnail.js 0.12 KiB
.svelte-kit/output/server/chunks/preview-ide-of-the-future.js 23.67 KiB
.svelte-kit/output/server/chunks/test.js 0.54 KiB
Error: Unknown variable dynamic import: ../posts/www.jetbrains.com.md
at file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js:7:92
at new Promise (<anonymous>)
at __variableDynamicImportRuntimeHelper (file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js:6:10)
at load (file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js:11:22)
at load_data (file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/index.js:626:39)
at async file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/index.js:1416:18
500 /pp-blog-svelte/blog/www.jetbrains.com (linked from /pp-blog-svelte/blog/preview-ide-of-the-future)
```
### System Info
```Shell
System:
OS: Linux 5.15 Ubuntu 20.04.5 LTS (Focal Fossa)
CPU: (16) x64 11th Gen Intel(R) Core(TM) i7-11800H @ 2.30GHz
Memory: 1.99 GB / 15.36 GB
Container: Yes
Shell: 5.0.17 - /bin/bash
Binaries:
Node: 17.9.1 - ~/.nvm/versions/node/v17.9.1/bin/node
Yarn: 1.22.19 - ~/.nvm/versions/node/v14.17.6/bin/yarn
npm: 8.11.0 - ~/.nvm/versions/node/v17.9.1/bin/npm
Browsers:
Brave Browser: 106.1.44.101
Chrome: 106.0.5249.61
Firefox: 105.0
npmPackages:
@sveltejs/adapter-static: ^1.0.0-next.46 => 1.0.0-next.46
@sveltejs/kit: 1.0.0-next.516 => 1.0.0-next.516
svelte: ^3.52.0 => 3.52.0
vite: ^3.1.8 => 3.1.8
```
### Severity
annoyance
### Additional Information
Hopefully it is not just me being dumb!
|
1.0
|
Prerender tries to render outgoing link - ### Describe the bug
First off, sorry if this is the wrong place, I am fairly new to all the things going on here.
I have been trying to create a blog which uses SvelteKit, mdsvex and a static adapter.
Everything works smooth until I start adding links inside the markdown files.
If any .md file has any outgoing link, it seems the prerender tries to find those sites on my site.
This causes it to fail ofcourse.
The issue can be resolved by setting `onerror` for the prerenderer into `continue`
but I would like to avoid that since now any missing posts will also be continuned.
### Reproduction
Clone the repository and try running the build command
```bash
git clone git@github.com:percybolmer/pp-blog-svelte.git
cd pp-blog-svelte
git checkout bug/prerender-outgoinglink
npm install
npm run build
```
### Logs
```Shell
percy@pc038:~/private/blog/pp-blog-svelte$ npm run build
> build
> vite build
vite v3.1.8 building for production...
✓ 75 modules transformed.
vite v3.1.8 building SSR bundle for production...
✓ 81 modules transformed.
Generated an empty chunk: "hooks"
.svelte-kit/output/server/vite-manifest.json 4.99 KiB
.svelte-kit/output/server/index.js 69.60 KiB
.svelte-kit/output/server/entries/endpoints/rss/_server.js 1.42 KiB
.svelte-kit/output/server/entries/endpoints/api/posts/_server.js 0.43 KiB
.svelte-kit/output/server/entries/pages/_layout.svelte.js 1.58 KiB
.svelte-kit/output/server/entries/pages/_layout.js 0.15 KiB
.svelte-kit/output/server/entries/fallbacks/error.svelte.js 1.50 KiB
.svelte-kit/output/server/entries/pages/_page.svelte.js 0.65 KiB
.svelte-kit/output/server/entries/pages/about/_page.svelte.js 0.38 KiB
.svelte-kit/output/server/entries/pages/blog/_page.svelte.js 2.57 KiB
.svelte-kit/output/server/entries/pages/blog/_page.js 0.16 KiB
.svelte-kit/output/server/entries/pages/blog/_slug_/_page.svelte.js 0.81 KiB
.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js 0.87 KiB
.svelte-kit/output/server/entries/pages/contact/_page.svelte.js 0.38 KiB
.svelte-kit/output/server/entries/pages/uses/_page.md.js 0.45 KiB
.svelte-kit/output/server/chunks/paths.js 0.17 KiB
.svelte-kit/output/server/chunks/index.js 3.53 KiB
.svelte-kit/output/server/chunks/index2.js 0.97 KiB
.svelte-kit/output/server/chunks/index3.js 0.79 KiB
.svelte-kit/output/server/chunks/hooks.js 0.00 KiB
.svelte-kit/output/server/chunks/2.js 0.91 KiB
.svelte-kit/output/server/chunks/thumbnail.js 0.12 KiB
.svelte-kit/output/server/chunks/preview-ide-of-the-future.js 23.67 KiB
.svelte-kit/output/server/chunks/test.js 0.54 KiB
Error: Unknown variable dynamic import: ../posts/www.jetbrains.com.md
at file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js:7:92
at new Promise (<anonymous>)
at __variableDynamicImportRuntimeHelper (file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js:6:10)
at load (file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/entries/pages/blog/_slug_/_page.js:11:22)
at load_data (file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/index.js:626:39)
at async file:///home/percy/private/blog/pp-blog-svelte/.svelte-kit/output/server/index.js:1416:18
500 /pp-blog-svelte/blog/www.jetbrains.com (linked from /pp-blog-svelte/blog/preview-ide-of-the-future)
```
### System Info
```Shell
System:
OS: Linux 5.15 Ubuntu 20.04.5 LTS (Focal Fossa)
CPU: (16) x64 11th Gen Intel(R) Core(TM) i7-11800H @ 2.30GHz
Memory: 1.99 GB / 15.36 GB
Container: Yes
Shell: 5.0.17 - /bin/bash
Binaries:
Node: 17.9.1 - ~/.nvm/versions/node/v17.9.1/bin/node
Yarn: 1.22.19 - ~/.nvm/versions/node/v14.17.6/bin/yarn
npm: 8.11.0 - ~/.nvm/versions/node/v17.9.1/bin/npm
Browsers:
Brave Browser: 106.1.44.101
Chrome: 106.0.5249.61
Firefox: 105.0
npmPackages:
@sveltejs/adapter-static: ^1.0.0-next.46 => 1.0.0-next.46
@sveltejs/kit: 1.0.0-next.516 => 1.0.0-next.516
svelte: ^3.52.0 => 3.52.0
vite: ^3.1.8 => 3.1.8
```
### Severity
annoyance
### Additional Information
Hopefully it is not just me being dumb!
|
non_test
|
prerender tries to render outgoing link describe the bug first off sorry if this is the wrong place i am fairly new to all the things going on here i have been trying to create a blog which uses sveltekit mdsvex and a static adapter everything works smooth until i start adding links inside the markdown files if any md file has any outgoing link it seems the prerender tries to find those sites on my site this causes it to fail ofcourse the issue can be resolved by setting onerror for the prerenderer into continue but i would like to avoid that since now any missing posts will also be continuned reproduction clone the repository and try running the build command bash git clone git github com percybolmer pp blog svelte git cd pp blog svelte git checkout bug prerender outgoinglink npm install npm run build logs shell percy private blog pp blog svelte npm run build build vite build vite building for production ✓ modules transformed vite building ssr bundle for production ✓ modules transformed generated an empty chunk hooks svelte kit output server vite manifest json kib svelte kit output server index js kib svelte kit output server entries endpoints rss server js kib svelte kit output server entries endpoints api posts server js kib svelte kit output server entries pages layout svelte js kib svelte kit output server entries pages layout js kib svelte kit output server entries fallbacks error svelte js kib svelte kit output server entries pages page svelte js kib svelte kit output server entries pages about page svelte js kib svelte kit output server entries pages blog page svelte js kib svelte kit output server entries pages blog page js kib svelte kit output server entries pages blog slug page svelte js kib svelte kit output server entries pages blog slug page js kib svelte kit output server entries pages contact page svelte js kib svelte kit output server entries pages uses page md js kib svelte kit output server chunks paths js kib svelte kit output server chunks index js kib svelte kit output server chunks js kib svelte kit output server chunks js kib svelte kit output server chunks hooks js kib svelte kit output server chunks js kib svelte kit output server chunks thumbnail js kib svelte kit output server chunks preview ide of the future js kib svelte kit output server chunks test js kib error unknown variable dynamic import posts at file home percy private blog pp blog svelte svelte kit output server entries pages blog slug page js at new promise at variabledynamicimportruntimehelper file home percy private blog pp blog svelte svelte kit output server entries pages blog slug page js at load file home percy private blog pp blog svelte svelte kit output server entries pages blog slug page js at load data file home percy private blog pp blog svelte svelte kit output server index js at async file home percy private blog pp blog svelte svelte kit output server index js pp blog svelte blog linked from pp blog svelte blog preview ide of the future system info shell system os linux ubuntu lts focal fossa cpu gen intel r core tm memory gb gb container yes shell bin bash binaries node nvm versions node bin node yarn nvm versions node bin yarn npm nvm versions node bin npm browsers brave browser chrome firefox npmpackages sveltejs adapter static next next sveltejs kit next next svelte vite severity annoyance additional information hopefully it is not just me being dumb
| 0
|
59,032
| 14,524,957,564
|
IssuesEvent
|
2020-12-14 12:12:41
|
neovim/neovim
|
https://api.github.com/repos/neovim/neovim
|
closed
|
`brew install --HEAD` failing with errors
|
bug build duplicate platform:macos
|
<!-- Before reporting: search existing issues and check the FAQ. -->
- `nvim --version`: NVIM v0.5.0-dev+944-g82100a6bd
- `vim -u DEFAULTS` (version: ) behaves differently? no
- Operating system/version: macOS 11.0.1 (on MacBook Air with Apple Silicon M1 SoC)
- Terminal name/version: iTerm2
- `$TERM`: xterm-256color
### Actual behaviour
Note: I've already found #8809 and followed the build-from-source prerequisite instructions.
```
❯ brew install --HEAD neovim
Updating Homebrew...
==> Downloading https://github.com/libmpack/libmpack-lua/releases/download/1.0.8
Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/4b91349796a66e38e93acd77e6ca14212c869090429fd88271def2be5293d83f--libmpack-lua-1.0.8.tar.gz
==> Downloading https://luarocks.org/manifests/gvvaughan/lpeg-1.0.2-1.src.rock
Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/0bc17975f0cf3981083172aa7da329b874e5780fa81494f5dd07d0665fcec23b--lpeg-1.0.2-1.src.rock
==> Downloading https://luarocks.org/manifests/kikito/inspect-3.1.1-0.src.rock
Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/ecc1fa84ce4c83c9406f6d1202d4a88c80ede4a0a2bfd8d8d2bd54fd042226bb--inspect-3.1.1-0.src.rock
==> Downloading https://github.com/keplerproject/lua-compat-5.3/archive/v0.7.tar
Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/5f4e552e95b9a6223a5e4fabb7f04bf4b886057024507c76b1d5b803218c9363--lua-compat-5.3-0.7.tar.gz
==> Downloading https://github.com/luvit/luv/releases/download/1.30.0-0/luv-1.30
Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/a9ce5a949370965d8f0fe70a47e7d102a3b4b6b831dbd4d088a83bbe1c00a3a2--luv-1.30.0-0.tar.gz
==> Cloning https://github.com/neovim/neovim.git
Updating /Users/mpcsh/Library/Caches/Homebrew/neovim--git
==> Checking out branch master
Already on 'master'
Your branch is up to date with 'origin/master'.
HEAD is now at 82100a6bd healthcheck: fix health check issue with shada file (#13291)
==> luarocks make --lua-dir=/usr/local/opt/luajit --tree=/private/tmp/neovim-202
==> luarocks make --lua-dir=/usr/local/opt/luajit --tree=/private/tmp/neovim-202
==> luarocks make --lua-dir=/usr/local/opt/luajit --tree=/private/tmp/neovim-202
==> cmake . -DCMAKE_C_COMPILER=/usr/bin/clang -DCMAKE_CXX_COMPILER=/usr/bin/clan
==> make install
==> cmake .. -DCMAKE_C_COMPILER=/usr/bin/clang -DCMAKE_CXX_COMPILER=/usr/bin/cla
==> make install
Last 15 lines from /Users/mpcsh/Library/Logs/Homebrew/neovim/07.make:
^
/tmp/neovim-20201214-42949-1h90lp3/src/nvim/os/lang.c:63:15: error: use of undeclared identifier 'LC_ALL'; did you mean 'P_ALL'?
setlocale(LC_ALL, "");
^~~~~~
P_ALL
/Library/Developer/CommandLineTools/SDKs/MacOSX11.0.sdk/usr/include/sys/wait.h:80:2: note: 'P_ALL' declared here
P_ALL,
^
2 errors generated.
[ 50%] Building C object src/nvim/CMakeFiles/nvim.dir/os/time.c.o
cd /tmp/neovim-20201214-42949-1h90lp3/build/src/nvim && /usr/bin/clang -DINCLUDE_GENERATED_DECLARATIONS -DMIN_LOG_LEVEL=3 -DNVIM_MSGPACK_HAS_FLOAT32 -D_GNU_SOURCE -Dnvim_EXPORTS -I/tmp/neovim-20201214-42949-1h90lp3/build/config -I/tmp/neovim-20201214-42949-1h90lp3/src -I/tmp/neovim-20201214-42949-1h90lp3/build/src/nvim/auto -I/tmp/neovim-20201214-42949-1h90lp3/build/include -I/usr/local/include/luajit-2.0 -isystem /usr/local/include -isystem /tmp/neovim-20201214-42949-1h90lp3/deps-build/include -isystem /usr/local/opt/gettext/include -U_FORTIFY_SOURCE -D_FORTIFY_SOURCE=1 -DNDEBUG -flto=thin -arch arm64 -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX11.0.sdk -Wall -Wextra -pedantic -Wno-unused-parameter -Wstrict-prototypes -std=gnu99 -Wshadow -Wconversion -Wmissing-prototypes -Wimplicit-fallthrough -Wvla -fstack-protector-strong -fno-common -fdiagnostics-color=auto -o CMakeFiles/nvim.dir/os/time.c.o -c /tmp/neovim-20201214-42949-1h90lp3/src/nvim/os/time.c
make[2]: *** [src/nvim/CMakeFiles/nvim.dir/os/lang.c.o] Error 1
make[2]: *** Waiting for unfinished jobs....
make[1]: *** [src/nvim/CMakeFiles/nvim.dir/all] Error 2
make: *** [all] Error 2
READ THIS: https://docs.brew.sh/Troubleshooting
Please create pull requests instead of asking for help on Homebrew's GitHub,
Twitter or any other official channels.
```
### Expected behaviour
`brew install --HEAD neovim` works
|
1.0
|
`brew install --HEAD` failing with errors - <!-- Before reporting: search existing issues and check the FAQ. -->
- `nvim --version`: NVIM v0.5.0-dev+944-g82100a6bd
- `vim -u DEFAULTS` (version: ) behaves differently? no
- Operating system/version: macOS 11.0.1 (on MacBook Air with Apple Silicon M1 SoC)
- Terminal name/version: iTerm2
- `$TERM`: xterm-256color
### Actual behaviour
Note: I've already found #8809 and followed the build-from-source prerequisite instructions.
```
❯ brew install --HEAD neovim
Updating Homebrew...
==> Downloading https://github.com/libmpack/libmpack-lua/releases/download/1.0.8
Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/4b91349796a66e38e93acd77e6ca14212c869090429fd88271def2be5293d83f--libmpack-lua-1.0.8.tar.gz
==> Downloading https://luarocks.org/manifests/gvvaughan/lpeg-1.0.2-1.src.rock
Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/0bc17975f0cf3981083172aa7da329b874e5780fa81494f5dd07d0665fcec23b--lpeg-1.0.2-1.src.rock
==> Downloading https://luarocks.org/manifests/kikito/inspect-3.1.1-0.src.rock
Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/ecc1fa84ce4c83c9406f6d1202d4a88c80ede4a0a2bfd8d8d2bd54fd042226bb--inspect-3.1.1-0.src.rock
==> Downloading https://github.com/keplerproject/lua-compat-5.3/archive/v0.7.tar
Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/5f4e552e95b9a6223a5e4fabb7f04bf4b886057024507c76b1d5b803218c9363--lua-compat-5.3-0.7.tar.gz
==> Downloading https://github.com/luvit/luv/releases/download/1.30.0-0/luv-1.30
Already downloaded: /Users/mpcsh/Library/Caches/Homebrew/downloads/a9ce5a949370965d8f0fe70a47e7d102a3b4b6b831dbd4d088a83bbe1c00a3a2--luv-1.30.0-0.tar.gz
==> Cloning https://github.com/neovim/neovim.git
Updating /Users/mpcsh/Library/Caches/Homebrew/neovim--git
==> Checking out branch master
Already on 'master'
Your branch is up to date with 'origin/master'.
HEAD is now at 82100a6bd healthcheck: fix health check issue with shada file (#13291)
==> luarocks make --lua-dir=/usr/local/opt/luajit --tree=/private/tmp/neovim-202
==> luarocks make --lua-dir=/usr/local/opt/luajit --tree=/private/tmp/neovim-202
==> luarocks make --lua-dir=/usr/local/opt/luajit --tree=/private/tmp/neovim-202
==> cmake . -DCMAKE_C_COMPILER=/usr/bin/clang -DCMAKE_CXX_COMPILER=/usr/bin/clan
==> make install
==> cmake .. -DCMAKE_C_COMPILER=/usr/bin/clang -DCMAKE_CXX_COMPILER=/usr/bin/cla
==> make install
Last 15 lines from /Users/mpcsh/Library/Logs/Homebrew/neovim/07.make:
^
/tmp/neovim-20201214-42949-1h90lp3/src/nvim/os/lang.c:63:15: error: use of undeclared identifier 'LC_ALL'; did you mean 'P_ALL'?
setlocale(LC_ALL, "");
^~~~~~
P_ALL
/Library/Developer/CommandLineTools/SDKs/MacOSX11.0.sdk/usr/include/sys/wait.h:80:2: note: 'P_ALL' declared here
P_ALL,
^
2 errors generated.
[ 50%] Building C object src/nvim/CMakeFiles/nvim.dir/os/time.c.o
cd /tmp/neovim-20201214-42949-1h90lp3/build/src/nvim && /usr/bin/clang -DINCLUDE_GENERATED_DECLARATIONS -DMIN_LOG_LEVEL=3 -DNVIM_MSGPACK_HAS_FLOAT32 -D_GNU_SOURCE -Dnvim_EXPORTS -I/tmp/neovim-20201214-42949-1h90lp3/build/config -I/tmp/neovim-20201214-42949-1h90lp3/src -I/tmp/neovim-20201214-42949-1h90lp3/build/src/nvim/auto -I/tmp/neovim-20201214-42949-1h90lp3/build/include -I/usr/local/include/luajit-2.0 -isystem /usr/local/include -isystem /tmp/neovim-20201214-42949-1h90lp3/deps-build/include -isystem /usr/local/opt/gettext/include -U_FORTIFY_SOURCE -D_FORTIFY_SOURCE=1 -DNDEBUG -flto=thin -arch arm64 -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX11.0.sdk -Wall -Wextra -pedantic -Wno-unused-parameter -Wstrict-prototypes -std=gnu99 -Wshadow -Wconversion -Wmissing-prototypes -Wimplicit-fallthrough -Wvla -fstack-protector-strong -fno-common -fdiagnostics-color=auto -o CMakeFiles/nvim.dir/os/time.c.o -c /tmp/neovim-20201214-42949-1h90lp3/src/nvim/os/time.c
make[2]: *** [src/nvim/CMakeFiles/nvim.dir/os/lang.c.o] Error 1
make[2]: *** Waiting for unfinished jobs....
make[1]: *** [src/nvim/CMakeFiles/nvim.dir/all] Error 2
make: *** [all] Error 2
READ THIS: https://docs.brew.sh/Troubleshooting
Please create pull requests instead of asking for help on Homebrew's GitHub,
Twitter or any other official channels.
```
### Expected behaviour
`brew install --HEAD neovim` works
|
non_test
|
brew install head failing with errors nvim version nvim dev vim u defaults version behaves differently no operating system version macos on macbook air with apple silicon soc terminal name version term xterm actual behaviour note i ve already found and followed the build from source prerequisite instructions ❯ brew install head neovim updating homebrew downloading already downloaded users mpcsh library caches homebrew downloads libmpack lua tar gz downloading already downloaded users mpcsh library caches homebrew downloads lpeg src rock downloading already downloaded users mpcsh library caches homebrew downloads inspect src rock downloading already downloaded users mpcsh library caches homebrew downloads lua compat tar gz downloading already downloaded users mpcsh library caches homebrew downloads luv tar gz cloning updating users mpcsh library caches homebrew neovim git checking out branch master already on master your branch is up to date with origin master head is now at healthcheck fix health check issue with shada file luarocks make lua dir usr local opt luajit tree private tmp neovim luarocks make lua dir usr local opt luajit tree private tmp neovim luarocks make lua dir usr local opt luajit tree private tmp neovim cmake dcmake c compiler usr bin clang dcmake cxx compiler usr bin clan make install cmake dcmake c compiler usr bin clang dcmake cxx compiler usr bin cla make install last lines from users mpcsh library logs homebrew neovim make tmp neovim src nvim os lang c error use of undeclared identifier lc all did you mean p all setlocale lc all p all library developer commandlinetools sdks sdk usr include sys wait h note p all declared here p all errors generated building c object src nvim cmakefiles nvim dir os time c o cd tmp neovim build src nvim usr bin clang dinclude generated declarations dmin log level dnvim msgpack has d gnu source dnvim exports i tmp neovim build config i tmp neovim src i tmp neovim build src nvim auto i tmp neovim build include i usr local include luajit isystem usr local include isystem tmp neovim deps build include isystem usr local opt gettext include u fortify source d fortify source dndebug flto thin arch isysroot library developer commandlinetools sdks sdk wall wextra pedantic wno unused parameter wstrict prototypes std wshadow wconversion wmissing prototypes wimplicit fallthrough wvla fstack protector strong fno common fdiagnostics color auto o cmakefiles nvim dir os time c o c tmp neovim src nvim os time c make error make waiting for unfinished jobs make error make error read this please create pull requests instead of asking for help on homebrew s github twitter or any other official channels expected behaviour brew install head neovim works
| 0
|
781,312
| 27,432,310,653
|
IssuesEvent
|
2023-03-02 03:01:14
|
AY2223S2-CS2103-F11-3/tp
|
https://api.github.com/repos/AY2223S2-CS2103-F11-3/tp
|
opened
|
Update command syntax - argument parsing
|
priority.High type.Task
|
Argument flags are delimited by a common `--` delimiter.
|
1.0
|
Update command syntax - argument parsing - Argument flags are delimited by a common `--` delimiter.
|
non_test
|
update command syntax argument parsing argument flags are delimited by a common delimiter
| 0
|
232,088
| 17,771,421,977
|
IssuesEvent
|
2021-08-30 14:05:18
|
stakwork/sphinx-kotlin
|
https://api.github.com/repos/stakwork/sphinx-kotlin
|
closed
|
Contact detail screen
|
documentation
|
When tapping on chat header of a chat of type Conversation, instead of taking user to TribeDetail screen, we should show ContactDetail screen.

The only 2 fields editable are:
- Nickname
- Route Hint
Address can't be edited, and the QR code icon on the right side should take user to QR code details view so it can be scanned or copied to clipboard.
|
1.0
|
Contact detail screen - When tapping on chat header of a chat of type Conversation, instead of taking user to TribeDetail screen, we should show ContactDetail screen.

The only 2 fields editable are:
- Nickname
- Route Hint
Address can't be edited, and the QR code icon on the right side should take user to QR code details view so it can be scanned or copied to clipboard.
|
non_test
|
contact detail screen when tapping on chat header of a chat of type conversation instead of taking user to tribedetail screen we should show contactdetail screen the only fields editable are nickname route hint address can t be edited and the qr code icon on the right side should take user to qr code details view so it can be scanned or copied to clipboard
| 0
|
72,613
| 7,302,942,093
|
IssuesEvent
|
2018-02-27 11:21:25
|
Microsoft/vscode
|
https://api.github.com/repos/Microsoft/vscode
|
closed
|
Test: Language Pack cache cleanup
|
testplan-item
|
Tests cleaning up language pack caches
- [x] anyOS: @sandy081 - tested on Windows
- [x] anyOS: @isidorn - tested on OS X
Complexity: 2
## Steps
- install a Language pack from the market place (e.g Bulgarian)
- close VS Code.
- restart VS Code with --locale=`the language pack you downloaded` (e.g --locale=bg)
- go to the user data directory on your file system. Under Windows that is `C:\Users\${user}\AppData\Roaming\Code - Insiders`
- ensure that there is a `languagepacks.json` file and a directorry `clp`
- navigate into `clp`and ensure that there is a directory with a hash and the locale of the language pack you installed
- close VS Code
- restart again with the locale switch
- open the log for the shared process. After ~40 seconds you should see a message that the language pack cleanup starts and that the installed language pack is skipped since it is still in use.
- uninstall the language pack
- close VS Code and restart it without a locale switch
- open the log for the shared process. Again after ~40 seconds you should see a message that the language pack cleanup starts . This time a language pack cache entry should be deleted.
- go to the file system again. Ensure that the directory entry below `clp` got deleted.
|
1.0
|
Test: Language Pack cache cleanup - Tests cleaning up language pack caches
- [x] anyOS: @sandy081 - tested on Windows
- [x] anyOS: @isidorn - tested on OS X
Complexity: 2
## Steps
- install a Language pack from the market place (e.g Bulgarian)
- close VS Code.
- restart VS Code with --locale=`the language pack you downloaded` (e.g --locale=bg)
- go to the user data directory on your file system. Under Windows that is `C:\Users\${user}\AppData\Roaming\Code - Insiders`
- ensure that there is a `languagepacks.json` file and a directorry `clp`
- navigate into `clp`and ensure that there is a directory with a hash and the locale of the language pack you installed
- close VS Code
- restart again with the locale switch
- open the log for the shared process. After ~40 seconds you should see a message that the language pack cleanup starts and that the installed language pack is skipped since it is still in use.
- uninstall the language pack
- close VS Code and restart it without a locale switch
- open the log for the shared process. Again after ~40 seconds you should see a message that the language pack cleanup starts . This time a language pack cache entry should be deleted.
- go to the file system again. Ensure that the directory entry below `clp` got deleted.
|
test
|
test language pack cache cleanup tests cleaning up language pack caches anyos tested on windows anyos isidorn tested on os x complexity steps install a language pack from the market place e g bulgarian close vs code restart vs code with locale the language pack you downloaded e g locale bg go to the user data directory on your file system under windows that is c users user appdata roaming code insiders ensure that there is a languagepacks json file and a directorry clp navigate into clp and ensure that there is a directory with a hash and the locale of the language pack you installed close vs code restart again with the locale switch open the log for the shared process after seconds you should see a message that the language pack cleanup starts and that the installed language pack is skipped since it is still in use uninstall the language pack close vs code and restart it without a locale switch open the log for the shared process again after seconds you should see a message that the language pack cleanup starts this time a language pack cache entry should be deleted go to the file system again ensure that the directory entry below clp got deleted
| 1
|
61,202
| 14,619,099,206
|
IssuesEvent
|
2020-12-22 17:16:21
|
kenferrara/calcite
|
https://api.github.com/repos/kenferrara/calcite
|
opened
|
CVE-2020-13956 (Medium) detected in httpclient-4.5.9.jar
|
security vulnerability
|
## CVE-2020-13956 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>httpclient-4.5.9.jar</b></p></summary>
<p>Apache HttpComponents Client</p>
<p>Library home page: <a href="http://hc.apache.org/">http://hc.apache.org/</a></p>
<p>Path to dependency file: calcite/server/build.gradle.kts</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.9/a25c1be5ce99d0ce99aa43eb982868c796dd0775/httpclient-4.5.9.jar,canner/.gradle/caches/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.9/a25c1be5ce99d0ce99aa43eb982868c796dd0775/httpclient-4.5.9.jar</p>
<p>
Dependency Hierarchy:
- pig-0.16.0.jar (Root Library)
- jets3t-0.9.0.jar
- :x: **httpclient-4.5.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kenferrara/calcite/commit/5e2bb91de24b843d35df04b29fe7158fad414ee0">5e2bb91de24b843d35df04b29fe7158fad414ee0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956>CVE-2020-13956</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.httpcomponents","packageName":"httpclient","packageVersion":"4.5.9","isTransitiveDependency":true,"dependencyTree":"org.apache.pig:pig:0.16.0;net.java.dev.jets3t:jets3t:0.9.0;org.apache.httpcomponents:httpclient:4.5.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3"}],"vulnerabilityIdentifier":"CVE-2020-13956","vulnerabilityDetails":"Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-13956 (Medium) detected in httpclient-4.5.9.jar - ## CVE-2020-13956 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>httpclient-4.5.9.jar</b></p></summary>
<p>Apache HttpComponents Client</p>
<p>Library home page: <a href="http://hc.apache.org/">http://hc.apache.org/</a></p>
<p>Path to dependency file: calcite/server/build.gradle.kts</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.9/a25c1be5ce99d0ce99aa43eb982868c796dd0775/httpclient-4.5.9.jar,canner/.gradle/caches/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.9/a25c1be5ce99d0ce99aa43eb982868c796dd0775/httpclient-4.5.9.jar</p>
<p>
Dependency Hierarchy:
- pig-0.16.0.jar (Root Library)
- jets3t-0.9.0.jar
- :x: **httpclient-4.5.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kenferrara/calcite/commit/5e2bb91de24b843d35df04b29fe7158fad414ee0">5e2bb91de24b843d35df04b29fe7158fad414ee0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956>CVE-2020-13956</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.httpcomponents","packageName":"httpclient","packageVersion":"4.5.9","isTransitiveDependency":true,"dependencyTree":"org.apache.pig:pig:0.16.0;net.java.dev.jets3t:jets3t:0.9.0;org.apache.httpcomponents:httpclient:4.5.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3"}],"vulnerabilityIdentifier":"CVE-2020-13956","vulnerabilityDetails":"Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve medium detected in httpclient jar cve medium severity vulnerability vulnerable library httpclient jar apache httpcomponents client library home page a href path to dependency file calcite server build gradle kts path to vulnerable library home wss scanner gradle caches modules files org apache httpcomponents httpclient httpclient jar canner gradle caches modules files org apache httpcomponents httpclient httpclient jar dependency hierarchy pig jar root library jar x httpclient jar vulnerable library found in head commit a href found in base branch master vulnerability details apache httpclient versions prior to version and can misinterpret malformed authority component in request uris passed to the library as java net uri object and pick the wrong target host for request execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache httpcomponents httpclient org apache httpcomponents httpclient osgi org apache httpcomponents org apache httpcomponents osgi isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails apache httpclient versions prior to version and can misinterpret malformed authority component in request uris passed to the library as java net uri object and pick the wrong target host for request execution vulnerabilityurl
| 0
|
18,234
| 5,612,817,055
|
IssuesEvent
|
2017-04-03 07:11:06
|
se-edu/addressbook-level4
|
https://api.github.com/repos/se-edu/addressbook-level4
|
opened
|
Remove magic string
|
a-CodeQuality
|
we should remove magic string like the command word "add" and prefix "p/" from where they are defined, as discussed [here](https://github.com/se-edu/addressbook-level4/pull/335#discussion_r107157865) and [here](https://github.com/se-edu/addressbook-level4/pull/373#pullrequestreview-30395535).
Places to look for these magic strings include various tests (including testutil) and some commands' MESSAGE_USAGE.
|
1.0
|
Remove magic string - we should remove magic string like the command word "add" and prefix "p/" from where they are defined, as discussed [here](https://github.com/se-edu/addressbook-level4/pull/335#discussion_r107157865) and [here](https://github.com/se-edu/addressbook-level4/pull/373#pullrequestreview-30395535).
Places to look for these magic strings include various tests (including testutil) and some commands' MESSAGE_USAGE.
|
non_test
|
remove magic string we should remove magic string like the command word add and prefix p from where they are defined as discussed and places to look for these magic strings include various tests including testutil and some commands message usage
| 0
|
128,328
| 10,525,265,020
|
IssuesEvent
|
2019-09-30 14:47:19
|
sbt/sbt
|
https://api.github.com/repos/sbt/sbt
|
closed
|
JUnitXmlReportPlugin generates invalid testcase name
|
area/test_framework uncategorized
|
It seems that in certain cases an invalid test-case name is generated in the JUnit XML output.
## steps
This is one arbitrary example demonstrating the issue:
```scala
"This example" should {
"given <VER.123 FOOBAR> yield <ver.123 foobar>" in {
ok
}
}
```
## problem
With this example (using specs2 `3.8.6`) the generated JUnit XML contains the following test-case:
```xml
<testcase classname="org.example.ExampleSpec" name="123 foobar>" time="0.006">
</testcase>
```
## expectation
Now, obviously the correct XML should be as follows:
```xml
<testcase classname="org.example.ExampleSpec" name="This example should::given <VER.123 FOOBAR> yield <ver.123 foobar>" time="0.006">
</testcase>
```
## notes
* See https://github.com/Pyppe/specs2-junitxml-testcase-name as a simple project demonstrating the issue.
* I initially reported this in https://github.com/etorreborre/specs2/issues/537, but @etorreborre graciously debugged it to be part of sbt's `JUnitXmlReportPlugin`.
|
1.0
|
JUnitXmlReportPlugin generates invalid testcase name - It seems that in certain cases an invalid test-case name is generated in the JUnit XML output.
## steps
This is one arbitrary example demonstrating the issue:
```scala
"This example" should {
"given <VER.123 FOOBAR> yield <ver.123 foobar>" in {
ok
}
}
```
## problem
With this example (using specs2 `3.8.6`) the generated JUnit XML contains the following test-case:
```xml
<testcase classname="org.example.ExampleSpec" name="123 foobar>" time="0.006">
</testcase>
```
## expectation
Now, obviously the correct XML should be as follows:
```xml
<testcase classname="org.example.ExampleSpec" name="This example should::given <VER.123 FOOBAR> yield <ver.123 foobar>" time="0.006">
</testcase>
```
## notes
* See https://github.com/Pyppe/specs2-junitxml-testcase-name as a simple project demonstrating the issue.
* I initially reported this in https://github.com/etorreborre/specs2/issues/537, but @etorreborre graciously debugged it to be part of sbt's `JUnitXmlReportPlugin`.
|
test
|
junitxmlreportplugin generates invalid testcase name it seems that in certain cases an invalid test case name is generated in the junit xml output steps this is one arbitrary example demonstrating the issue scala this example should given yield in ok problem with this example using the generated junit xml contains the following test case xml expectation now obviously the correct xml should be as follows xml notes see as a simple project demonstrating the issue i initially reported this in but etorreborre graciously debugged it to be part of sbt s junitxmlreportplugin
| 1
|
85,353
| 10,606,331,140
|
IssuesEvent
|
2019-10-10 22:58:07
|
EdgeFortress/OSS-13
|
https://api.github.com/repos/EdgeFortress/OSS-13
|
closed
|
UI Design Layout
|
design ui
|
Original game's interface looks a bit outdated:

We need new UI Design Layout with next prerequisites:
1. Optimum TileGrid size is 960 x 960 pixels (64 pixels one tile * 15 tiles);
2. The chat should be large enough to accommodate about as much text as the original one.
3. Least possible number of elements should placed on the TileGrid.
4. Mostly all GUI elements above chat can be moved to hotkeys and sub-windows.
5. RMB menu can be implemented as pop-up bubbles with objects/actions around clicked point.
6. You can use as much subwindows and hotkeys as you needed.
7. Console at the bottom of window is obsolete and will not be implemented.
Key gameplay UI elements need to be designed:
- [ ] Chat
- [ ] Inventory
- [ ] Right Mouse Button menu (now it's simple Desktop-like drop-down menu).
- [ ] Health, Body Temperature, Gas indicators.
|
1.0
|
UI Design Layout - Original game's interface looks a bit outdated:

We need new UI Design Layout with next prerequisites:
1. Optimum TileGrid size is 960 x 960 pixels (64 pixels one tile * 15 tiles);
2. The chat should be large enough to accommodate about as much text as the original one.
3. Least possible number of elements should placed on the TileGrid.
4. Mostly all GUI elements above chat can be moved to hotkeys and sub-windows.
5. RMB menu can be implemented as pop-up bubbles with objects/actions around clicked point.
6. You can use as much subwindows and hotkeys as you needed.
7. Console at the bottom of window is obsolete and will not be implemented.
Key gameplay UI elements need to be designed:
- [ ] Chat
- [ ] Inventory
- [ ] Right Mouse Button menu (now it's simple Desktop-like drop-down menu).
- [ ] Health, Body Temperature, Gas indicators.
|
non_test
|
ui design layout original game s interface looks a bit outdated we need new ui design layout with next prerequisites optimum tilegrid size is x pixels pixels one tile tiles the chat should be large enough to accommodate about as much text as the original one least possible number of elements should placed on the tilegrid mostly all gui elements above chat can be moved to hotkeys and sub windows rmb menu can be implemented as pop up bubbles with objects actions around clicked point you can use as much subwindows and hotkeys as you needed console at the bottom of window is obsolete and will not be implemented key gameplay ui elements need to be designed chat inventory right mouse button menu now it s simple desktop like drop down menu health body temperature gas indicators
| 0
|
88,020
| 25,281,481,510
|
IssuesEvent
|
2022-11-16 16:04:38
|
intel/media-driver
|
https://api.github.com/repos/intel/media-driver
|
closed
|
[Bug]: Cannot build on musl
|
Build Common
|
### Which component impacted?
Build
### Is it regression? Good in old configuration?
_No response_
### What happened?
Using musl as libc, I am not able to build because execinfo.h is not found.
### What's the usage scenario when you are seeing the problem?
Others
### What impacted?
_No response_
### Debug Information
Here is the output of the build :
media_softlet/linux/common/os/osservice/mos_utilities_specific.cpp:37:10: fatal error: execinfo.h: No such file or directory
### Do you want to contribute a patch to fix the issue?
[Here](https://raw.githubusercontent.com/void-linux/void-packages/master/srcpkgs/intel-media-driver/patches/execinfo.patch) is a patch from void linux. It doesn't work on master, but I'll try to edit it.
|
1.0
|
[Bug]: Cannot build on musl - ### Which component impacted?
Build
### Is it regression? Good in old configuration?
_No response_
### What happened?
Using musl as libc, I am not able to build because execinfo.h is not found.
### What's the usage scenario when you are seeing the problem?
Others
### What impacted?
_No response_
### Debug Information
Here is the output of the build :
media_softlet/linux/common/os/osservice/mos_utilities_specific.cpp:37:10: fatal error: execinfo.h: No such file or directory
### Do you want to contribute a patch to fix the issue?
[Here](https://raw.githubusercontent.com/void-linux/void-packages/master/srcpkgs/intel-media-driver/patches/execinfo.patch) is a patch from void linux. It doesn't work on master, but I'll try to edit it.
|
non_test
|
cannot build on musl which component impacted build is it regression good in old configuration no response what happened using musl as libc i am not able to build because execinfo h is not found what s the usage scenario when you are seeing the problem others what impacted no response debug information here is the output of the build media softlet linux common os osservice mos utilities specific cpp fatal error execinfo h no such file or directory do you want to contribute a patch to fix the issue is a patch from void linux it doesn t work on master but i ll try to edit it
| 0
|
288,813
| 31,930,913,635
|
IssuesEvent
|
2023-09-19 07:21:32
|
Trinadh465/linux-4.1.15_CVE-2023-4128
|
https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2023-4128
|
opened
|
CVE-2021-26931 (Medium) detected in linux-stable-rtv4.1.33, linux-stable-rtv4.1.33
|
Mend: dependency security vulnerability
|
## CVE-2021-26931 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv4.1.33</b>, <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel 2.6.39 through 5.10.16, as used in Xen. Block, net, and SCSI backends consider certain errors a plain bug, deliberately causing a kernel crash. For errors potentially being at least under the influence of guests (such as out of memory conditions), it isn't correct to assume a plain bug. Memory allocations potentially causing such crashes occur only when Linux is running in PV mode, though. This affects drivers/block/xen-blkback/blkback.c and drivers/xen/xen-scsiback.c.
<p>Publish Date: 2021-02-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-26931>CVE-2021-26931</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-26931">https://nvd.nist.gov/vuln/detail/CVE-2021-26931</a></p>
<p>Release Date: 2021-02-17</p>
<p>Fix Resolution: linux-libc-headers - 5.13;linux-yocto - 5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1,4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-26931 (Medium) detected in linux-stable-rtv4.1.33, linux-stable-rtv4.1.33 - ## CVE-2021-26931 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv4.1.33</b>, <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel 2.6.39 through 5.10.16, as used in Xen. Block, net, and SCSI backends consider certain errors a plain bug, deliberately causing a kernel crash. For errors potentially being at least under the influence of guests (such as out of memory conditions), it isn't correct to assume a plain bug. Memory allocations potentially causing such crashes occur only when Linux is running in PV mode, though. This affects drivers/block/xen-blkback/blkback.c and drivers/xen/xen-scsiback.c.
<p>Publish Date: 2021-02-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-26931>CVE-2021-26931</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-26931">https://nvd.nist.gov/vuln/detail/CVE-2021-26931</a></p>
<p>Release Date: 2021-02-17</p>
<p>Fix Resolution: linux-libc-headers - 5.13;linux-yocto - 5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1,4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in linux stable linux stable cve medium severity vulnerability vulnerable libraries linux stable linux stable vulnerability details an issue was discovered in the linux kernel through as used in xen block net and scsi backends consider certain errors a plain bug deliberately causing a kernel crash for errors potentially being at least under the influence of guests such as out of memory conditions it isn t correct to assume a plain bug memory allocations potentially causing such crashes occur only when linux is running in pv mode though this affects drivers block xen blkback blkback c and drivers xen xen scsiback c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux libc headers linux yocto gitautoinc gitautoinc step up your open source security game with mend
| 0
|
107,784
| 9,223,256,978
|
IssuesEvent
|
2019-03-12 02:37:54
|
brave/browser-android-tabs
|
https://api.github.com/repos/brave/browser-android-tabs
|
closed
|
twitch.tv crashes when trying to view Desktop Site
|
QA/Test-plan-specified QA/Yes crash
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description <!--Provide a brief description of the issue-->
When I go to twitch.tv and try to select 'Desktop Site' from hamburger menu, Brave crashes.
## Steps to Reproduce <!--Please add a series of steps to reproduce the issue-->
1. Install 1.0.86 (BR_6)
2. Navigate to twitch.tv, page loads
3. Tap on hamburger menu
4. Select Desktop site.
## Actual result: <!--Please add screenshots if needed-->
Get message which says
'Unfortunately, Brave has stopped.'
## Expected result:
No crash, desktop site should open.
## Issue reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
easily
## Issue happens on: <!-- Mention yes or no -->
- Current Playstore version? no, does not reproduce on 1.0.84
- Beta build? Yes, Rewards Beta
## Device Details:
- Install Type(ARM, x86): all
- Device(Phone, Tablet, Phablet): all
- Android Version: all
## Brave version:
1.0.86 (BR_6)
### Website problems only:
- Does the issue resolve itself when disabling Brave Shields? no
- Is the issue reproducible on the latest version of Chrome? no
### Additional Information
Have only experienced on the Rewards build so far, but do not need to enable Rewards to experience the crash.
|
1.0
|
twitch.tv crashes when trying to view Desktop Site - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description <!--Provide a brief description of the issue-->
When I go to twitch.tv and try to select 'Desktop Site' from hamburger menu, Brave crashes.
## Steps to Reproduce <!--Please add a series of steps to reproduce the issue-->
1. Install 1.0.86 (BR_6)
2. Navigate to twitch.tv, page loads
3. Tap on hamburger menu
4. Select Desktop site.
## Actual result: <!--Please add screenshots if needed-->
Get message which says
'Unfortunately, Brave has stopped.'
## Expected result:
No crash, desktop site should open.
## Issue reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
easily
## Issue happens on: <!-- Mention yes or no -->
- Current Playstore version? no, does not reproduce on 1.0.84
- Beta build? Yes, Rewards Beta
## Device Details:
- Install Type(ARM, x86): all
- Device(Phone, Tablet, Phablet): all
- Android Version: all
## Brave version:
1.0.86 (BR_6)
### Website problems only:
- Does the issue resolve itself when disabling Brave Shields? no
- Is the issue reproducible on the latest version of Chrome? no
### Additional Information
Have only experienced on the Rewards build so far, but do not need to enable Rewards to experience the crash.
|
test
|
twitch tv crashes when trying to view desktop site have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description when i go to twitch tv and try to select desktop site from hamburger menu brave crashes steps to reproduce install br navigate to twitch tv page loads tap on hamburger menu select desktop site actual result get message which says unfortunately brave has stopped expected result no crash desktop site should open issue reproduces how often easily issue happens on current playstore version no does not reproduce on beta build yes rewards beta device details install type arm all device phone tablet phablet all android version all brave version br website problems only does the issue resolve itself when disabling brave shields no is the issue reproducible on the latest version of chrome no additional information have only experienced on the rewards build so far but do not need to enable rewards to experience the crash
| 1
|
61,970
| 14,648,070,031
|
IssuesEvent
|
2020-12-27 00:02:22
|
wallanpsantos/cursoAngularNove
|
https://api.github.com/repos/wallanpsantos/cursoAngularNove
|
opened
|
CVE-2020-7774 (High) detected in y18n-4.0.0.tgz
|
security vulnerability
|
## CVE-2020-7774 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>y18n-4.0.0.tgz</b></p></summary>
<p>the bare-bones internationalization library used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz">https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz</a></p>
<p>Path to dependency file: cursoAngularNove/backend/package.json</p>
<p>Path to vulnerable library: cursoAngularNove/backend/node_modules/y18n/package.json,cursoAngularNove/frontend/node_modules/y18n/package.json</p>
<p>
Dependency Hierarchy:
- json-server-0.16.1.tgz (Root Library)
- yargs-15.4.1.tgz
- :x: **y18n-4.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/wallanpsantos/cursoAngularNove/commit/c6b9eaff3e9f65596ffd12cd6b64751337db2942">c6b9eaff3e9f65596ffd12cd6b64751337db2942</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package y18n before 4.0.1 and 5.0.5. PoC by po6ix: const y18n = require('y18n')(); y18n.setLocale('__proto__'); y18n.updateLocale({polluted: true}); console.log(polluted); // true
<p>Publish Date: 2020-11-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7774>CVE-2020-7774</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7774">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7774</a></p>
<p>Release Date: 2020-11-17</p>
<p>Fix Resolution: 5.0.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7774 (High) detected in y18n-4.0.0.tgz - ## CVE-2020-7774 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>y18n-4.0.0.tgz</b></p></summary>
<p>the bare-bones internationalization library used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz">https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz</a></p>
<p>Path to dependency file: cursoAngularNove/backend/package.json</p>
<p>Path to vulnerable library: cursoAngularNove/backend/node_modules/y18n/package.json,cursoAngularNove/frontend/node_modules/y18n/package.json</p>
<p>
Dependency Hierarchy:
- json-server-0.16.1.tgz (Root Library)
- yargs-15.4.1.tgz
- :x: **y18n-4.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/wallanpsantos/cursoAngularNove/commit/c6b9eaff3e9f65596ffd12cd6b64751337db2942">c6b9eaff3e9f65596ffd12cd6b64751337db2942</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package y18n before 4.0.1 and 5.0.5. PoC by po6ix: const y18n = require('y18n')(); y18n.setLocale('__proto__'); y18n.updateLocale({polluted: true}); console.log(polluted); // true
<p>Publish Date: 2020-11-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7774>CVE-2020-7774</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7774">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7774</a></p>
<p>Release Date: 2020-11-17</p>
<p>Fix Resolution: 5.0.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in tgz cve high severity vulnerability vulnerable library tgz the bare bones internationalization library used by yargs library home page a href path to dependency file cursoangularnove backend package json path to vulnerable library cursoangularnove backend node modules package json cursoangularnove frontend node modules package json dependency hierarchy json server tgz root library yargs tgz x tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects the package before and poc by const require setlocale proto updatelocale polluted true console log polluted true publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
356,411
| 25,176,183,737
|
IssuesEvent
|
2022-11-11 09:27:50
|
Santosh3007/pe
|
https://api.github.com/repos/Santosh3007/pe
|
opened
|
Uneven Text size in table of contents
|
severity.VeryLow type.DocumentationBug
|
## Issue
- Uneven text size for `Appendix: Instructions for manual testing`
## Screenshot

<!--session: 1668151550995-b01faa43-82fc-43e2-93df-11e1d10a284a-->
<!--Version: Web v3.4.4-->
|
1.0
|
Uneven Text size in table of contents - ## Issue
- Uneven text size for `Appendix: Instructions for manual testing`
## Screenshot

<!--session: 1668151550995-b01faa43-82fc-43e2-93df-11e1d10a284a-->
<!--Version: Web v3.4.4-->
|
non_test
|
uneven text size in table of contents issue uneven text size for appendix instructions for manual testing screenshot
| 0
|
294,571
| 25,383,735,207
|
IssuesEvent
|
2022-11-21 19:53:33
|
chocolatey/choco
|
https://api.github.com/repos/chocolatey/choco
|
opened
|
PushCommand tests should push to a mocked endpoint and not Chocolatey Community Repository
|
0 - Backlog Tests
|
## Is Your Feature Request Related To A Problem? Please describe.
Currently PushCommand tests require an API Key to push to the Chocolatey Community Repository. This means that community members are not able to run the full test suite as they do not have access to this API Key.
## Describe The Solution. Why is it needed?
PowerShell has the ability to respond to HTTP requests. We could therefore have our tests listen for Chocolatey to push a package to it and respond with exactly the message we want.
## Additional Context.
NA
## Related Issues
NA
|
1.0
|
PushCommand tests should push to a mocked endpoint and not Chocolatey Community Repository - ## Is Your Feature Request Related To A Problem? Please describe.
Currently PushCommand tests require an API Key to push to the Chocolatey Community Repository. This means that community members are not able to run the full test suite as they do not have access to this API Key.
## Describe The Solution. Why is it needed?
PowerShell has the ability to respond to HTTP requests. We could therefore have our tests listen for Chocolatey to push a package to it and respond with exactly the message we want.
## Additional Context.
NA
## Related Issues
NA
|
test
|
pushcommand tests should push to a mocked endpoint and not chocolatey community repository is your feature request related to a problem please describe currently pushcommand tests require an api key to push to the chocolatey community repository this means that community members are not able to run the full test suite as they do not have access to this api key describe the solution why is it needed powershell has the ability to respond to http requests we could therefore have our tests listen for chocolatey to push a package to it and respond with exactly the message we want additional context na related issues na
| 1
|
152,464
| 12,109,742,436
|
IssuesEvent
|
2020-04-21 09:15:41
|
WoWManiaUK/Redemption
|
https://api.github.com/repos/WoWManiaUK/Redemption
|
closed
|
[NPC/ICC] Nerub'ar Broodkeeper (Web Wrap)
|
Fix - Tester Confirmed
|
**Links:** https://www.wowhead.com/npc=36725/nerubar-broodkeeper#abilities
https://www.wowhead.com/spell=70980/web-wrap
https://www.wowhead.com/npc=38028/web-wrap
**What is Happening:** The [Nerub'ar Broodkeeper](https://www.wowhead.com/npc=36725/nerubar-broodkeeper#abilities) in Icecrown Citadel currently cast their [Web Wraps](https://www.wowhead.com/spell=70980/web-wrap) properly, but only the first part of the spell works. The target gets immobilized for 3 seconds, but are then freed and nothing more happens.
**What Should happen:** After the 3s immobilization of the Web Wrap cast time, the target should become encased in a [cocoon](https://www.wowhead.com/npc=38028/web-wrap) .
|
1.0
|
[NPC/ICC] Nerub'ar Broodkeeper (Web Wrap) - **Links:** https://www.wowhead.com/npc=36725/nerubar-broodkeeper#abilities
https://www.wowhead.com/spell=70980/web-wrap
https://www.wowhead.com/npc=38028/web-wrap
**What is Happening:** The [Nerub'ar Broodkeeper](https://www.wowhead.com/npc=36725/nerubar-broodkeeper#abilities) in Icecrown Citadel currently cast their [Web Wraps](https://www.wowhead.com/spell=70980/web-wrap) properly, but only the first part of the spell works. The target gets immobilized for 3 seconds, but are then freed and nothing more happens.
**What Should happen:** After the 3s immobilization of the Web Wrap cast time, the target should become encased in a [cocoon](https://www.wowhead.com/npc=38028/web-wrap) .
|
test
|
nerub ar broodkeeper web wrap links what is happening the in icecrown citadel currently cast their properly but only the first part of the spell works the target gets immobilized for seconds but are then freed and nothing more happens what should happen after the immobilization of the web wrap cast time the target should become encased in a
| 1
|
210,205
| 16,090,673,503
|
IssuesEvent
|
2021-04-26 16:18:08
|
spring-projects/spring-framework
|
https://api.github.com/repos/spring-projects/spring-framework
|
closed
|
MockMvc's MVC_RESULT_ATTRIBUTE lost with HandlerMappingIntrospector and RouterFunctions in use
|
in: test in: web type: bug
|
Related to https://github.com/spring-projects/spring-security-samples/issues/9
A contributor shared the following sample application: https://github.com/hantsy/spring-webmvc-auth0-sample
The tests result in a `NullPointerException` because the `MockMvc.MVC_RESULT_ATTRIBUTE` is missing.
```java
HandlerExecutionChain chain = super.getHandler(request);
if (chain != null) {
DefaultMvcResult mvcResult = getMvcResult(request); // returns null
mvcResult.setHandler(chain.getHandler());
mvcResult.setInterceptors(chain.getInterceptors());
}
return chain;
```
It gets removed due to the following arrangement:
* Spring Security's `CorsFilter` by default uses `HandlerMappingIntrospector`
* `HandlerMappingIntrospector` uses `RequestAttributeChangeIgnoringWrapper` which ignores all but `PATH_ATTRIBUTE`
* `RequestPredicates#restoreAttributes` attempts to restore the attributes to a previous state by clearing the attribute set and then re-adding each attribute one by one
Before `CorsFilter` runs, `MVC_RESULT_ATTRIBUTE` is present in the request. When `RequestPredicates#restoreAttributes` is run, it removes all attributes. Then, when it tries to add the original set back in, `RequestAttributeChangeIgnoringWrapper` only adds `PATH_ATTRIBUTE` back in.
For the specific sample, the tests can be repaired by removing the `CorsFilter` or by exposing a custom `CorsConfigurationSource` bean since either of those will prevent `RequestAttributeChangeIgnoringWrapper` from wrapping the request.
I was also able to fix the tests by adding the following to `RequestAttributeChangeIgnoringWrapper`:
```java
@Override
public void removeAttribute(String name) {
if (name.equals(ServletRequestPathUtils.PATH_ATTRIBUTE) || name.equals(UrlPathHelper.PATH_ATTRIBUTE)) {
super.removeAttribute(name);
}
}
```
At least in this isolated case, it seems reasonable that if an attribute cannot be set, it also should not be able to be removed.
|
1.0
|
MockMvc's MVC_RESULT_ATTRIBUTE lost with HandlerMappingIntrospector and RouterFunctions in use - Related to https://github.com/spring-projects/spring-security-samples/issues/9
A contributor shared the following sample application: https://github.com/hantsy/spring-webmvc-auth0-sample
The tests result in a `NullPointerException` because the `MockMvc.MVC_RESULT_ATTRIBUTE` is missing.
```java
HandlerExecutionChain chain = super.getHandler(request);
if (chain != null) {
DefaultMvcResult mvcResult = getMvcResult(request); // returns null
mvcResult.setHandler(chain.getHandler());
mvcResult.setInterceptors(chain.getInterceptors());
}
return chain;
```
It gets removed due to the following arrangement:
* Spring Security's `CorsFilter` by default uses `HandlerMappingIntrospector`
* `HandlerMappingIntrospector` uses `RequestAttributeChangeIgnoringWrapper` which ignores all but `PATH_ATTRIBUTE`
* `RequestPredicates#restoreAttributes` attempts to restore the attributes to a previous state by clearing the attribute set and then re-adding each attribute one by one
Before `CorsFilter` runs, `MVC_RESULT_ATTRIBUTE` is present in the request. When `RequestPredicates#restoreAttributes` is run, it removes all attributes. Then, when it tries to add the original set back in, `RequestAttributeChangeIgnoringWrapper` only adds `PATH_ATTRIBUTE` back in.
For the specific sample, the tests can be repaired by removing the `CorsFilter` or by exposing a custom `CorsConfigurationSource` bean since either of those will prevent `RequestAttributeChangeIgnoringWrapper` from wrapping the request.
I was also able to fix the tests by adding the following to `RequestAttributeChangeIgnoringWrapper`:
```java
@Override
public void removeAttribute(String name) {
if (name.equals(ServletRequestPathUtils.PATH_ATTRIBUTE) || name.equals(UrlPathHelper.PATH_ATTRIBUTE)) {
super.removeAttribute(name);
}
}
```
At least in this isolated case, it seems reasonable that if an attribute cannot be set, it also should not be able to be removed.
|
test
|
mockmvc s mvc result attribute lost with handlermappingintrospector and routerfunctions in use related to a contributor shared the following sample application the tests result in a nullpointerexception because the mockmvc mvc result attribute is missing java handlerexecutionchain chain super gethandler request if chain null defaultmvcresult mvcresult getmvcresult request returns null mvcresult sethandler chain gethandler mvcresult setinterceptors chain getinterceptors return chain it gets removed due to the following arrangement spring security s corsfilter by default uses handlermappingintrospector handlermappingintrospector uses requestattributechangeignoringwrapper which ignores all but path attribute requestpredicates restoreattributes attempts to restore the attributes to a previous state by clearing the attribute set and then re adding each attribute one by one before corsfilter runs mvc result attribute is present in the request when requestpredicates restoreattributes is run it removes all attributes then when it tries to add the original set back in requestattributechangeignoringwrapper only adds path attribute back in for the specific sample the tests can be repaired by removing the corsfilter or by exposing a custom corsconfigurationsource bean since either of those will prevent requestattributechangeignoringwrapper from wrapping the request i was also able to fix the tests by adding the following to requestattributechangeignoringwrapper java override public void removeattribute string name if name equals servletrequestpathutils path attribute name equals urlpathhelper path attribute super removeattribute name at least in this isolated case it seems reasonable that if an attribute cannot be set it also should not be able to be removed
| 1
|
307,646
| 23,209,893,949
|
IssuesEvent
|
2022-08-02 09:13:09
|
Shard-Labs/starknet-devnet
|
https://api.github.com/repos/Shard-Labs/starknet-devnet
|
closed
|
Error: Installing startnet-devnet on Windows
|
documentation enhancement help wanted
|
**Describe the bug**
I get this bug when I run the command `pip install starknet-devnet`

**Machine specifications**
- OS: Windows
- CPU: I5 ( Intel64 Family 6 Model 158 Stepping 9 GenuineIntel ~2500 Mhz)
- Memory: 16GB RAM
|
1.0
|
Error: Installing startnet-devnet on Windows - **Describe the bug**
I get this bug when I run the command `pip install starknet-devnet`

**Machine specifications**
- OS: Windows
- CPU: I5 ( Intel64 Family 6 Model 158 Stepping 9 GenuineIntel ~2500 Mhz)
- Memory: 16GB RAM
|
non_test
|
error installing startnet devnet on windows describe the bug i get this bug when i run the command pip install starknet devnet machine specifications os windows cpu family model stepping genuineintel mhz memory ram
| 0
|
52,810
| 13,771,901,607
|
IssuesEvent
|
2020-10-07 23:04:27
|
jgeraigery/please-del-this-repo
|
https://api.github.com/repos/jgeraigery/please-del-this-repo
|
opened
|
CVE-2017-16137 (Medium) detected in multiple libraries
|
security vulnerability
|
## CVE-2017-16137 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>debug-2.2.0.tgz</b>, <b>debug-2.3.3.tgz</b>, <b>debug-2.6.8.tgz</b>, <b>debug-2.6.0.tgz</b></p></summary>
<p>
<details><summary><b>debug-2.2.0.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p>
<p>Path to dependency file: please-del-this-repo/package.json</p>
<p>Path to vulnerable library: please-del-this-repo/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- express-4.13.4.tgz (Root Library)
- :x: **debug-2.2.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>debug-2.3.3.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.3.3.tgz">https://registry.npmjs.org/debug/-/debug-2.3.3.tgz</a></p>
<p>Path to dependency file: please-del-this-repo/package.json</p>
<p>Path to vulnerable library: please-del-this-repo/node_modules/socket.io-adapter/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- socket.io-1.7.3.tgz (Root Library)
- :x: **debug-2.3.3.tgz** (Vulnerable Library)
</details>
<details><summary><b>debug-2.6.8.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.8.tgz">https://registry.npmjs.org/debug/-/debug-2.6.8.tgz</a></p>
<p>Path to dependency file: please-del-this-repo/package.json</p>
<p>Path to vulnerable library: please-del-this-repo/node_modules/eslint-module-utils/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-1.11.0.tgz (Root Library)
- chokidar-1.7.0.tgz
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- tar-pack-3.4.0.tgz
- :x: **debug-2.6.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>debug-2.6.0.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.0.tgz">https://registry.npmjs.org/debug/-/debug-2.6.0.tgz</a></p>
<p>Path to dependency file: please-del-this-repo/package.json</p>
<p>Path to vulnerable library: please-del-this-repo/node_modules/mocha/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- mocha-3.4.2.tgz (Root Library)
- :x: **debug-2.6.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/please-del-this-repo/commit/e701d80cb459747be96e9c0e79c8640eca26c232">e701d80cb459747be96e9c0e79c8640eca26c232</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 2.6.9</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.2.0","isTransitiveDependency":true,"dependencyTree":"express:4.13.4;debug:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"},{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.3.3","isTransitiveDependency":true,"dependencyTree":"socket.io:1.7.3;debug:2.3.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"},{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.6.8","isTransitiveDependency":true,"dependencyTree":"nodemon:1.11.0;chokidar:1.7.0;fsevents:1.1.2;node-pre-gyp:0.6.36;tar-pack:3.4.0;debug:2.6.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"},{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.6.0","isTransitiveDependency":true,"dependencyTree":"mocha:3.4.2;debug:2.6.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"}],"vulnerabilityIdentifier":"CVE-2017-16137","vulnerabilityDetails":"The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2017-16137 (Medium) detected in multiple libraries - ## CVE-2017-16137 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>debug-2.2.0.tgz</b>, <b>debug-2.3.3.tgz</b>, <b>debug-2.6.8.tgz</b>, <b>debug-2.6.0.tgz</b></p></summary>
<p>
<details><summary><b>debug-2.2.0.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p>
<p>Path to dependency file: please-del-this-repo/package.json</p>
<p>Path to vulnerable library: please-del-this-repo/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- express-4.13.4.tgz (Root Library)
- :x: **debug-2.2.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>debug-2.3.3.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.3.3.tgz">https://registry.npmjs.org/debug/-/debug-2.3.3.tgz</a></p>
<p>Path to dependency file: please-del-this-repo/package.json</p>
<p>Path to vulnerable library: please-del-this-repo/node_modules/socket.io-adapter/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- socket.io-1.7.3.tgz (Root Library)
- :x: **debug-2.3.3.tgz** (Vulnerable Library)
</details>
<details><summary><b>debug-2.6.8.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.8.tgz">https://registry.npmjs.org/debug/-/debug-2.6.8.tgz</a></p>
<p>Path to dependency file: please-del-this-repo/package.json</p>
<p>Path to vulnerable library: please-del-this-repo/node_modules/eslint-module-utils/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-1.11.0.tgz (Root Library)
- chokidar-1.7.0.tgz
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- tar-pack-3.4.0.tgz
- :x: **debug-2.6.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>debug-2.6.0.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.6.0.tgz">https://registry.npmjs.org/debug/-/debug-2.6.0.tgz</a></p>
<p>Path to dependency file: please-del-this-repo/package.json</p>
<p>Path to vulnerable library: please-del-this-repo/node_modules/mocha/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- mocha-3.4.2.tgz (Root Library)
- :x: **debug-2.6.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/please-del-this-repo/commit/e701d80cb459747be96e9c0e79c8640eca26c232">e701d80cb459747be96e9c0e79c8640eca26c232</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 2.6.9</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.2.0","isTransitiveDependency":true,"dependencyTree":"express:4.13.4;debug:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"},{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.3.3","isTransitiveDependency":true,"dependencyTree":"socket.io:1.7.3;debug:2.3.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"},{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.6.8","isTransitiveDependency":true,"dependencyTree":"nodemon:1.11.0;chokidar:1.7.0;fsevents:1.1.2;node-pre-gyp:0.6.36;tar-pack:3.4.0;debug:2.6.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"},{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.6.0","isTransitiveDependency":true,"dependencyTree":"mocha:3.4.2;debug:2.6.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"}],"vulnerabilityIdentifier":"CVE-2017-16137","vulnerabilityDetails":"The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries debug tgz debug tgz debug tgz debug tgz debug tgz small debugging utility library home page a href path to dependency file please del this repo package json path to vulnerable library please del this repo node modules debug package json dependency hierarchy express tgz root library x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file please del this repo package json path to vulnerable library please del this repo node modules socket io adapter node modules debug package json dependency hierarchy socket io tgz root library x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file please del this repo package json path to vulnerable library please del this repo node modules eslint module utils node modules debug package json dependency hierarchy nodemon tgz root library chokidar tgz fsevents tgz node pre gyp tgz tar pack tgz x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file please del this repo package json path to vulnerable library please del this repo node modules mocha node modules debug package json dependency hierarchy mocha tgz root library x debug tgz vulnerable library found in head commit a href found in base branch master vulnerability details the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue vulnerabilityurl
| 0
|
51,815
| 6,198,568,817
|
IssuesEvent
|
2017-07-05 19:28:43
|
fedora-infra/bodhi
|
https://api.github.com/repos/fedora-infra/bodhi
|
closed
|
The pydocstyle tests do not work with tox
|
Tests
|
I noticed today that the pydocstyle tests do not actually check the paths that are given. I noticed that deleting a docblock from ```bodhi/server/views/generic.py``` doesn't cause any errors when btest is run, and adding all of bodhi to the ```match``` argument in ```tox.ini``` doesn't cause any failures (and should cause plenty).
|
1.0
|
The pydocstyle tests do not work with tox - I noticed today that the pydocstyle tests do not actually check the paths that are given. I noticed that deleting a docblock from ```bodhi/server/views/generic.py``` doesn't cause any errors when btest is run, and adding all of bodhi to the ```match``` argument in ```tox.ini``` doesn't cause any failures (and should cause plenty).
|
test
|
the pydocstyle tests do not work with tox i noticed today that the pydocstyle tests do not actually check the paths that are given i noticed that deleting a docblock from bodhi server views generic py doesn t cause any errors when btest is run and adding all of bodhi to the match argument in tox ini doesn t cause any failures and should cause plenty
| 1
|
256,039
| 27,552,556,975
|
IssuesEvent
|
2023-03-07 15:49:25
|
billmcchesney1/flow
|
https://api.github.com/repos/billmcchesney1/flow
|
closed
|
CVE-2022-41915 (Medium) detected in netty-codec-http-4.1.49.Final.jar - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2022-41915 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-codec-http-4.1.49.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-codec-http/4.1.49.Final/netty-codec-http-4.1.49.Final.jar</p>
<p>
Dependency Hierarchy:
- transport-7.9.3.jar (Root Library)
- transport-netty4-client-7.9.3.jar
- :x: **netty-codec-http-4.1.49.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/flow/commit/eb687271afab9d7c61ca82fce2ed4fdb3d5e1a70">eb687271afab9d7c61ca82fce2ed4fdb3d5e1a70</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Netty project is an event-driven asynchronous network application framework. Starting in version 4.1.83.Final and prior to 4.1.86.Final, when calling `DefaultHttpHeadesr.set` with an _iterator_ of values, header value validation was not performed, allowing malicious header values in the iterator to perform HTTP Response Splitting. This issue has been patched in version 4.1.86.Final. Integrators can work around the issue by changing the `DefaultHttpHeaders.set(CharSequence, Iterator<?>)` call, into a `remove()` call, and call `add()` in a loop over the iterator of values.
<p>Publish Date: 2022-12-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41915>CVE-2022-41915</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-12-13</p>
<p>Fix Resolution: io.netty:netty-codec-http:4.1.86.Final</p>
</p>
</details>
<p></p>
|
True
|
CVE-2022-41915 (Medium) detected in netty-codec-http-4.1.49.Final.jar - autoclosed - ## CVE-2022-41915 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-codec-http-4.1.49.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-codec-http/4.1.49.Final/netty-codec-http-4.1.49.Final.jar</p>
<p>
Dependency Hierarchy:
- transport-7.9.3.jar (Root Library)
- transport-netty4-client-7.9.3.jar
- :x: **netty-codec-http-4.1.49.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/flow/commit/eb687271afab9d7c61ca82fce2ed4fdb3d5e1a70">eb687271afab9d7c61ca82fce2ed4fdb3d5e1a70</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Netty project is an event-driven asynchronous network application framework. Starting in version 4.1.83.Final and prior to 4.1.86.Final, when calling `DefaultHttpHeadesr.set` with an _iterator_ of values, header value validation was not performed, allowing malicious header values in the iterator to perform HTTP Response Splitting. This issue has been patched in version 4.1.86.Final. Integrators can work around the issue by changing the `DefaultHttpHeaders.set(CharSequence, Iterator<?>)` call, into a `remove()` call, and call `add()` in a loop over the iterator of values.
<p>Publish Date: 2022-12-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41915>CVE-2022-41915</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-12-13</p>
<p>Fix Resolution: io.netty:netty-codec-http:4.1.86.Final</p>
</p>
</details>
<p></p>
|
non_test
|
cve medium detected in netty codec http final jar autoclosed cve medium severity vulnerability vulnerable library netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository io netty netty codec http final netty codec http final jar dependency hierarchy transport jar root library transport client jar x netty codec http final jar vulnerable library found in head commit a href found in base branch master vulnerability details netty project is an event driven asynchronous network application framework starting in version final and prior to final when calling defaulthttpheadesr set with an iterator of values header value validation was not performed allowing malicious header values in the iterator to perform http response splitting this issue has been patched in version final integrators can work around the issue by changing the defaulthttpheaders set charsequence iterator call into a remove call and call add in a loop over the iterator of values publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution io netty netty codec http final
| 0
|
242,515
| 20,252,599,487
|
IssuesEvent
|
2022-02-14 19:27:50
|
rspott/WAF-test02
|
https://api.github.com/repos/rspott/WAF-test02
|
opened
|
Discover and remediate common risks to improve Secure Score in Azure Security Center
|
WARP-Import test1 Security Security & Compliance Security Center
|
<a href="https://docs.microsoft.com/azure/architecture/framework/Security/governance#discover-and-remediate-common-risks">Discover and remediate common risks to improve Secure Score in Azure Security Center</a>
<p><b>Why Consider This?</b></p>
Secure Score is a snapshot of your security posture relative to Microsoft recommendations. Identifying and remediating common security hygiene risks significantly reduces overall risk to the organization by increasing cost to attackers.
<p><b>Context</b></p>
<p><span>When routine and well-established attack vectors are mitigated, attackers are forced to acquire and use advanced or untested attack methods. Azure Secure Score in Azure Security Center is one method to easily identify and address well-known issues in Azure.</span></p>
<p><b>Suggested Actions</b></p>
<p><span>Identify well-known risks for your Azure tenants, remediate those risks, and track your progress using Secure Score</span></p>
<p><b>Learn More</b></p>
<p><a href="https://docs.microsoft.com/en-us/azure/architecture/framework/Security/governance#discover-and-remediate-common-risks" target="_blank"><span>Discover and remediate common risks</span></a><span /></p>
|
1.0
|
Discover and remediate common risks to improve Secure Score in Azure Security Center - <a href="https://docs.microsoft.com/azure/architecture/framework/Security/governance#discover-and-remediate-common-risks">Discover and remediate common risks to improve Secure Score in Azure Security Center</a>
<p><b>Why Consider This?</b></p>
Secure Score is a snapshot of your security posture relative to Microsoft recommendations. Identifying and remediating common security hygiene risks significantly reduces overall risk to the organization by increasing cost to attackers.
<p><b>Context</b></p>
<p><span>When routine and well-established attack vectors are mitigated, attackers are forced to acquire and use advanced or untested attack methods. Azure Secure Score in Azure Security Center is one method to easily identify and address well-known issues in Azure.</span></p>
<p><b>Suggested Actions</b></p>
<p><span>Identify well-known risks for your Azure tenants, remediate those risks, and track your progress using Secure Score</span></p>
<p><b>Learn More</b></p>
<p><a href="https://docs.microsoft.com/en-us/azure/architecture/framework/Security/governance#discover-and-remediate-common-risks" target="_blank"><span>Discover and remediate common risks</span></a><span /></p>
|
test
|
discover and remediate common risks to improve secure score in azure security center why consider this secure score is a snapshot of your security posture relative to microsoft recommendations identifying and remediating common security hygiene risks significantly reduces overall risk to the organization by increasing cost to attackers context when routine and well established attack vectors are mitigated attackers are forced to acquire and use advanced or untested attack methods azure secure score in azure security center is one method to easily identify and address well known issues in azure suggested actions identify well known risks for your azure tenants remediate those risks and track your progress using secure score learn more discover and remediate common risks
| 1
|
80,447
| 10,013,920,848
|
IssuesEvent
|
2019-07-15 16:13:42
|
Kotlin/kotlinx.coroutines
|
https://api.github.com/repos/Kotlin/kotlinx.coroutines
|
opened
|
ReceiveChannel.asFlow operator
|
design flow
|
A missing piece of the Flow puzzle is a connection between regular `ReceiveChannel` and `Flow` (note that we already have a pretty non-obvious `ReceieveChannel.asFlux`) with a proper migration path for channel operators.
For example, all operators on top of the channels are deprecated, but users have no **clear** migration path from their channel operators sequence to flow.
The main focus here is to provide clear consumption semantics and decide whether we want to introduce such primitive at all
|
1.0
|
ReceiveChannel.asFlow operator - A missing piece of the Flow puzzle is a connection between regular `ReceiveChannel` and `Flow` (note that we already have a pretty non-obvious `ReceieveChannel.asFlux`) with a proper migration path for channel operators.
For example, all operators on top of the channels are deprecated, but users have no **clear** migration path from their channel operators sequence to flow.
The main focus here is to provide clear consumption semantics and decide whether we want to introduce such primitive at all
|
non_test
|
receivechannel asflow operator a missing piece of the flow puzzle is a connection between regular receivechannel and flow note that we already have a pretty non obvious receievechannel asflux with a proper migration path for channel operators for example all operators on top of the channels are deprecated but users have no clear migration path from their channel operators sequence to flow the main focus here is to provide clear consumption semantics and decide whether we want to introduce such primitive at all
| 0
|
200,199
| 15,093,543,901
|
IssuesEvent
|
2021-02-07 01:09:33
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
Out of order tags across active branches are causing tests to fail
|
kind/failing-test priority/critical-urgent sig/release triage/accepted
|
_Note: this is the same issue we encountered in https://github.com/kubernetes/kubernetes/issues/86182._
For the `v1.20.0` release, as well as the subsequent patch releases for `v1.17.15` / `v1.18.13` / `v1.19.5`, the release version was tagged either on the same or a later commit than the `rc.0` tag, causing subsequent builds to fail the version regex check in `get-kube.sh`. As summarized in @justaugustus [email to k-dev](https://groups.google.com/g/kubernetes-dev/c/angCw-zHejE/discussion) the last time this happened:
> We use a combination of git describe and regexes (of course) to determine the tag and whether or not the tag is a CI tag. As the tags landed on the same commit, our git describe was ambiguous and picked up something in the form of v1.y.z-<number-of-commits-past-tag>+<commit-ish>, which makes our CI version regex unhappy.
This will likely need to be solved by manually tagging new versions. This issue thread will be updated with progress towards resolution.
An example test failure: https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-cos-k8sbeta-default/1336844198506139648
_Very long_ slack thread from last time this happened with rationale behind doing manual tag and push: https://kubernetes.slack.com/archives/CJH2GBF7Y/p1576268807016400
/cc @kubernetes/release-managers
/assign
/assign @justaugustus
/priority critical-urgent
|
1.0
|
Out of order tags across active branches are causing tests to fail - _Note: this is the same issue we encountered in https://github.com/kubernetes/kubernetes/issues/86182._
For the `v1.20.0` release, as well as the subsequent patch releases for `v1.17.15` / `v1.18.13` / `v1.19.5`, the release version was tagged either on the same or a later commit than the `rc.0` tag, causing subsequent builds to fail the version regex check in `get-kube.sh`. As summarized in @justaugustus [email to k-dev](https://groups.google.com/g/kubernetes-dev/c/angCw-zHejE/discussion) the last time this happened:
> We use a combination of git describe and regexes (of course) to determine the tag and whether or not the tag is a CI tag. As the tags landed on the same commit, our git describe was ambiguous and picked up something in the form of v1.y.z-<number-of-commits-past-tag>+<commit-ish>, which makes our CI version regex unhappy.
This will likely need to be solved by manually tagging new versions. This issue thread will be updated with progress towards resolution.
An example test failure: https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-cos-k8sbeta-default/1336844198506139648
_Very long_ slack thread from last time this happened with rationale behind doing manual tag and push: https://kubernetes.slack.com/archives/CJH2GBF7Y/p1576268807016400
/cc @kubernetes/release-managers
/assign
/assign @justaugustus
/priority critical-urgent
|
test
|
out of order tags across active branches are causing tests to fail note this is the same issue we encountered in for the release as well as the subsequent patch releases for the release version was tagged either on the same or a later commit than the rc tag causing subsequent builds to fail the version regex check in get kube sh as summarized in justaugustus the last time this happened we use a combination of git describe and regexes of course to determine the tag and whether or not the tag is a ci tag as the tags landed on the same commit our git describe was ambiguous and picked up something in the form of y z which makes our ci version regex unhappy this will likely need to be solved by manually tagging new versions this issue thread will be updated with progress towards resolution an example test failure very long slack thread from last time this happened with rationale behind doing manual tag and push cc kubernetes release managers assign assign justaugustus priority critical urgent
| 1
|
233,091
| 18,947,335,290
|
IssuesEvent
|
2021-11-18 11:37:27
|
hyperledger-labs/go-perun
|
https://api.github.com/repos/hyperledger-labs/go-perun
|
opened
|
[sim/wallet] Address.String() length
|
good first issue testing-only
|
### Location
`[backend/sim/wallet] Address.String()` function
Function only returns the first 4 bytes as string for easier comparability in test output.
### Problem
>I am not really convinced that shortening the string representation is a good idea. Your debug sessions may be easier to read, but on the other hand it may create confusion. In any case, this function could be a one-liner: return fmt.Sprintf("0x%x...", a.X.Bytes()[:length]).
from @matthiasgeihs https://github.com/hyperledger-labs/go-perun/pull/260#discussion_r751680877
### Proposal
Change to full length.
|
1.0
|
[sim/wallet] Address.String() length - ### Location
`[backend/sim/wallet] Address.String()` function
Function only returns the first 4 bytes as string for easier comparability in test output.
### Problem
>I am not really convinced that shortening the string representation is a good idea. Your debug sessions may be easier to read, but on the other hand it may create confusion. In any case, this function could be a one-liner: return fmt.Sprintf("0x%x...", a.X.Bytes()[:length]).
from @matthiasgeihs https://github.com/hyperledger-labs/go-perun/pull/260#discussion_r751680877
### Proposal
Change to full length.
|
test
|
address string length location address string function function only returns the first bytes as string for easier comparability in test output problem i am not really convinced that shortening the string representation is a good idea your debug sessions may be easier to read but on the other hand it may create confusion in any case this function could be a one liner return fmt sprintf x a x bytes from matthiasgeihs proposal change to full length
| 1
|
93,788
| 15,932,116,301
|
IssuesEvent
|
2021-04-14 05:05:02
|
samq-wsdemo/SecurityShepherd
|
https://api.github.com/repos/samq-wsdemo/SecurityShepherd
|
opened
|
CVE-2020-2933 (Low) detected in mysql-connector-java-5.1.24.jar
|
security vulnerability
|
## CVE-2020-2933 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.24.jar</b></p></summary>
<p>MySQL JDBC Type 4 driver</p>
<p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p>
<p>Path to dependency file: SecurityShepherd/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/mysql/mysql-connector-java/5.1.24/mysql-connector-java-5.1.24.jar,SecurityShepherd/target/owaspSecurityShepherd/WEB-INF/lib/mysql-connector-java-5.1.24.jar</p>
<p>
Dependency Hierarchy:
- :x: **mysql-connector-java-5.1.24.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-wsdemo/SecurityShepherd/commit/00289f7750953c5aab43147df68b8dcf4063441b">00289f7750953c5aab43147df68b8dcf4063441b</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 5.1.48 and prior. Difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 2.2 (Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:H/UI:N/S:U/C:N/I:N/A:L).
<p>Publish Date: 2020-04-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2933>CVE-2020-2933</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://docs.oracle.com/javase/7/docs/api/javax/xml/XMLConstants.html#FEATURE_SECURE_PROCESSING">https://docs.oracle.com/javase/7/docs/api/javax/xml/XMLConstants.html#FEATURE_SECURE_PROCESSING</a></p>
<p>Release Date: 2020-04-15</p>
<p>Fix Resolution: mysql:mysql-connector-java:5.1.49</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.24","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.24","isMinimumFixVersionAvailable":true,"minimumFixVersion":"mysql:mysql-connector-java:5.1.49"}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-2933","vulnerabilityDetails":"Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 5.1.48 and prior. Difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 2.2 (Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:H/UI:N/S:U/C:N/I:N/A:L).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2933","cvss3Severity":"low","cvss3Score":"2.2","cvss3Metrics":{"A":"Low","AC":"High","PR":"High","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-2933 (Low) detected in mysql-connector-java-5.1.24.jar - ## CVE-2020-2933 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.24.jar</b></p></summary>
<p>MySQL JDBC Type 4 driver</p>
<p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p>
<p>Path to dependency file: SecurityShepherd/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/mysql/mysql-connector-java/5.1.24/mysql-connector-java-5.1.24.jar,SecurityShepherd/target/owaspSecurityShepherd/WEB-INF/lib/mysql-connector-java-5.1.24.jar</p>
<p>
Dependency Hierarchy:
- :x: **mysql-connector-java-5.1.24.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-wsdemo/SecurityShepherd/commit/00289f7750953c5aab43147df68b8dcf4063441b">00289f7750953c5aab43147df68b8dcf4063441b</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 5.1.48 and prior. Difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 2.2 (Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:H/UI:N/S:U/C:N/I:N/A:L).
<p>Publish Date: 2020-04-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2933>CVE-2020-2933</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://docs.oracle.com/javase/7/docs/api/javax/xml/XMLConstants.html#FEATURE_SECURE_PROCESSING">https://docs.oracle.com/javase/7/docs/api/javax/xml/XMLConstants.html#FEATURE_SECURE_PROCESSING</a></p>
<p>Release Date: 2020-04-15</p>
<p>Fix Resolution: mysql:mysql-connector-java:5.1.49</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.24","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.24","isMinimumFixVersionAvailable":true,"minimumFixVersion":"mysql:mysql-connector-java:5.1.49"}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-2933","vulnerabilityDetails":"Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 5.1.48 and prior. Difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 2.2 (Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:H/UI:N/S:U/C:N/I:N/A:L).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2933","cvss3Severity":"low","cvss3Score":"2.2","cvss3Metrics":{"A":"Low","AC":"High","PR":"High","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve low detected in mysql connector java jar cve low severity vulnerability vulnerable library mysql connector java jar mysql jdbc type driver library home page a href path to dependency file securityshepherd pom xml path to vulnerable library canner repository mysql mysql connector java mysql connector java jar securityshepherd target owaspsecurityshepherd web inf lib mysql connector java jar dependency hierarchy x mysql connector java jar vulnerable library found in head commit a href found in base branch dev vulnerability details vulnerability in the mysql connectors product of oracle mysql component connector j supported versions that are affected are and prior difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise mysql connectors successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service partial dos of mysql connectors cvss base score availability impacts cvss vector cvss av n ac h pr h ui n s u c n i n a l publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution mysql mysql connector java rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree mysql mysql connector java isminimumfixversionavailable true minimumfixversion mysql mysql connector java basebranches vulnerabilityidentifier cve vulnerabilitydetails vulnerability in the mysql connectors product of oracle mysql component connector j supported versions that are affected are and prior difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise mysql connectors successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service partial dos of mysql connectors cvss base score availability impacts cvss vector cvss av n ac h pr h ui n s u c n i n a l vulnerabilityurl
| 0
|
164,154
| 6,219,864,529
|
IssuesEvent
|
2017-07-09 17:31:03
|
tferreira/piggydime
|
https://api.github.com/repos/tferreira/piggydime
|
closed
|
Dockerize the app
|
deployment priority - medium
|
For easier deployment and best support of a wide range of operating systems, it would be necessary to create mutiple Docker images:
- The python server (maybe using uWSGI and Nginx)
- The React app
- A MySQL DB (using volumes to keep the datas)
Using docker-compose is probably the best way to achieve this.
Also see: [https://github.com/geerlingguy/docker-examples/tree/master/flask](https://github.com/geerlingguy/docker-examples/tree/master/flask)
|
1.0
|
Dockerize the app - For easier deployment and best support of a wide range of operating systems, it would be necessary to create mutiple Docker images:
- The python server (maybe using uWSGI and Nginx)
- The React app
- A MySQL DB (using volumes to keep the datas)
Using docker-compose is probably the best way to achieve this.
Also see: [https://github.com/geerlingguy/docker-examples/tree/master/flask](https://github.com/geerlingguy/docker-examples/tree/master/flask)
|
non_test
|
dockerize the app for easier deployment and best support of a wide range of operating systems it would be necessary to create mutiple docker images the python server maybe using uwsgi and nginx the react app a mysql db using volumes to keep the datas using docker compose is probably the best way to achieve this also see
| 0
|
380,786
| 26,431,097,649
|
IssuesEvent
|
2023-01-14 20:36:58
|
Luanftg/codigo-do-futuro-api-dotnet
|
https://api.github.com/repos/Luanftg/codigo-do-futuro-api-dotnet
|
closed
|
Criação da Conta Google
|
documentation
|
### 9-Registrar a chave para utilizar a Api do Google Maps
- [ ] Criar conta no google Cloud
- [ ] Cria Key para APIs do google maps
- [ ] Configurar os acessos das APIs
### Texto Auxiliar
[Documentação de Referência APi Google Maps](https://developers.google.com/maps/documentation/maps-static/start?hl=pt-br#Addresses)
|
1.0
|
Criação da Conta Google - ### 9-Registrar a chave para utilizar a Api do Google Maps
- [ ] Criar conta no google Cloud
- [ ] Cria Key para APIs do google maps
- [ ] Configurar os acessos das APIs
### Texto Auxiliar
[Documentação de Referência APi Google Maps](https://developers.google.com/maps/documentation/maps-static/start?hl=pt-br#Addresses)
|
non_test
|
criação da conta google registrar a chave para utilizar a api do google maps criar conta no google cloud cria key para apis do google maps configurar os acessos das apis texto auxiliar
| 0
|
68,322
| 7,093,010,315
|
IssuesEvent
|
2018-01-12 18:43:59
|
truetandem/e-QIP-prototype
|
https://api.github.com/repos/truetandem/e-QIP-prototype
|
opened
|
Safari - Foreign activities - review screen
|
cross browser testing - safari
|
Foreign activities - review screen dropped user midway down the screen instead of at top.

|
1.0
|
Safari - Foreign activities - review screen - Foreign activities - review screen dropped user midway down the screen instead of at top.

|
test
|
safari foreign activities review screen foreign activities review screen dropped user midway down the screen instead of at top
| 1
|
87,569
| 8,100,760,973
|
IssuesEvent
|
2018-08-12 03:41:53
|
codercom/bugs
|
https://api.github.com/repos/codercom/bugs
|
closed
|
Project data not removed when the project is deleted.
|
bug testing tracking
|
**Describe the bug**
Data is not removed when a project is deleted.
**To Reproduce**
Steps to reproduce the behavior:
1. Create a project.
2. Create a file in the project.
3. Delete the project.
4. Create a new project with the exact same name and the files will still be there.
**Expected behavior**
A new project would be created without old data as it is supposed to delete when the project is deleted.
**Desktop (please complete the following information):**
- OS: Windows 10
- Browser: Google Chrome
- Version: 67.0.3396.99 (Official Build) (64-bit)
|
1.0
|
Project data not removed when the project is deleted. - **Describe the bug**
Data is not removed when a project is deleted.
**To Reproduce**
Steps to reproduce the behavior:
1. Create a project.
2. Create a file in the project.
3. Delete the project.
4. Create a new project with the exact same name and the files will still be there.
**Expected behavior**
A new project would be created without old data as it is supposed to delete when the project is deleted.
**Desktop (please complete the following information):**
- OS: Windows 10
- Browser: Google Chrome
- Version: 67.0.3396.99 (Official Build) (64-bit)
|
test
|
project data not removed when the project is deleted describe the bug data is not removed when a project is deleted to reproduce steps to reproduce the behavior create a project create a file in the project delete the project create a new project with the exact same name and the files will still be there expected behavior a new project would be created without old data as it is supposed to delete when the project is deleted desktop please complete the following information os windows browser google chrome version official build bit
| 1
|
31,830
| 6,643,248,573
|
IssuesEvent
|
2017-09-27 10:31:57
|
contao/core-bundle
|
https://api.github.com/repos/contao/core-bundle
|
closed
|
Important part cannot be set within picker overlay
|
defect
|
If you go to the file manager within the picker overlay and edit an image, you are not able to set the important part with your cursor anymore. Instead, only the input fields to manually set the important part appear.
Possibly related to #1092 ?
|
1.0
|
Important part cannot be set within picker overlay - If you go to the file manager within the picker overlay and edit an image, you are not able to set the important part with your cursor anymore. Instead, only the input fields to manually set the important part appear.
Possibly related to #1092 ?
|
non_test
|
important part cannot be set within picker overlay if you go to the file manager within the picker overlay and edit an image you are not able to set the important part with your cursor anymore instead only the input fields to manually set the important part appear possibly related to
| 0
|
699,548
| 24,020,807,536
|
IssuesEvent
|
2022-09-15 07:29:11
|
PrefectHQ/prefect
|
https://api.github.com/repos/PrefectHQ/prefect
|
closed
|
Only one default parameter is deployed when multiple are specified (in deployment.py)
|
bug status:backlog priority:high component:deployment
|
### First check
- [X] I added a descriptive title to this issue.
- [X] I used the GitHub search to find a similar issue and didn't find it.
- [X] I searched the Prefect documentation for this issue.
- [X] I checked that this issue is related to Prefect and not one of its dependencies.
### Bug summary
I am scheduling a dbt flow with Prefect. I tried to specify two parameters in my deployment.py file, but only the first of them is stored in the database after deployment. I have not tried deployment via CLI/yaml yet, but though this is already worth mentioning.
### Reproduction
```python
# deployment.py
from dbt_debug import dbt_debug_flow
from prefect.deployments import Deployment
deployment = Deployment.build_from_flow(
flow=dbt_debug_flow,
description ="Testing a prefect-based dbt workflow.",
name="dbt-debug-test",
parameters={'file_path': 'C:/Users/Public/File',
'project_dir':'C:/Users/User/Documents/Git/Project/dbt_dev'},
infra_overrides={"env": {"PREFECT_LOGGING_LEVEL": "DEBUG"}},
work_queue_name="test",
)
if __name__ == "__main__":
deployment.apply()
# deploy via python deployment.py
```
### Error
# On Prefect Orion

```
Flow run 'cuddly-lyrebird' - Validation of flow parameters failed with error: SignatureMismatchError("Function expects parameters ['file_path'] but was provided with parameters ['file_path', 'project_dir']")
```
### Versions
```
Version: 2.4.0
API version: 0.8.0
Python version: 3.10.7
Git commit: 513639e8
Built: Tue, Sep 13, 2022 2:15 PM
OS/Arch: win32/AMD64
Profile: default
Server type: ephemeral
Server:
Database: sqlite
SQLite version: 3.37.2
```
### Additional context
_No response_
|
1.0
|
Only one default parameter is deployed when multiple are specified (in deployment.py) - ### First check
- [X] I added a descriptive title to this issue.
- [X] I used the GitHub search to find a similar issue and didn't find it.
- [X] I searched the Prefect documentation for this issue.
- [X] I checked that this issue is related to Prefect and not one of its dependencies.
### Bug summary
I am scheduling a dbt flow with Prefect. I tried to specify two parameters in my deployment.py file, but only the first of them is stored in the database after deployment. I have not tried deployment via CLI/yaml yet, but though this is already worth mentioning.
### Reproduction
```python
# deployment.py
from dbt_debug import dbt_debug_flow
from prefect.deployments import Deployment
deployment = Deployment.build_from_flow(
flow=dbt_debug_flow,
description ="Testing a prefect-based dbt workflow.",
name="dbt-debug-test",
parameters={'file_path': 'C:/Users/Public/File',
'project_dir':'C:/Users/User/Documents/Git/Project/dbt_dev'},
infra_overrides={"env": {"PREFECT_LOGGING_LEVEL": "DEBUG"}},
work_queue_name="test",
)
if __name__ == "__main__":
deployment.apply()
# deploy via python deployment.py
```
### Error
# On Prefect Orion

```
Flow run 'cuddly-lyrebird' - Validation of flow parameters failed with error: SignatureMismatchError("Function expects parameters ['file_path'] but was provided with parameters ['file_path', 'project_dir']")
```
### Versions
```
Version: 2.4.0
API version: 0.8.0
Python version: 3.10.7
Git commit: 513639e8
Built: Tue, Sep 13, 2022 2:15 PM
OS/Arch: win32/AMD64
Profile: default
Server type: ephemeral
Server:
Database: sqlite
SQLite version: 3.37.2
```
### Additional context
_No response_
|
non_test
|
only one default parameter is deployed when multiple are specified in deployment py first check i added a descriptive title to this issue i used the github search to find a similar issue and didn t find it i searched the prefect documentation for this issue i checked that this issue is related to prefect and not one of its dependencies bug summary i am scheduling a dbt flow with prefect i tried to specify two parameters in my deployment py file but only the first of them is stored in the database after deployment i have not tried deployment via cli yaml yet but though this is already worth mentioning reproduction python deployment py from dbt debug import dbt debug flow from prefect deployments import deployment deployment deployment build from flow flow dbt debug flow description testing a prefect based dbt workflow name dbt debug test parameters file path c users public file project dir c users user documents git project dbt dev infra overrides env prefect logging level debug work queue name test if name main deployment apply deploy via python deployment py error on prefect orion flow run cuddly lyrebird validation of flow parameters failed with error signaturemismatcherror function expects parameters but was provided with parameters versions version api version python version git commit built tue sep pm os arch profile default server type ephemeral server database sqlite sqlite version additional context no response
| 0
|
807,332
| 29,995,841,602
|
IssuesEvent
|
2023-06-26 05:24:13
|
KingSupernova31/RulesGuru
|
https://api.github.com/repos/KingSupernova31/RulesGuru
|
opened
|
Talk to Kyle Ryc about their phone issues
|
bug medium priority
|
Canadian L2 Kyle Ryc was telling me about several issues they were having with RG on their phone. I or someone else needs to reach out to them, get a list of the issues, and fix them.
|
1.0
|
Talk to Kyle Ryc about their phone issues - Canadian L2 Kyle Ryc was telling me about several issues they were having with RG on their phone. I or someone else needs to reach out to them, get a list of the issues, and fix them.
|
non_test
|
talk to kyle ryc about their phone issues canadian kyle ryc was telling me about several issues they were having with rg on their phone i or someone else needs to reach out to them get a list of the issues and fix them
| 0
|
57,034
| 3,081,232,950
|
IssuesEvent
|
2015-08-22 14:21:05
|
bitfighter/bitfighter
|
https://api.github.com/repos/bitfighter/bitfighter
|
closed
|
Multi-line text items
|
020 bug imported Priority-Medium
|
_From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on October 23, 2013 07:40:11_
Allow users to enter multi-line text items, with line breaks specified with \n.
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=265_
|
1.0
|
Multi-line text items - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on October 23, 2013 07:40:11_
Allow users to enter multi-line text items, with line breaks specified with \n.
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=265_
|
non_test
|
multi line text items from on october allow users to enter multi line text items with line breaks specified with n original issue
| 0
|
424,205
| 12,307,575,547
|
IssuesEvent
|
2020-05-12 05:03:21
|
googleapis/java-billing
|
https://api.github.com/repos/googleapis/java-billing
|
opened
|
Synthesis failed for java-billing
|
autosynth failure priority: p1 type: bug
|
Hello! Autosynth couldn't regenerate java-billing. :broken_heart:
Here's the output from running `synth.py`:
```
2020-05-11 21:52:44,383 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/java-billing
2020-05-11 21:52:44,900 autosynth > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore
2020-05-11 21:52:44,905 autosynth > Running: git config user.name yoshi-automation
2020-05-11 21:52:44,910 autosynth > Running: git config user.email yoshi-automation@google.com
2020-05-11 21:52:44,915 autosynth > Running: git config push.default simple
2020-05-11 21:52:44,920 autosynth > Running: git branch -f autosynth
2020-05-11 21:52:44,926 autosynth > Running: git checkout autosynth
2020-05-11 21:52:44,939 autosynth > autosynth flags:
2020-05-11 21:52:44,940 autosynth > AUTOSYNTH_MULTIPLE_COMMITS: true
2020-05-11 21:52:44,940 autosynth > AUTOSYNTH_MULTIPLE_PRS: true
2020-05-11 21:52:44,941 autosynth > Running: git rev-parse --show-toplevel
2020-05-11 21:52:44,946 autosynth > Running: git log -1 --pretty=%H
2020-05-11 21:52:44,951 autosynth > Running: git remote get-url origin
2020-05-11 21:52:44,956 autosynth > Running: git clone --single-branch https://github.com/googleapis/googleapis.git -- /tmpfs/tmp/tmpaaavgl1n/googleapis
2020-05-11 21:52:46,450 autosynth > Running: git log d741cd976975c745d0199987aff0e908b8352992..HEAD --pretty=%H --no-decorate
2020-05-11 21:52:46,457 autosynth > Running: git clone --single-branch https://github.com/googleapis/synthtool.git -- /tmpfs/tmp/tmpaaavgl1n/synthtool
2020-05-11 21:52:47,010 autosynth > Running: git log f8a9933e5e98202b04ef427f28d1d79221190fa4..HEAD --pretty=%H --no-decorate
2020-05-11 21:52:47,017 autosynth > Running: git log -1 --pretty=%at d741cd976975c745d0199987aff0e908b8352992
2020-05-11 21:52:47,022 autosynth > Running: git log -1 --pretty=%at e4f4b23e07315492b533746e6a9255a1e6b3e748
2020-05-11 21:52:47,028 autosynth > Running: git log -1 --pretty=%at ee4ea76504aa60c2bff9b7c11269c155d8c21e0d
2020-05-11 21:52:47,033 autosynth > Running: git log -1 --pretty=%at 6114e2d30c0792522720d1c2cf8f5e111be7545b
2020-05-11 21:52:47,038 autosynth > Running: git log -1 --pretty=%at af71ea52236b05ce73253f6bc935fcf35c866664
2020-05-11 21:52:47,044 autosynth > Running: git log -1 --pretty=%at c451a4457f38e5582e9794a0094c773355a9eaad
2020-05-11 21:52:47,050 autosynth > Running: git log -1 --pretty=%at 9af7212bc2cba8a7184f032e1ab6662ccf604b3a
2020-05-11 21:52:47,055 autosynth > Running: git log -1 --pretty=%at ac1f904979899bc598fb0567cf7a367f08b40b43
2020-05-11 21:52:47,061 autosynth > Running: git log -1 --pretty=%at fe8b17aff5b419ffb418a77228fa049190488233
2020-05-11 21:52:47,067 autosynth > Running: git log -1 --pretty=%at fd35dc6865fa9f47f283d5a6c37c4901162007cf
2020-05-11 21:52:47,072 autosynth > Running: git log -1 --pretty=%at 96c0db42dbcb3bb0116cfa55c054dd139ffb3912
2020-05-11 21:52:47,077 autosynth > Running: git log -1 --pretty=%at b8c7c4e8da56547fc32bee9b563d306fadabdb6f
2020-05-11 21:52:47,082 autosynth > Running: git log -1 --pretty=%at b21bc9b46e3263cd4ebe50f60c61589e3fe797f2
2020-05-11 21:52:47,087 autosynth > Running: git log -1 --pretty=%at 41d5dee23527df3fc9127838a027df393d79ff3c
2020-05-11 21:52:47,092 autosynth > Running: git log -1 --pretty=%at 801c3fb47b1a533989a1fe5af4aedf02af5f2eae
2020-05-11 21:52:47,097 autosynth > Running: git log -1 --pretty=%at 1692c3d51ee7d4366ce81892df1b9ca9846480d3
2020-05-11 21:52:47,102 autosynth > Running: git log -1 --pretty=%at 1df336ab4a989e7b0bc363baa9fc91eb66b0625e
2020-05-11 21:52:47,107 autosynth > Running: git log -1 --pretty=%at 1e92c343790ae63f88817afde60bd2d5bc775f36
2020-05-11 21:52:47,111 autosynth > Running: git log -1 --pretty=%at cf18ab5200e41e9d0cfef88c6c4bcdd394faf1ca
2020-05-11 21:52:47,116 autosynth > Running: git log -1 --pretty=%at ab869e90c6ced0ef30fcc8e8fbe8e527a695d3cb
2020-05-11 21:52:47,121 autosynth > Running: git log -1 --pretty=%at 13736ead90e0fa65103eb03fea775a1c010564af
2020-05-11 21:52:47,126 autosynth > Running: git log -1 --pretty=%at f9d76ac088f5d444b16d5e17fe58a9eea57b7a99
2020-05-11 21:52:47,131 autosynth > Running: git log -1 --pretty=%at 1bd77e8ce6f953ac641af7966d0c52646afc16a8
2020-05-11 21:52:47,136 autosynth > Running: git log -1 --pretty=%at 3824f547aa44df459580615c73cabb30a2a78ee0
2020-05-11 21:52:47,141 autosynth > Running: git log -1 --pretty=%at 4d61e1cb40184a7ad63ef37b1813f6608718674a
2020-05-11 21:52:47,147 autosynth > Running: git log -1 --pretty=%at 9ea892ebe8a7cd66da9bd62eb7086b8dc2ddd559
2020-05-11 21:52:47,152 autosynth > Running: git log -1 --pretty=%at 0feb08065d91d6deebffef3019751f0619a72140
2020-05-11 21:52:47,157 autosynth > Running: git log -1 --pretty=%at 789fd7e0680b04a589447139811d55c5d1f34afe
2020-05-11 21:52:47,163 autosynth > Running: git log -1 --pretty=%at 4efd71279f0781579ced60c374336ce1b8b8cea8
2020-05-11 21:52:47,168 autosynth > Running: git log -1 --pretty=%at a34f4eb96a94bfd5b3e7f8de253db28d0a3e9237
2020-05-11 21:52:47,173 autosynth > Running: git log -1 --pretty=%at f561f57eea013327be27d8f6799b3cc2d18ee804
2020-05-11 21:52:47,177 autosynth > Running: git log -1 --pretty=%at c9942c00e8f35b5b37cd589ef7a0323aeda264e2
2020-05-11 21:52:47,182 autosynth > Running: git log -1 --pretty=%at e2a59ff2f861b1b92b2f15b242780dccddd77ab9
2020-05-11 21:52:47,188 autosynth > Running: git log -1 --pretty=%at df955af5d9508cf9215de6f1fbdad033d5b8fef6
2020-05-11 21:52:47,193 autosynth > Running: git log -1 --pretty=%at dea1f52034d188ca371c3a70c716ca2a6b3f2aa1
2020-05-11 21:52:47,198 autosynth > Running: git log -1 --pretty=%at 35672e7936b923eff048d0ee1529faa163cc81b8
2020-05-11 21:52:47,203 autosynth > Running: git log -1 --pretty=%at 26523a96798ce1a6caa1b3c912119059cfcc98a7
2020-05-11 21:52:47,208 autosynth > Running: git log -1 --pretty=%at 04eff659376f40a7111bf0e8e49fbb38bc7739f7
2020-05-11 21:52:47,213 autosynth > Running: git log -1 --pretty=%at 3028060618e8024af9a32b3ab3456c160091ecb7
2020-05-11 21:52:47,218 autosynth > Running: git log -1 --pretty=%at b5aa7d668d9513cf0d305c8b23262ee0ea9b850c
2020-05-11 21:52:47,223 autosynth > Running: git log -1 --pretty=%at c0a072159eda3c7d5a98201d6500feaf78a18073
2020-05-11 21:52:47,228 autosynth > Running: git log -1 --pretty=%at fea22b1d9f27f86ef355c1d0dba00e0791a08a19
2020-05-11 21:52:47,233 autosynth > Running: git log -1 --pretty=%at 80dac49dd5f407b14f15ac86b4922967b5dd6508
2020-05-11 21:52:47,238 autosynth > Running: git log -1 --pretty=%at b8495354535dc1370377cbb15a3ef7a07ec3a114
2020-05-11 21:52:47,243 autosynth > Running: git log -1 --pretty=%at ab5d07c2883026eec395d73ff59a17f4ad0153db
2020-05-11 21:52:47,249 autosynth > Running: git log -1 --pretty=%at 160e3187dc93265ed635ad68e12c483c8ab4f7be
2020-05-11 21:52:47,254 autosynth > Running: git log -1 --pretty=%at 629e6bea8418ad4a5d081ab9eb48bb327ad368e1
2020-05-11 21:52:47,259 autosynth > Running: git log -1 --pretty=%at c3f3cdeda70f60d808eaaae8fea44d306b6d5cc2
2020-05-11 21:52:47,264 autosynth > Running: git log -1 --pretty=%at cdf13efacdea0649e940452f9c5d320b93735974
2020-05-11 21:52:47,269 autosynth > Running: git log -1 --pretty=%at 101d31acd73076c52d78e18322be01f3debe8cb5
2020-05-11 21:52:47,274 autosynth > Running: git log -1 --pretty=%at cab24c7c9b570e2b483e4f742a7a4a111167238b
2020-05-11 21:52:47,279 autosynth > Running: git log -1 --pretty=%at 6e0ac551e057e20a6c41ab0b21ffa9623e463326
2020-05-11 21:52:47,285 autosynth > Running: git log -1 --pretty=%at 69b4dc56d21394e7c39bc62fe11c76a88aa5511f
2020-05-11 21:52:47,290 autosynth > Running: git log -1 --pretty=%at bcc476396e799806d3355e87246c6becf6250a70
2020-05-11 21:52:47,295 autosynth > Running: git log -1 --pretty=%at b5e9b0b09cc15c263f07da92fa64db31b2e4dd07
2020-05-11 21:52:47,300 autosynth > Running: git log -1 --pretty=%at c09ec5156db66d326cce56ed3a2a8092e02c3abd
2020-05-11 21:52:47,306 autosynth > Running: git log -1 --pretty=%at 42ee97c1b93a0e3759bbba3013da309f670a90ab
2020-05-11 21:52:47,310 autosynth > Running: git log -1 --pretty=%at 45b363d3ab4692115cc5018eb26dcdaa49b320a8
2020-05-11 21:52:47,315 autosynth > Running: git log -1 --pretty=%at 012c553112fcceb7b69822e24335575bf99ec62b
2020-05-11 21:52:47,320 autosynth > Running: git log -1 --pretty=%at 2bf9f22da3f101aeb536e1e88fe036323b8e9188
2020-05-11 21:52:47,326 autosynth > Running: git log -1 --pretty=%at 862cd840dbb0101a0630913217bf8b6d4101c84c
2020-05-11 21:52:47,331 autosynth > Running: git log -1 --pretty=%at b3bcfd059ec684445a38db693f6e36f14a1d2618
2020-05-11 21:52:47,336 autosynth > Running: git log -1 --pretty=%at 1d520eaa7bbd8b40f53942ff03b5496fff887b53
2020-05-11 21:52:47,341 autosynth > Running: git log -1 --pretty=%at 0d7c3565dd942d0cc41b16e895046486ae29e44e
2020-05-11 21:52:47,347 autosynth > Running: git log -1 --pretty=%at d14bf59a446c14ef16e9931ebfc8e63ab549bf07
2020-05-11 21:52:47,352 autosynth > Running: git log -1 --pretty=%at 28e76243c23cc282efbb288cb558c174e3e5e9ee
2020-05-11 21:52:47,357 autosynth > Running: git log -1 --pretty=%at 5dca7f1079a4c127245c0f1dbb8293b434fbd0c8
2020-05-11 21:52:47,363 autosynth > Running: git log -1 --pretty=%at c2bfceedfc9f01356229bd206ad25e250aa9551c
2020-05-11 21:52:47,368 autosynth > Running: git log -1 --pretty=%at 371206424e6db4df2cb3734bb1013dcaac9dbe4b
2020-05-11 21:52:47,373 autosynth > Running: git log -1 --pretty=%at ff17eefa184f6985e5d9a8ee3e725da95a06cd23
2020-05-11 21:52:47,379 autosynth > Running: git log -1 --pretty=%at c284e4f849e4b085a297d336cef8721ab69ba013
2020-05-11 21:52:47,384 autosynth > Running: git log -1 --pretty=%at 756b174de4a122461993c1c583345533d819936d
2020-05-11 21:52:47,390 autosynth > Running: git log -1 --pretty=%at 3f5f8a2258c6a41f9fbf7b80acbca631dda0a952
2020-05-11 21:52:47,395 autosynth > Running: git log -1 --pretty=%at 7e1c7603c8cb2f05ef29842405654ff8a9d3d108
2020-05-11 21:52:47,400 autosynth > Running: git log -1 --pretty=%at f78da0d54c78aa8f66d52a448db1f7ec0e0fd591
2020-05-11 21:52:47,406 autosynth > Running: git log -1 --pretty=%at aaff764c185e18a6c73227357c3df5fa60fec85a
2020-05-11 21:52:47,412 autosynth > Running: git log -1 --pretty=%at ee9e8e4e67f06aba881392a1a2956fbe7a42d216
2020-05-11 21:52:47,418 autosynth > Running: git log -1 --pretty=%at 93b25607689c4bcdb3cb35c657fd3c16249dcd54
2020-05-11 21:52:47,423 autosynth > Running: git log -1 --pretty=%at ab3468cf63abc521be76c9eceb620d61c5ddd76a
2020-05-11 21:52:47,428 autosynth > Running: git log -1 --pretty=%at a737a68e214d1cf927457c996b73e2600b14a949
2020-05-11 21:52:47,433 autosynth > Running: git log -1 --pretty=%at 4ca990cb1a6771f11469b1fdf2fc578271703db0
2020-05-11 21:52:47,439 autosynth > Running: git log -1 --pretty=%at e0f9d9e1f9de890db765be46f45ca8490723e3eb
2020-05-11 21:52:47,445 autosynth > Running: git log -1 --pretty=%at 1b5a8d2bbb69c5a04db26bd546d2888e609c6bab
2020-05-11 21:52:47,450 autosynth > Running: git log -1 --pretty=%at 4bef0001ac7040431ea24b6187424fdec9c08b1b
2020-05-11 21:52:47,456 autosynth > Running: git log -1 --pretty=%at a3a0bf0f6291d69f2ff3df7fcd63d28ee20ac727
2020-05-11 21:52:47,461 autosynth > Running: git log -1 --pretty=%at 6dfd72d028a0d0a43764e060f7b15e004385c3a1
2020-05-11 21:52:47,466 autosynth > Running: git log -1 --pretty=%at 839fae42335ee1bb1e70767f3e6c51738683892b
2020-05-11 21:52:47,471 autosynth > Running: git log -1 --pretty=%at 30cfca094376e4904e32e71c838a81169fd4a2e2
2020-05-11 21:52:47,476 autosynth > Running: git log -1 --pretty=%at c08dcec05ce1c181bcdbce59cabba36e0e541ff6
2020-05-11 21:52:47,481 autosynth > Running: git log -1 --pretty=%at 2fc2caaacb15949c7f80426bfc7dafdd41dbc333
2020-05-11 21:52:47,486 autosynth > Running: git log -1 --pretty=%at 684dfea7decfeca7a7526ea96a8e9256694dd5d8
2020-05-11 21:52:47,491 autosynth > Running: git log -1 --pretty=%at fb8f62b6784f43faf4b64179c57ce4b4931b1a00
2020-05-11 21:52:47,495 autosynth > Running: git log -1 --pretty=%at d3cc7bbf41afdfd6677cc1ab0d53965625cc073b
2020-05-11 21:52:47,500 autosynth > Running: git log -1 --pretty=%at 83816bb3093686a28af2891db5b7506614a820b1
2020-05-11 21:52:47,505 autosynth > Running: git log -1 --pretty=%at 73d4b5d9a791f8b1ee63d439ffe909bb8ffa07f7
2020-05-11 21:52:47,510 autosynth > Running: git log -1 --pretty=%at 482e5206e05e4ba8cd79738f6fb7a521c8c23555
2020-05-11 21:52:47,516 autosynth > Running: git log -1 --pretty=%at 1019b0d9dbd999e1107f90247fcc478678d1105c
2020-05-11 21:52:47,521 autosynth > Running: git log -1 --pretty=%at 4186d3e5424edc088a0f4dbce78f02dd860f3de8
2020-05-11 21:52:47,526 autosynth > Running: git log -1 --pretty=%at aed11c01e52921613b9ee469c2d85f5f33175fb7
2020-05-11 21:52:47,531 autosynth > Running: git log -1 --pretty=%at edd3b80fb770548d6ad780105f1782de6ff73ea0
2020-05-11 21:52:47,536 autosynth > Running: git log -1 --pretty=%at f8a9933e5e98202b04ef427f28d1d79221190fa4
2020-05-11 21:52:47,541 autosynth > Running: git log -1 --pretty=%at 19465d3ec5e5acdb01521d8f3bddd311bcbee28d
2020-05-11 21:52:47,546 autosynth > Running: git log -1 --pretty=%at f5e4c17dc78a966dbf29961dd01f9bbd63e20a04
2020-05-11 21:52:47,551 autosynth > Running: git log -1 --pretty=%at 21ce80f9d4785dc75aa36d26944bddd5a1b7e25b
2020-05-11 21:52:47,556 autosynth > Running: git log -1 --pretty=%at 716f741f2d307b48cbe8a5bc3bc883571212344a
2020-05-11 21:52:47,561 autosynth > Running: git log -1 --pretty=%at 01b6f23d24b27878b48667ce597876d66b59780e
2020-05-11 21:52:47,566 autosynth > Running: git log -1 --pretty=%at 969a2340e74c73227e7c1638ed7650abcac22ee4
2020-05-11 21:52:47,571 autosynth > Running: git log -1 --pretty=%at 720b7bff0da03ed9e67cb2362f2d291241fa59ae
2020-05-11 21:52:47,576 autosynth > Running: git log -1 --pretty=%at 7c1a92d00de654db2fb440243129c61393a694e2
2020-05-11 21:52:47,581 autosynth > Running: git log -1 --pretty=%at 46fb62a3ebda62cc6c1c98e4f1eaf91965eab580
2020-05-11 21:52:47,586 autosynth > Running: git log -1 --pretty=%at 6b685a261479e27fd959dd7b1ff253d6cf595172
2020-05-11 21:52:47,591 autosynth > Running: git log -1 --pretty=%at 958a803ee0c43523fa3e6e0bb016276a69614c8e
2020-05-11 21:52:47,596 autosynth > Running: git log -1 --pretty=%at cdddf139b36000b3a7c65fd2a7781e253262359a
2020-05-11 21:52:47,601 autosynth > Running: git log -1 --pretty=%at 04cb397eb7590ea1e6c10a39fca3d8fe0fb3d256
2020-05-11 21:52:47,605 autosynth > Running: git log -1 --pretty=%at a2c1c4f89a5e220e7b39420ebea33623c7c72804
2020-05-11 21:52:47,610 autosynth > Running: git log -1 --pretty=%at ea3ae392fced25d7d6f5c7ef9c8a1f24768a0936
2020-05-11 21:52:47,614 autosynth > Running: git log -1 --pretty=%at 09c48461232ce929c34386259eb59018ad2d8eef
2020-05-11 21:52:47,619 autosynth > Running: git log -1 --pretty=%at ab883569eb0257bbf16a6d825fd018b3adde3912
2020-05-11 21:52:47,624 autosynth > Running: git log -1 --pretty=%at 756bc4dfc24e8bc4c5dd4116daa41a0440ebf5a0
2020-05-11 21:52:47,629 autosynth > Running: git log -1 --pretty=%at 558bb0d70fa98ea228483b44885704d8941a6a80
2020-05-11 21:52:47,634 autosynth > Running: git log -1 --pretty=%at be74d3e532faa47eb59f1a0eaebde0860d1d8ab4
2020-05-11 21:52:47,638 autosynth > Running: git log -1 --pretty=%at 4674113712c0c7ada19e6a8219d7963ff174b392
2020-05-11 21:52:47,643 autosynth > Running: git log -1 --pretty=%at 5bbfd095faedfe273819d266f21e402192a29041
2020-05-11 21:52:47,648 autosynth > Running: git log -1 --pretty=%at 4fa923bd3dafb91df8613accbe2230299cc5b98e
2020-05-11 21:52:47,652 autosynth > Running: git log -1 --pretty=%at 55cdc844877d97139f25004229842624a6a86a02
2020-05-11 21:52:47,657 autosynth > Running: git log -1 --pretty=%at 98c50772ec23295c64cf0d2ddf199ea52961fd4c
2020-05-11 21:52:47,662 autosynth > Forking toolbox
2020-05-11 21:52:47,662 autosynth > forking: autosynth-self
2020-05-11 21:52:47,663 autosynth > Running: git branch -f autosynth-self
2020-05-11 21:52:47,668 autosynth > forking: autosynth-googleapis
2020-05-11 21:52:47,669 autosynth > Running: git branch -f autosynth-googleapis
2020-05-11 21:52:47,674 autosynth > forking: autosynth-synthtool
2020-05-11 21:52:47,675 autosynth > Running: git branch -f autosynth-synthtool
2020-05-11 21:52:47,679 autosynth > Forked branch: autosynth-self
2020-05-11 21:52:47,997 autosynth > Running: git checkout autosynth-self
2020-05-11 21:52:48,012 autosynth > Building most recent versions
2020-05-11 21:52:48,013 autosynth > Running: git checkout 50afb0e301f97194bf58c78b0133606cf140784f
2020-05-11 21:52:48,020 autosynth > Running: git checkout d741cd976975c745d0199987aff0e908b8352992
2020-05-11 21:52:48,176 autosynth > Running: git checkout f8a9933e5e98202b04ef427f28d1d79221190fa4
2020-05-11 21:52:48,189 autosynth > Running: git branch -f autosynth-self-2
2020-05-11 21:52:48,196 autosynth > Running: git checkout autosynth-self-2
2020-05-11 21:52:48,204 autosynth > Running synthtool
2020-05-11 21:52:48,204 autosynth > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py --
2020-05-11 21:52:48,204 autosynth > -> /tmpfs/src/github/synthtool/logs/googleapis/java-billing/self/2/sponge_log.log
2020-05-11 21:53:10,684 autosynth > Failed executing /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py --
2020-05-11 21:53:10,685 autosynth > Synthesis failed
2020-05-11 21:53:10,688 autosynth > Running: git reset --hard HEAD
2020-05-11 21:53:10,817 autosynth > Running: git checkout autosynth-self
2020-05-11 21:53:10,829 autosynth > Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
2020-05-11 21:53:10,830 autosynth > Building most recent versions
2020-05-11 21:53:10,831 autosynth > Running: git checkout 50afb0e301f97194bf58c78b0133606cf140784f
2020-05-11 21:53:10,839 autosynth > Running: git checkout edd3b80fb770548d6ad780105f1782de6ff73ea0
2020-05-11 21:53:11,242 autosynth > Running: git checkout 98c50772ec23295c64cf0d2ddf199ea52961fd4c
2020-05-11 21:53:11,363 autosynth > Running: git branch -f autosynth-126
2020-05-11 21:53:11,371 autosynth > Running: git checkout autosynth-126
2020-05-11 21:53:11,378 autosynth > Running synthtool
2020-05-11 21:53:11,378 autosynth > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py --
2020-05-11 21:53:11,378 autosynth > -> /tmpfs/src/github/synthtool/logs/googleapis/java-billing/126/sponge_log.log
2020-05-11 22:03:18,709 autosynth > Failed executing /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py --
2020-05-11 22:03:18,711 autosynth > Synthesis failed
2020-05-11 22:03:18,717 autosynth > Running: git reset --hard HEAD
2020-05-11 22:03:18,911 autosynth > Running: git checkout autosynth
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 642, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 489, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 624, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 379, in synthesize_loop
synthesize_inner_loop(toolbox, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 390, in synthesize_inner_loop
synthesizer, len(toolbox.versions) - 1
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 274, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 110, in synthesize
proc.check_returncode()
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/c5f7338e-3c88-4aa0-9102-e6d02a16d4ef).
|
1.0
|
Synthesis failed for java-billing - Hello! Autosynth couldn't regenerate java-billing. :broken_heart:
Here's the output from running `synth.py`:
```
2020-05-11 21:52:44,383 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/java-billing
2020-05-11 21:52:44,900 autosynth > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore
2020-05-11 21:52:44,905 autosynth > Running: git config user.name yoshi-automation
2020-05-11 21:52:44,910 autosynth > Running: git config user.email yoshi-automation@google.com
2020-05-11 21:52:44,915 autosynth > Running: git config push.default simple
2020-05-11 21:52:44,920 autosynth > Running: git branch -f autosynth
2020-05-11 21:52:44,926 autosynth > Running: git checkout autosynth
2020-05-11 21:52:44,939 autosynth > autosynth flags:
2020-05-11 21:52:44,940 autosynth > AUTOSYNTH_MULTIPLE_COMMITS: true
2020-05-11 21:52:44,940 autosynth > AUTOSYNTH_MULTIPLE_PRS: true
2020-05-11 21:52:44,941 autosynth > Running: git rev-parse --show-toplevel
2020-05-11 21:52:44,946 autosynth > Running: git log -1 --pretty=%H
2020-05-11 21:52:44,951 autosynth > Running: git remote get-url origin
2020-05-11 21:52:44,956 autosynth > Running: git clone --single-branch https://github.com/googleapis/googleapis.git -- /tmpfs/tmp/tmpaaavgl1n/googleapis
2020-05-11 21:52:46,450 autosynth > Running: git log d741cd976975c745d0199987aff0e908b8352992..HEAD --pretty=%H --no-decorate
2020-05-11 21:52:46,457 autosynth > Running: git clone --single-branch https://github.com/googleapis/synthtool.git -- /tmpfs/tmp/tmpaaavgl1n/synthtool
2020-05-11 21:52:47,010 autosynth > Running: git log f8a9933e5e98202b04ef427f28d1d79221190fa4..HEAD --pretty=%H --no-decorate
2020-05-11 21:52:47,017 autosynth > Running: git log -1 --pretty=%at d741cd976975c745d0199987aff0e908b8352992
2020-05-11 21:52:47,022 autosynth > Running: git log -1 --pretty=%at e4f4b23e07315492b533746e6a9255a1e6b3e748
2020-05-11 21:52:47,028 autosynth > Running: git log -1 --pretty=%at ee4ea76504aa60c2bff9b7c11269c155d8c21e0d
2020-05-11 21:52:47,033 autosynth > Running: git log -1 --pretty=%at 6114e2d30c0792522720d1c2cf8f5e111be7545b
2020-05-11 21:52:47,038 autosynth > Running: git log -1 --pretty=%at af71ea52236b05ce73253f6bc935fcf35c866664
2020-05-11 21:52:47,044 autosynth > Running: git log -1 --pretty=%at c451a4457f38e5582e9794a0094c773355a9eaad
2020-05-11 21:52:47,050 autosynth > Running: git log -1 --pretty=%at 9af7212bc2cba8a7184f032e1ab6662ccf604b3a
2020-05-11 21:52:47,055 autosynth > Running: git log -1 --pretty=%at ac1f904979899bc598fb0567cf7a367f08b40b43
2020-05-11 21:52:47,061 autosynth > Running: git log -1 --pretty=%at fe8b17aff5b419ffb418a77228fa049190488233
2020-05-11 21:52:47,067 autosynth > Running: git log -1 --pretty=%at fd35dc6865fa9f47f283d5a6c37c4901162007cf
2020-05-11 21:52:47,072 autosynth > Running: git log -1 --pretty=%at 96c0db42dbcb3bb0116cfa55c054dd139ffb3912
2020-05-11 21:52:47,077 autosynth > Running: git log -1 --pretty=%at b8c7c4e8da56547fc32bee9b563d306fadabdb6f
2020-05-11 21:52:47,082 autosynth > Running: git log -1 --pretty=%at b21bc9b46e3263cd4ebe50f60c61589e3fe797f2
2020-05-11 21:52:47,087 autosynth > Running: git log -1 --pretty=%at 41d5dee23527df3fc9127838a027df393d79ff3c
2020-05-11 21:52:47,092 autosynth > Running: git log -1 --pretty=%at 801c3fb47b1a533989a1fe5af4aedf02af5f2eae
2020-05-11 21:52:47,097 autosynth > Running: git log -1 --pretty=%at 1692c3d51ee7d4366ce81892df1b9ca9846480d3
2020-05-11 21:52:47,102 autosynth > Running: git log -1 --pretty=%at 1df336ab4a989e7b0bc363baa9fc91eb66b0625e
2020-05-11 21:52:47,107 autosynth > Running: git log -1 --pretty=%at 1e92c343790ae63f88817afde60bd2d5bc775f36
2020-05-11 21:52:47,111 autosynth > Running: git log -1 --pretty=%at cf18ab5200e41e9d0cfef88c6c4bcdd394faf1ca
2020-05-11 21:52:47,116 autosynth > Running: git log -1 --pretty=%at ab869e90c6ced0ef30fcc8e8fbe8e527a695d3cb
2020-05-11 21:52:47,121 autosynth > Running: git log -1 --pretty=%at 13736ead90e0fa65103eb03fea775a1c010564af
2020-05-11 21:52:47,126 autosynth > Running: git log -1 --pretty=%at f9d76ac088f5d444b16d5e17fe58a9eea57b7a99
2020-05-11 21:52:47,131 autosynth > Running: git log -1 --pretty=%at 1bd77e8ce6f953ac641af7966d0c52646afc16a8
2020-05-11 21:52:47,136 autosynth > Running: git log -1 --pretty=%at 3824f547aa44df459580615c73cabb30a2a78ee0
2020-05-11 21:52:47,141 autosynth > Running: git log -1 --pretty=%at 4d61e1cb40184a7ad63ef37b1813f6608718674a
2020-05-11 21:52:47,147 autosynth > Running: git log -1 --pretty=%at 9ea892ebe8a7cd66da9bd62eb7086b8dc2ddd559
2020-05-11 21:52:47,152 autosynth > Running: git log -1 --pretty=%at 0feb08065d91d6deebffef3019751f0619a72140
2020-05-11 21:52:47,157 autosynth > Running: git log -1 --pretty=%at 789fd7e0680b04a589447139811d55c5d1f34afe
2020-05-11 21:52:47,163 autosynth > Running: git log -1 --pretty=%at 4efd71279f0781579ced60c374336ce1b8b8cea8
2020-05-11 21:52:47,168 autosynth > Running: git log -1 --pretty=%at a34f4eb96a94bfd5b3e7f8de253db28d0a3e9237
2020-05-11 21:52:47,173 autosynth > Running: git log -1 --pretty=%at f561f57eea013327be27d8f6799b3cc2d18ee804
2020-05-11 21:52:47,177 autosynth > Running: git log -1 --pretty=%at c9942c00e8f35b5b37cd589ef7a0323aeda264e2
2020-05-11 21:52:47,182 autosynth > Running: git log -1 --pretty=%at e2a59ff2f861b1b92b2f15b242780dccddd77ab9
2020-05-11 21:52:47,188 autosynth > Running: git log -1 --pretty=%at df955af5d9508cf9215de6f1fbdad033d5b8fef6
2020-05-11 21:52:47,193 autosynth > Running: git log -1 --pretty=%at dea1f52034d188ca371c3a70c716ca2a6b3f2aa1
2020-05-11 21:52:47,198 autosynth > Running: git log -1 --pretty=%at 35672e7936b923eff048d0ee1529faa163cc81b8
2020-05-11 21:52:47,203 autosynth > Running: git log -1 --pretty=%at 26523a96798ce1a6caa1b3c912119059cfcc98a7
2020-05-11 21:52:47,208 autosynth > Running: git log -1 --pretty=%at 04eff659376f40a7111bf0e8e49fbb38bc7739f7
2020-05-11 21:52:47,213 autosynth > Running: git log -1 --pretty=%at 3028060618e8024af9a32b3ab3456c160091ecb7
2020-05-11 21:52:47,218 autosynth > Running: git log -1 --pretty=%at b5aa7d668d9513cf0d305c8b23262ee0ea9b850c
2020-05-11 21:52:47,223 autosynth > Running: git log -1 --pretty=%at c0a072159eda3c7d5a98201d6500feaf78a18073
2020-05-11 21:52:47,228 autosynth > Running: git log -1 --pretty=%at fea22b1d9f27f86ef355c1d0dba00e0791a08a19
2020-05-11 21:52:47,233 autosynth > Running: git log -1 --pretty=%at 80dac49dd5f407b14f15ac86b4922967b5dd6508
2020-05-11 21:52:47,238 autosynth > Running: git log -1 --pretty=%at b8495354535dc1370377cbb15a3ef7a07ec3a114
2020-05-11 21:52:47,243 autosynth > Running: git log -1 --pretty=%at ab5d07c2883026eec395d73ff59a17f4ad0153db
2020-05-11 21:52:47,249 autosynth > Running: git log -1 --pretty=%at 160e3187dc93265ed635ad68e12c483c8ab4f7be
2020-05-11 21:52:47,254 autosynth > Running: git log -1 --pretty=%at 629e6bea8418ad4a5d081ab9eb48bb327ad368e1
2020-05-11 21:52:47,259 autosynth > Running: git log -1 --pretty=%at c3f3cdeda70f60d808eaaae8fea44d306b6d5cc2
2020-05-11 21:52:47,264 autosynth > Running: git log -1 --pretty=%at cdf13efacdea0649e940452f9c5d320b93735974
2020-05-11 21:52:47,269 autosynth > Running: git log -1 --pretty=%at 101d31acd73076c52d78e18322be01f3debe8cb5
2020-05-11 21:52:47,274 autosynth > Running: git log -1 --pretty=%at cab24c7c9b570e2b483e4f742a7a4a111167238b
2020-05-11 21:52:47,279 autosynth > Running: git log -1 --pretty=%at 6e0ac551e057e20a6c41ab0b21ffa9623e463326
2020-05-11 21:52:47,285 autosynth > Running: git log -1 --pretty=%at 69b4dc56d21394e7c39bc62fe11c76a88aa5511f
2020-05-11 21:52:47,290 autosynth > Running: git log -1 --pretty=%at bcc476396e799806d3355e87246c6becf6250a70
2020-05-11 21:52:47,295 autosynth > Running: git log -1 --pretty=%at b5e9b0b09cc15c263f07da92fa64db31b2e4dd07
2020-05-11 21:52:47,300 autosynth > Running: git log -1 --pretty=%at c09ec5156db66d326cce56ed3a2a8092e02c3abd
2020-05-11 21:52:47,306 autosynth > Running: git log -1 --pretty=%at 42ee97c1b93a0e3759bbba3013da309f670a90ab
2020-05-11 21:52:47,310 autosynth > Running: git log -1 --pretty=%at 45b363d3ab4692115cc5018eb26dcdaa49b320a8
2020-05-11 21:52:47,315 autosynth > Running: git log -1 --pretty=%at 012c553112fcceb7b69822e24335575bf99ec62b
2020-05-11 21:52:47,320 autosynth > Running: git log -1 --pretty=%at 2bf9f22da3f101aeb536e1e88fe036323b8e9188
2020-05-11 21:52:47,326 autosynth > Running: git log -1 --pretty=%at 862cd840dbb0101a0630913217bf8b6d4101c84c
2020-05-11 21:52:47,331 autosynth > Running: git log -1 --pretty=%at b3bcfd059ec684445a38db693f6e36f14a1d2618
2020-05-11 21:52:47,336 autosynth > Running: git log -1 --pretty=%at 1d520eaa7bbd8b40f53942ff03b5496fff887b53
2020-05-11 21:52:47,341 autosynth > Running: git log -1 --pretty=%at 0d7c3565dd942d0cc41b16e895046486ae29e44e
2020-05-11 21:52:47,347 autosynth > Running: git log -1 --pretty=%at d14bf59a446c14ef16e9931ebfc8e63ab549bf07
2020-05-11 21:52:47,352 autosynth > Running: git log -1 --pretty=%at 28e76243c23cc282efbb288cb558c174e3e5e9ee
2020-05-11 21:52:47,357 autosynth > Running: git log -1 --pretty=%at 5dca7f1079a4c127245c0f1dbb8293b434fbd0c8
2020-05-11 21:52:47,363 autosynth > Running: git log -1 --pretty=%at c2bfceedfc9f01356229bd206ad25e250aa9551c
2020-05-11 21:52:47,368 autosynth > Running: git log -1 --pretty=%at 371206424e6db4df2cb3734bb1013dcaac9dbe4b
2020-05-11 21:52:47,373 autosynth > Running: git log -1 --pretty=%at ff17eefa184f6985e5d9a8ee3e725da95a06cd23
2020-05-11 21:52:47,379 autosynth > Running: git log -1 --pretty=%at c284e4f849e4b085a297d336cef8721ab69ba013
2020-05-11 21:52:47,384 autosynth > Running: git log -1 --pretty=%at 756b174de4a122461993c1c583345533d819936d
2020-05-11 21:52:47,390 autosynth > Running: git log -1 --pretty=%at 3f5f8a2258c6a41f9fbf7b80acbca631dda0a952
2020-05-11 21:52:47,395 autosynth > Running: git log -1 --pretty=%at 7e1c7603c8cb2f05ef29842405654ff8a9d3d108
2020-05-11 21:52:47,400 autosynth > Running: git log -1 --pretty=%at f78da0d54c78aa8f66d52a448db1f7ec0e0fd591
2020-05-11 21:52:47,406 autosynth > Running: git log -1 --pretty=%at aaff764c185e18a6c73227357c3df5fa60fec85a
2020-05-11 21:52:47,412 autosynth > Running: git log -1 --pretty=%at ee9e8e4e67f06aba881392a1a2956fbe7a42d216
2020-05-11 21:52:47,418 autosynth > Running: git log -1 --pretty=%at 93b25607689c4bcdb3cb35c657fd3c16249dcd54
2020-05-11 21:52:47,423 autosynth > Running: git log -1 --pretty=%at ab3468cf63abc521be76c9eceb620d61c5ddd76a
2020-05-11 21:52:47,428 autosynth > Running: git log -1 --pretty=%at a737a68e214d1cf927457c996b73e2600b14a949
2020-05-11 21:52:47,433 autosynth > Running: git log -1 --pretty=%at 4ca990cb1a6771f11469b1fdf2fc578271703db0
2020-05-11 21:52:47,439 autosynth > Running: git log -1 --pretty=%at e0f9d9e1f9de890db765be46f45ca8490723e3eb
2020-05-11 21:52:47,445 autosynth > Running: git log -1 --pretty=%at 1b5a8d2bbb69c5a04db26bd546d2888e609c6bab
2020-05-11 21:52:47,450 autosynth > Running: git log -1 --pretty=%at 4bef0001ac7040431ea24b6187424fdec9c08b1b
2020-05-11 21:52:47,456 autosynth > Running: git log -1 --pretty=%at a3a0bf0f6291d69f2ff3df7fcd63d28ee20ac727
2020-05-11 21:52:47,461 autosynth > Running: git log -1 --pretty=%at 6dfd72d028a0d0a43764e060f7b15e004385c3a1
2020-05-11 21:52:47,466 autosynth > Running: git log -1 --pretty=%at 839fae42335ee1bb1e70767f3e6c51738683892b
2020-05-11 21:52:47,471 autosynth > Running: git log -1 --pretty=%at 30cfca094376e4904e32e71c838a81169fd4a2e2
2020-05-11 21:52:47,476 autosynth > Running: git log -1 --pretty=%at c08dcec05ce1c181bcdbce59cabba36e0e541ff6
2020-05-11 21:52:47,481 autosynth > Running: git log -1 --pretty=%at 2fc2caaacb15949c7f80426bfc7dafdd41dbc333
2020-05-11 21:52:47,486 autosynth > Running: git log -1 --pretty=%at 684dfea7decfeca7a7526ea96a8e9256694dd5d8
2020-05-11 21:52:47,491 autosynth > Running: git log -1 --pretty=%at fb8f62b6784f43faf4b64179c57ce4b4931b1a00
2020-05-11 21:52:47,495 autosynth > Running: git log -1 --pretty=%at d3cc7bbf41afdfd6677cc1ab0d53965625cc073b
2020-05-11 21:52:47,500 autosynth > Running: git log -1 --pretty=%at 83816bb3093686a28af2891db5b7506614a820b1
2020-05-11 21:52:47,505 autosynth > Running: git log -1 --pretty=%at 73d4b5d9a791f8b1ee63d439ffe909bb8ffa07f7
2020-05-11 21:52:47,510 autosynth > Running: git log -1 --pretty=%at 482e5206e05e4ba8cd79738f6fb7a521c8c23555
2020-05-11 21:52:47,516 autosynth > Running: git log -1 --pretty=%at 1019b0d9dbd999e1107f90247fcc478678d1105c
2020-05-11 21:52:47,521 autosynth > Running: git log -1 --pretty=%at 4186d3e5424edc088a0f4dbce78f02dd860f3de8
2020-05-11 21:52:47,526 autosynth > Running: git log -1 --pretty=%at aed11c01e52921613b9ee469c2d85f5f33175fb7
2020-05-11 21:52:47,531 autosynth > Running: git log -1 --pretty=%at edd3b80fb770548d6ad780105f1782de6ff73ea0
2020-05-11 21:52:47,536 autosynth > Running: git log -1 --pretty=%at f8a9933e5e98202b04ef427f28d1d79221190fa4
2020-05-11 21:52:47,541 autosynth > Running: git log -1 --pretty=%at 19465d3ec5e5acdb01521d8f3bddd311bcbee28d
2020-05-11 21:52:47,546 autosynth > Running: git log -1 --pretty=%at f5e4c17dc78a966dbf29961dd01f9bbd63e20a04
2020-05-11 21:52:47,551 autosynth > Running: git log -1 --pretty=%at 21ce80f9d4785dc75aa36d26944bddd5a1b7e25b
2020-05-11 21:52:47,556 autosynth > Running: git log -1 --pretty=%at 716f741f2d307b48cbe8a5bc3bc883571212344a
2020-05-11 21:52:47,561 autosynth > Running: git log -1 --pretty=%at 01b6f23d24b27878b48667ce597876d66b59780e
2020-05-11 21:52:47,566 autosynth > Running: git log -1 --pretty=%at 969a2340e74c73227e7c1638ed7650abcac22ee4
2020-05-11 21:52:47,571 autosynth > Running: git log -1 --pretty=%at 720b7bff0da03ed9e67cb2362f2d291241fa59ae
2020-05-11 21:52:47,576 autosynth > Running: git log -1 --pretty=%at 7c1a92d00de654db2fb440243129c61393a694e2
2020-05-11 21:52:47,581 autosynth > Running: git log -1 --pretty=%at 46fb62a3ebda62cc6c1c98e4f1eaf91965eab580
2020-05-11 21:52:47,586 autosynth > Running: git log -1 --pretty=%at 6b685a261479e27fd959dd7b1ff253d6cf595172
2020-05-11 21:52:47,591 autosynth > Running: git log -1 --pretty=%at 958a803ee0c43523fa3e6e0bb016276a69614c8e
2020-05-11 21:52:47,596 autosynth > Running: git log -1 --pretty=%at cdddf139b36000b3a7c65fd2a7781e253262359a
2020-05-11 21:52:47,601 autosynth > Running: git log -1 --pretty=%at 04cb397eb7590ea1e6c10a39fca3d8fe0fb3d256
2020-05-11 21:52:47,605 autosynth > Running: git log -1 --pretty=%at a2c1c4f89a5e220e7b39420ebea33623c7c72804
2020-05-11 21:52:47,610 autosynth > Running: git log -1 --pretty=%at ea3ae392fced25d7d6f5c7ef9c8a1f24768a0936
2020-05-11 21:52:47,614 autosynth > Running: git log -1 --pretty=%at 09c48461232ce929c34386259eb59018ad2d8eef
2020-05-11 21:52:47,619 autosynth > Running: git log -1 --pretty=%at ab883569eb0257bbf16a6d825fd018b3adde3912
2020-05-11 21:52:47,624 autosynth > Running: git log -1 --pretty=%at 756bc4dfc24e8bc4c5dd4116daa41a0440ebf5a0
2020-05-11 21:52:47,629 autosynth > Running: git log -1 --pretty=%at 558bb0d70fa98ea228483b44885704d8941a6a80
2020-05-11 21:52:47,634 autosynth > Running: git log -1 --pretty=%at be74d3e532faa47eb59f1a0eaebde0860d1d8ab4
2020-05-11 21:52:47,638 autosynth > Running: git log -1 --pretty=%at 4674113712c0c7ada19e6a8219d7963ff174b392
2020-05-11 21:52:47,643 autosynth > Running: git log -1 --pretty=%at 5bbfd095faedfe273819d266f21e402192a29041
2020-05-11 21:52:47,648 autosynth > Running: git log -1 --pretty=%at 4fa923bd3dafb91df8613accbe2230299cc5b98e
2020-05-11 21:52:47,652 autosynth > Running: git log -1 --pretty=%at 55cdc844877d97139f25004229842624a6a86a02
2020-05-11 21:52:47,657 autosynth > Running: git log -1 --pretty=%at 98c50772ec23295c64cf0d2ddf199ea52961fd4c
2020-05-11 21:52:47,662 autosynth > Forking toolbox
2020-05-11 21:52:47,662 autosynth > forking: autosynth-self
2020-05-11 21:52:47,663 autosynth > Running: git branch -f autosynth-self
2020-05-11 21:52:47,668 autosynth > forking: autosynth-googleapis
2020-05-11 21:52:47,669 autosynth > Running: git branch -f autosynth-googleapis
2020-05-11 21:52:47,674 autosynth > forking: autosynth-synthtool
2020-05-11 21:52:47,675 autosynth > Running: git branch -f autosynth-synthtool
2020-05-11 21:52:47,679 autosynth > Forked branch: autosynth-self
2020-05-11 21:52:47,997 autosynth > Running: git checkout autosynth-self
2020-05-11 21:52:48,012 autosynth > Building most recent versions
2020-05-11 21:52:48,013 autosynth > Running: git checkout 50afb0e301f97194bf58c78b0133606cf140784f
2020-05-11 21:52:48,020 autosynth > Running: git checkout d741cd976975c745d0199987aff0e908b8352992
2020-05-11 21:52:48,176 autosynth > Running: git checkout f8a9933e5e98202b04ef427f28d1d79221190fa4
2020-05-11 21:52:48,189 autosynth > Running: git branch -f autosynth-self-2
2020-05-11 21:52:48,196 autosynth > Running: git checkout autosynth-self-2
2020-05-11 21:52:48,204 autosynth > Running synthtool
2020-05-11 21:52:48,204 autosynth > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py --
2020-05-11 21:52:48,204 autosynth > -> /tmpfs/src/github/synthtool/logs/googleapis/java-billing/self/2/sponge_log.log
2020-05-11 21:53:10,684 autosynth > Failed executing /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py --
2020-05-11 21:53:10,685 autosynth > Synthesis failed
2020-05-11 21:53:10,688 autosynth > Running: git reset --hard HEAD
2020-05-11 21:53:10,817 autosynth > Running: git checkout autosynth-self
2020-05-11 21:53:10,829 autosynth > Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
2020-05-11 21:53:10,830 autosynth > Building most recent versions
2020-05-11 21:53:10,831 autosynth > Running: git checkout 50afb0e301f97194bf58c78b0133606cf140784f
2020-05-11 21:53:10,839 autosynth > Running: git checkout edd3b80fb770548d6ad780105f1782de6ff73ea0
2020-05-11 21:53:11,242 autosynth > Running: git checkout 98c50772ec23295c64cf0d2ddf199ea52961fd4c
2020-05-11 21:53:11,363 autosynth > Running: git branch -f autosynth-126
2020-05-11 21:53:11,371 autosynth > Running: git checkout autosynth-126
2020-05-11 21:53:11,378 autosynth > Running synthtool
2020-05-11 21:53:11,378 autosynth > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py --
2020-05-11 21:53:11,378 autosynth > -> /tmpfs/src/github/synthtool/logs/googleapis/java-billing/126/sponge_log.log
2020-05-11 22:03:18,709 autosynth > Failed executing /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py --
2020-05-11 22:03:18,711 autosynth > Synthesis failed
2020-05-11 22:03:18,717 autosynth > Running: git reset --hard HEAD
2020-05-11 22:03:18,911 autosynth > Running: git checkout autosynth
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 642, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 489, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 624, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 379, in synthesize_loop
synthesize_inner_loop(toolbox, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 390, in synthesize_inner_loop
synthesizer, len(toolbox.versions) - 1
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 274, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 110, in synthesize
proc.check_returncode()
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/c5f7338e-3c88-4aa0-9102-e6d02a16d4ef).
|
non_test
|
synthesis failed for java billing hello autosynth couldn t regenerate java billing broken heart here s the output from running synth py autosynth logs will be written to tmpfs src github synthtool logs googleapis java billing autosynth running git config global core excludesfile home kbuilder autosynth gitignore autosynth running git config user name yoshi automation autosynth running git config user email yoshi automation google com autosynth running git config push default simple autosynth running git branch f autosynth autosynth running git checkout autosynth autosynth autosynth flags autosynth autosynth multiple commits true autosynth autosynth multiple prs true autosynth running git rev parse show toplevel autosynth running git log pretty h autosynth running git remote get url origin autosynth running git clone single branch tmpfs tmp googleapis autosynth running git log head pretty h no decorate autosynth running git clone single branch tmpfs tmp synthtool autosynth running git log head pretty h no decorate autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth forking toolbox autosynth forking autosynth self autosynth running git branch f autosynth self autosynth forking autosynth googleapis autosynth running git branch f autosynth googleapis autosynth forking autosynth synthtool autosynth running git branch f autosynth synthtool autosynth forked branch autosynth self autosynth running git checkout autosynth self autosynth building most recent versions autosynth running git checkout autosynth running git checkout autosynth running git checkout autosynth running git branch f autosynth self autosynth running git checkout autosynth self autosynth running synthtool autosynth running tmpfs src github synthtool env bin m synthtool metadata synth metadata synth py autosynth tmpfs src github synthtool logs googleapis java billing self sponge log log autosynth failed executing tmpfs src github synthtool env bin m synthtool metadata synth metadata synth py autosynth synthesis failed autosynth running git reset hard head autosynth running git checkout autosynth self autosynth command returned non zero exit status autosynth building most recent versions autosynth running git checkout autosynth running git checkout autosynth running git checkout autosynth running git branch f autosynth autosynth running git checkout autosynth autosynth running synthtool autosynth running tmpfs src github synthtool env bin m synthtool metadata synth metadata synth py autosynth tmpfs src github synthtool logs googleapis java billing sponge log log autosynth failed executing tmpfs src github synthtool env bin m synthtool metadata synth metadata synth py autosynth synthesis failed autosynth running git reset hard head autosynth running git checkout autosynth traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop synthesize inner loop toolbox synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize inner loop synthesizer len toolbox versions file tmpfs src github synthtool autosynth synth py line in synthesize version in new branch synthesizer synthesize synth log path self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize proc check returncode file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
| 0
|
339,294
| 30,389,702,466
|
IssuesEvent
|
2023-07-13 05:46:42
|
dealii/dealii
|
https://api.github.com/repos/dealii/dealii
|
opened
|
Regression tester regressed 79ac9c
|
High Priority :warning: Regression tester
|
[start automated regression testsuite report]
Dear all,
this is the automated regression testsuite reporting a new regression between
- new revision: 79ac9cf0deb8c76604c4fd0dee46c185ff26803c [regression testsuite results](https://cdash.dealii.org/index.php?project=deal.II&filtercount=1&showfilters=1&field1=revision&compare1=63&value1=79ac9c)
- old revision: 9e4231ab33c5ee1feb527a01c8687506ea1c3049 [regression testsuite results](https://cdash.dealii.org/index.php?project=deal.II&filtercount=1&showfilters=1&field1=revision&compare1=63&value1=9e4231)
Summary:
- :rage: build errors for configuration [GNU-9.4.0-master-ubuntu-lts-20.04](https://cdash.dealii.org/build/352) on site tester-tng
I have identified the following pull requests as possible candidates:
- <!-- Use readability-qualified-auto clang-tidy check -->#15705 (authored-by: @masterleinad; merged-by: @bangerth)
- <!-- MatrixFree: Memory consumption of mapping data faces by cells in MappingInfo -->#15729 (authored-by: @bergbauer; merged-by: @bangerth)
Notes:
- The regression reported above is a subset of the following full set of regressions compared to the baseline:
- :rage: build errors for configuration [GNU-9.4.0-master-ubuntu-lts-20.04](https://cdash.dealii.org/build/352) on site tester-tng
- I will close and unpin this issue automatically once a full run is complete and compares cleanly to the baseline.
- If closed I will reopen the issue if the testsuite run identified an additional regression compared to what I have reported so far.
[end automated regression testsuite report]
|
1.0
|
Regression tester regressed 79ac9c - [start automated regression testsuite report]
Dear all,
this is the automated regression testsuite reporting a new regression between
- new revision: 79ac9cf0deb8c76604c4fd0dee46c185ff26803c [regression testsuite results](https://cdash.dealii.org/index.php?project=deal.II&filtercount=1&showfilters=1&field1=revision&compare1=63&value1=79ac9c)
- old revision: 9e4231ab33c5ee1feb527a01c8687506ea1c3049 [regression testsuite results](https://cdash.dealii.org/index.php?project=deal.II&filtercount=1&showfilters=1&field1=revision&compare1=63&value1=9e4231)
Summary:
- :rage: build errors for configuration [GNU-9.4.0-master-ubuntu-lts-20.04](https://cdash.dealii.org/build/352) on site tester-tng
I have identified the following pull requests as possible candidates:
- <!-- Use readability-qualified-auto clang-tidy check -->#15705 (authored-by: @masterleinad; merged-by: @bangerth)
- <!-- MatrixFree: Memory consumption of mapping data faces by cells in MappingInfo -->#15729 (authored-by: @bergbauer; merged-by: @bangerth)
Notes:
- The regression reported above is a subset of the following full set of regressions compared to the baseline:
- :rage: build errors for configuration [GNU-9.4.0-master-ubuntu-lts-20.04](https://cdash.dealii.org/build/352) on site tester-tng
- I will close and unpin this issue automatically once a full run is complete and compares cleanly to the baseline.
- If closed I will reopen the issue if the testsuite run identified an additional regression compared to what I have reported so far.
[end automated regression testsuite report]
|
test
|
regression tester regressed dear all this is the automated regression testsuite reporting a new regression between new revision old revision summary rage build errors for configuration on site tester tng i have identified the following pull requests as possible candidates authored by masterleinad merged by bangerth authored by bergbauer merged by bangerth notes the regression reported above is a subset of the following full set of regressions compared to the baseline rage build errors for configuration on site tester tng i will close and unpin this issue automatically once a full run is complete and compares cleanly to the baseline if closed i will reopen the issue if the testsuite run identified an additional regression compared to what i have reported so far
| 1
|
16,832
| 4,099,731,546
|
IssuesEvent
|
2016-06-03 13:49:04
|
F5Networks/f5-openstack-lbaasv2-driver
|
https://api.github.com/repos/F5Networks/f5-openstack-lbaasv2-driver
|
opened
|
Documentation link in the README is not correct.
|
critical documentation
|
#### Agent Version
N/A
#### Operating System
N/A
#### OpenStack Release
Liberty
#### Description
The link to the project documentation is not valid. When you click on it you get the RTD maze for page not found.
Here is the link that is in the README:
http://f5-openstack-lbaasv2.readthedocs.io/en/
#### Deployment
N/A
|
1.0
|
Documentation link in the README is not correct. - #### Agent Version
N/A
#### Operating System
N/A
#### OpenStack Release
Liberty
#### Description
The link to the project documentation is not valid. When you click on it you get the RTD maze for page not found.
Here is the link that is in the README:
http://f5-openstack-lbaasv2.readthedocs.io/en/
#### Deployment
N/A
|
non_test
|
documentation link in the readme is not correct agent version n a operating system n a openstack release liberty description the link to the project documentation is not valid when you click on it you get the rtd maze for page not found here is the link that is in the readme deployment n a
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.