Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
3,104
11,852,478,317
IssuesEvent
2020-03-24 20:01:04
tgstation/tgstation-server
https://api.github.com/repos/tgstation/tgstation-server
opened
Refactor the Chat Manager
Area: Chat Component Issue Maintainability Issue
Just the component. Everything around it is sane, it's just full of shit code that's a mess to understand.
True
Refactor the Chat Manager - Just the component. Everything around it is sane, it's just full of shit code that's a mess to understand.
main
refactor the chat manager just the component everything around it is sane it s just full of shit code that s a mess to understand
1
3,073
11,641,039,394
IssuesEvent
2020-02-29 01:08:48
plotly/dash-table
https://api.github.com/repos/plotly/dash-table
closed
Clean up server-side tests
dash-attribute-maintainability dash-stage-estimate_needed size: 3
Server-side tests are built using n-apps for different flavours of the table (e.g. https://github.com/plotly/dash-table/blob/dev/package.json#L25) -- instead, use a single app with routing to simplify writing & running server-side tests for the table. Alternatively, investigate rewriting the tests with Selenium. We've been having significant issues with Cypress for a while now and those issues don't look like they are going away. Work done in the last year in DCC also greatly improved our Selenium testing environment to the point that maintaining two different sets of testing environments might be moot.
True
Clean up server-side tests - Server-side tests are built using n-apps for different flavours of the table (e.g. https://github.com/plotly/dash-table/blob/dev/package.json#L25) -- instead, use a single app with routing to simplify writing & running server-side tests for the table. Alternatively, investigate rewriting the tests with Selenium. We've been having significant issues with Cypress for a while now and those issues don't look like they are going away. Work done in the last year in DCC also greatly improved our Selenium testing environment to the point that maintaining two different sets of testing environments might be moot.
main
clean up server side tests server side tests are built using n apps for different flavours of the table e g instead use a single app with routing to simplify writing running server side tests for the table alternatively investigate rewriting the tests with selenium we ve been having significant issues with cypress for a while now and those issues don t look like they are going away work done in the last year in dcc also greatly improved our selenium testing environment to the point that maintaining two different sets of testing environments might be moot
1
15,443
27,200,383,640
IssuesEvent
2023-02-20 09:18:15
EEHPCWG/PowerMeasurementMethodology
https://api.github.com/repos/EEHPCWG/PowerMeasurementMethodology
opened
Accuracy requirements
enhancement affects requirements
# Problem statement it is perceived to be hard to figure our the right level for a measurement with respect to accuracy. # Solution - Add the IEC 62053-21 accuracy classes - Reconsider level accuracy thresholds based on information on accuracy of submitted Green500 runs
1.0
Accuracy requirements - # Problem statement it is perceived to be hard to figure our the right level for a measurement with respect to accuracy. # Solution - Add the IEC 62053-21 accuracy classes - Reconsider level accuracy thresholds based on information on accuracy of submitted Green500 runs
non_main
accuracy requirements problem statement it is perceived to be hard to figure our the right level for a measurement with respect to accuracy solution add the iec accuracy classes reconsider level accuracy thresholds based on information on accuracy of submitted runs
0
393,876
11,625,915,409
IssuesEvent
2020-02-27 13:35:02
googleapis/elixir-google-api
https://api.github.com/repos/googleapis/elixir-google-api
opened
Synthesis failed for OAuth2
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate OAuth2. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth-oauth2' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/o_auth2/synth.metadata', 'synth.py', '--'] synthtool > Executing /tmpfs/src/git/autosynth/working_repo/synth.py. synthtool > Cloning https://github.com/googleapis/elixir-google-api.git. synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh OAuth2 synthtool > Wrote metadata to clients/o_auth2/synth.metadata. Changed files: M clients/o_auth2/README.md D clients/o_auth2/lib/google_api/o_auth2/v2/model/jwk.ex D clients/o_auth2/lib/google_api/o_auth2/v2/model/jwk_keys.ex M clients/o_auth2/mix.exs M clients/o_auth2/synth.metadata [autosynth-oauth2 68dbbdf5e] Regenerate OAuth2 client 5 files changed, 5 insertions(+), 111 deletions(-) delete mode 100644 clients/o_auth2/lib/google_api/o_auth2/v2/model/jwk.ex delete mode 100644 clients/o_auth2/lib/google_api/o_auth2/v2/model/jwk_keys.ex To https://github.com/googleapis/elixir-google-api.git + 04b3645ca...68dbbdf5e autosynth-oauth2 -> autosynth-oauth2 (forced update) Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 321, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 309, in main args.repository, branch=branch, title=pr_title, body=pr_body File "/tmpfs/src/git/autosynth/autosynth/github.py", line 65, in create_pull_request response.raise_for_status() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/requests/models.py", line 941, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 502 Server Error: Bad Gateway for url: https://api.github.com/repos/googleapis/elixir-google-api/pulls ``` Google internal developers can see the full log [here](https://sponge/b6c13870-fef3-4917-bf8f-a5d752845525).
1.0
Synthesis failed for OAuth2 - Hello! Autosynth couldn't regenerate OAuth2. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth-oauth2' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/o_auth2/synth.metadata', 'synth.py', '--'] synthtool > Executing /tmpfs/src/git/autosynth/working_repo/synth.py. synthtool > Cloning https://github.com/googleapis/elixir-google-api.git. synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh OAuth2 synthtool > Wrote metadata to clients/o_auth2/synth.metadata. Changed files: M clients/o_auth2/README.md D clients/o_auth2/lib/google_api/o_auth2/v2/model/jwk.ex D clients/o_auth2/lib/google_api/o_auth2/v2/model/jwk_keys.ex M clients/o_auth2/mix.exs M clients/o_auth2/synth.metadata [autosynth-oauth2 68dbbdf5e] Regenerate OAuth2 client 5 files changed, 5 insertions(+), 111 deletions(-) delete mode 100644 clients/o_auth2/lib/google_api/o_auth2/v2/model/jwk.ex delete mode 100644 clients/o_auth2/lib/google_api/o_auth2/v2/model/jwk_keys.ex To https://github.com/googleapis/elixir-google-api.git + 04b3645ca...68dbbdf5e autosynth-oauth2 -> autosynth-oauth2 (forced update) Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 321, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 309, in main args.repository, branch=branch, title=pr_title, body=pr_body File "/tmpfs/src/git/autosynth/autosynth/github.py", line 65, in create_pull_request response.raise_for_status() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/requests/models.py", line 941, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 502 Server Error: Bad Gateway for url: https://api.github.com/repos/googleapis/elixir-google-api/pulls ``` Google internal developers can see the full log [here](https://sponge/b6c13870-fef3-4917-bf8f-a5d752845525).
non_main
synthesis failed for hello autosynth couldn t regenerate broken heart here s the output from running synth py cloning into working repo switched to branch autosynth running synthtool synthtool executing tmpfs src git autosynth working repo synth py synthtool cloning synthtool running docker run rm v home kbuilder cache synthtool elixir google api workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh synthtool wrote metadata to clients o synth metadata changed files m clients o readme md d clients o lib google api o model jwk ex d clients o lib google api o model jwk keys ex m clients o mix exs m clients o synth metadata regenerate client files changed insertions deletions delete mode clients o lib google api o model jwk ex delete mode clients o lib google api o model jwk keys ex to autosynth autosynth forced update traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main args repository branch branch title pr title body pr body file tmpfs src git autosynth autosynth github py line in create pull request response raise for status file tmpfs src git autosynth env lib site packages requests models py line in raise for status raise httperror http error msg response self requests exceptions httperror server error bad gateway for url google internal developers can see the full log
0
3,115
11,904,959,842
IssuesEvent
2020-03-30 17:44:10
diofant/diofant
https://api.github.com/repos/diofant/diofant
opened
Port rootisolation module to use sparse polys
maintainability polys
This include implementing PolyElement.sturm() method to address [this TODO](https://github.com/diofant/diofant/blob/3d08f9ab8cd77359f97411382ad754b5dc09b96e/diofant/polys/rings.py#L2058-L2059).
True
Port rootisolation module to use sparse polys - This include implementing PolyElement.sturm() method to address [this TODO](https://github.com/diofant/diofant/blob/3d08f9ab8cd77359f97411382ad754b5dc09b96e/diofant/polys/rings.py#L2058-L2059).
main
port rootisolation module to use sparse polys this include implementing polyelement sturm method to address
1
325,753
27,961,266,142
IssuesEvent
2023-03-24 15:49:43
wazuh/wazuh-qa
https://api.github.com/repos/wazuh/wazuh-qa
closed
Red Hat Enterprise Linux 9 SCA policy rework - checks 6 to 6.2.16
team/qa feature/sca dev-testing subteam/qa-main level/task type/test
| Target version | Related issue | Related PR | |--------------------|--------------------|-----------------| | 4.4.x | #3391 | https://github.com/wazuh/wazuh/pull/16016 | |Check Id and Name| Status| Ready for QA| |---|---|---| |||| |6 System Maintenance||| |6.1 System File Permissions||| |6.1.1 Ensure permissions on /etc/passwd are configured (Automated)|🟒|🟒| |6.1.2 Ensure permissions on /etc/passwd- are configured (Automated)|🟒|🟒| |6.1.3 Ensure permissions on /etc/group are configured (Automated)|🟒|🟒| |6.1.4 Ensure permissions on /etc/group- are configured (Automated)|🟒|🟒| |6.1.5 Ensure permissions on /etc/shadow are configured (Automated)|🟒|🟒| |6.1.6 Ensure permissions on /etc/shadow- are configured (Automated)|🟒|🟒| |6.1.7 Ensure permissions on /etc/gshadow are configured (Automated)|🟒|🟒| |6.1.8 Ensure permissions on /etc/gshadow- are configured (Automated)|🟒|🟒| |6.1.9 Ensure no world writable files exist (Automated)|⚫|| |6.1.10 Ensure no unowned files or directories exist (Automated)|⚫|| |6.1.11 Ensure no ungrouped files or directories exist (Automated)|⚫|| |6.1.12 Ensure sticky bit is set on all world-writable directories (Automated)|⚫|| |6.1.13 Audit SUID executables (Manual)|⚫|| |6.1.14 Audit SGID executables (Manual)|⚫|| |6.1.15 Audit system file permissions (Manual)|⚫|| |||| |6.2 Local User and Group Settings||| |6.2.1 Ensure accounts in /etc/passwd use shadowed passwords (Automated)|⚫|| |6.2.2 Ensure /etc/shadow password fields are not empty (Automated)|🟒|🟒| |6.2.3 Ensure all groups in /etc/passwd exist in /etc/group (Automated)|⚫|| |6.2.4 Ensure no duplicate UIDs exist (Automated)|⚫|| |6.2.5 Ensure no duplicate GIDs exist (Automated)|⚫|| |6.2.6 Ensure no duplicate user names exist (Automated)|⚫|| |6.2.7 Ensure no duplicate group names exist (Automated)|⚫|| |6.2.8 Ensure root PATH Integrity (Automated)|⚫|| |6.2.9 Ensure root is the only UID 0 account (Automated)|🟒|🟒| |6.2.10 Ensure local interactive user home directories exist (Automated)|⚫|| |6.2.11 Ensure local interactive users own their home directories (Automated)|⚫|| |6.2.12 Ensure local interactive user home directories are mode 750 or more restrictive (Automated)|⚫|| |6.2.13 Ensure no local interactive user has .netrc files (Automated)|⚫|| |6.2.14 Ensure no local interactive user has .forward files (Automated)|⚫|| |6.2.15 Ensure no local interactive user has .rhosts files (Automated)|⚫|| |6.2.16 Ensure local interactive user dot files are not group or world writable (Automated)|⚫|| ||||
2.0
Red Hat Enterprise Linux 9 SCA policy rework - checks 6 to 6.2.16 - | Target version | Related issue | Related PR | |--------------------|--------------------|-----------------| | 4.4.x | #3391 | https://github.com/wazuh/wazuh/pull/16016 | |Check Id and Name| Status| Ready for QA| |---|---|---| |||| |6 System Maintenance||| |6.1 System File Permissions||| |6.1.1 Ensure permissions on /etc/passwd are configured (Automated)|🟒|🟒| |6.1.2 Ensure permissions on /etc/passwd- are configured (Automated)|🟒|🟒| |6.1.3 Ensure permissions on /etc/group are configured (Automated)|🟒|🟒| |6.1.4 Ensure permissions on /etc/group- are configured (Automated)|🟒|🟒| |6.1.5 Ensure permissions on /etc/shadow are configured (Automated)|🟒|🟒| |6.1.6 Ensure permissions on /etc/shadow- are configured (Automated)|🟒|🟒| |6.1.7 Ensure permissions on /etc/gshadow are configured (Automated)|🟒|🟒| |6.1.8 Ensure permissions on /etc/gshadow- are configured (Automated)|🟒|🟒| |6.1.9 Ensure no world writable files exist (Automated)|⚫|| |6.1.10 Ensure no unowned files or directories exist (Automated)|⚫|| |6.1.11 Ensure no ungrouped files or directories exist (Automated)|⚫|| |6.1.12 Ensure sticky bit is set on all world-writable directories (Automated)|⚫|| |6.1.13 Audit SUID executables (Manual)|⚫|| |6.1.14 Audit SGID executables (Manual)|⚫|| |6.1.15 Audit system file permissions (Manual)|⚫|| |||| |6.2 Local User and Group Settings||| |6.2.1 Ensure accounts in /etc/passwd use shadowed passwords (Automated)|⚫|| |6.2.2 Ensure /etc/shadow password fields are not empty (Automated)|🟒|🟒| |6.2.3 Ensure all groups in /etc/passwd exist in /etc/group (Automated)|⚫|| |6.2.4 Ensure no duplicate UIDs exist (Automated)|⚫|| |6.2.5 Ensure no duplicate GIDs exist (Automated)|⚫|| |6.2.6 Ensure no duplicate user names exist (Automated)|⚫|| |6.2.7 Ensure no duplicate group names exist (Automated)|⚫|| |6.2.8 Ensure root PATH Integrity (Automated)|⚫|| |6.2.9 Ensure root is the only UID 0 account (Automated)|🟒|🟒| |6.2.10 Ensure local interactive user home directories exist (Automated)|⚫|| |6.2.11 Ensure local interactive users own their home directories (Automated)|⚫|| |6.2.12 Ensure local interactive user home directories are mode 750 or more restrictive (Automated)|⚫|| |6.2.13 Ensure no local interactive user has .netrc files (Automated)|⚫|| |6.2.14 Ensure no local interactive user has .forward files (Automated)|⚫|| |6.2.15 Ensure no local interactive user has .rhosts files (Automated)|⚫|| |6.2.16 Ensure local interactive user dot files are not group or world writable (Automated)|⚫|| ||||
non_main
red hat enterprise linux sca policy rework checks to target version related issue related pr x check id and name status ready for qa system maintenance system file permissions ensure permissions on etc passwd are configured automated 🟒 🟒 ensure permissions on etc passwd are configured automated 🟒 🟒 ensure permissions on etc group are configured automated 🟒 🟒 ensure permissions on etc group are configured automated 🟒 🟒 ensure permissions on etc shadow are configured automated 🟒 🟒 ensure permissions on etc shadow are configured automated 🟒 🟒 ensure permissions on etc gshadow are configured automated 🟒 🟒 ensure permissions on etc gshadow are configured automated 🟒 🟒 ensure no world writable files exist automated ⚫ ensure no unowned files or directories exist automated ⚫ ensure no ungrouped files or directories exist automated ⚫ ensure sticky bit is set on all world writable directories automated ⚫ audit suid executables manual ⚫ audit sgid executables manual ⚫ audit system file permissions manual ⚫ local user and group settings ensure accounts in etc passwd use shadowed passwords automated ⚫ ensure etc shadow password fields are not empty automated 🟒 🟒 ensure all groups in etc passwd exist in etc group automated ⚫ ensure no duplicate uids exist automated ⚫ ensure no duplicate gids exist automated ⚫ ensure no duplicate user names exist automated ⚫ ensure no duplicate group names exist automated ⚫ ensure root path integrity automated ⚫ ensure root is the only uid account automated 🟒 🟒 ensure local interactive user home directories exist automated ⚫ ensure local interactive users own their home directories automated ⚫ ensure local interactive user home directories are mode or more restrictive automated ⚫ ensure no local interactive user has netrc files automated ⚫ ensure no local interactive user has forward files automated ⚫ ensure no local interactive user has rhosts files automated ⚫ ensure local interactive user dot files are not group or world writable automated ⚫
0
17,066
23,542,658,287
IssuesEvent
2022-08-20 16:43:13
sekiguchi-nagisa/ydsh
https://api.github.com/repos/sekiguchi-nagisa/ydsh
closed
control space insertion behavior in linenoise
incompatible change Interactive API
currently, if size of completion candidates is 1, always insert space after inserting a candidate. in some situation, space insertion is not needed (ex. variable, field, method name completion). need to change `DSState_complete` api
True
control space insertion behavior in linenoise - currently, if size of completion candidates is 1, always insert space after inserting a candidate. in some situation, space insertion is not needed (ex. variable, field, method name completion). need to change `DSState_complete` api
non_main
control space insertion behavior in linenoise currently if size of completion candidates is always insert space after inserting a candidate in some situation space insertion is not needed ex variable field method name completion need to change dsstate complete api
0
344,597
30,751,816,649
IssuesEvent
2023-07-28 20:03:09
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
[Backport v2.7.6] Node gets kicked out of Cluster after snapshots are restored.
kind/bug internal [zube]: To Test QA/S area/provisioning-v2 team/area2 regression JIRA
This is a backport issue for https://github.com/rancher/rancher/issues/42201, automatically created via rancherbot by @Sahota1225 Original issue description: <!--------- For bugs and general issues ---------> **Setup** Rancher version: 2.7.5 Downstream cluster: Custom cluster Nodes: 3 CNI: Cilium Kubernetes version: v1.25.7+rke2r1 **Describe the bug** When restoring a snapshot on a custom cluster a node gets deleted from the cluster. **To Reproduce** 1. Deploy a fresh RKE2 custom cluster 2. Take a snapshot 3. Restore snapshot 4. repeat steps 2-3 until the bug is hit. (usually 2 tries) **Result** A worker node gets deleted from the cluster. **Expected Result** All nodes remain in the cluster and the restore occurs properly. **Screenshots** <img width="1100" alt="image" src="https://github.com/rancher/dashboard/assets/136753565/3f0a0484-0733-4582-b47f-ede6da8053a7"> **Additional context** Tried to reproduce this on 2.7.4 but was unable to do so after 5-10 restores SURE-6669
1.0
[Backport v2.7.6] Node gets kicked out of Cluster after snapshots are restored. - This is a backport issue for https://github.com/rancher/rancher/issues/42201, automatically created via rancherbot by @Sahota1225 Original issue description: <!--------- For bugs and general issues ---------> **Setup** Rancher version: 2.7.5 Downstream cluster: Custom cluster Nodes: 3 CNI: Cilium Kubernetes version: v1.25.7+rke2r1 **Describe the bug** When restoring a snapshot on a custom cluster a node gets deleted from the cluster. **To Reproduce** 1. Deploy a fresh RKE2 custom cluster 2. Take a snapshot 3. Restore snapshot 4. repeat steps 2-3 until the bug is hit. (usually 2 tries) **Result** A worker node gets deleted from the cluster. **Expected Result** All nodes remain in the cluster and the restore occurs properly. **Screenshots** <img width="1100" alt="image" src="https://github.com/rancher/dashboard/assets/136753565/3f0a0484-0733-4582-b47f-ede6da8053a7"> **Additional context** Tried to reproduce this on 2.7.4 but was unable to do so after 5-10 restores SURE-6669
non_main
node gets kicked out of cluster after snapshots are restored this is a backport issue for automatically created via rancherbot by original issue description setup rancher version downstream cluster custom cluster nodes cni cilium kubernetes version describe the bug when restoring a snapshot on a custom cluster a node gets deleted from the cluster to reproduce deploy a fresh custom cluster take a snapshot restore snapshot repeat steps until the bug is hit usually tries result a worker node gets deleted from the cluster expected result all nodes remain in the cluster and the restore occurs properly screenshots img width alt image src additional context tried to reproduce this on but was unable to do so after restores sure
0
234,953
7,733,088,437
IssuesEvent
2018-05-26 06:58:06
dbpiper/PathOfTrading
https://api.github.com/repos/dbpiper/PathOfTrading
opened
Issues with decreasing the height of page (2)
Medium Priority bug
When the height is severely decreased the following issues are observed: - The body can go below the fixed header - The menu icon can be hidden by the tabs These issues should be investigated and fixed.
1.0
Issues with decreasing the height of page (2) - When the height is severely decreased the following issues are observed: - The body can go below the fixed header - The menu icon can be hidden by the tabs These issues should be investigated and fixed.
non_main
issues with decreasing the height of page when the height is severely decreased the following issues are observed the body can go below the fixed header the menu icon can be hidden by the tabs these issues should be investigated and fixed
0
2,338
8,365,342,640
IssuesEvent
2018-10-04 04:30:30
chocolatey/chocolatey-package-requests
https://api.github.com/repos/chocolatey/chocolatey-package-requests
closed
RFP - openjdk11
Status: Available For Maintainer(s)
https://jdk.java.net/11/ There is [concern about Oracle JDK](https://www.reddit.com/r/programming/comments/9j25je/do_not_fall_into_oracles_java_11_trap/), so I think OpenJDK would be useful in addition to future `javaruntime` 11.
True
RFP - openjdk11 - https://jdk.java.net/11/ There is [concern about Oracle JDK](https://www.reddit.com/r/programming/comments/9j25je/do_not_fall_into_oracles_java_11_trap/), so I think OpenJDK would be useful in addition to future `javaruntime` 11.
main
rfp there is so i think openjdk would be useful in addition to future javaruntime
1
197,605
22,596,929,113
IssuesEvent
2022-06-29 04:46:06
dmyers87/headerstrip
https://api.github.com/repos/dmyers87/headerstrip
opened
CVE-2021-23440 (High) detected in set-value-2.0.0.tgz, set-value-0.4.3.tgz
security vulnerability
## CVE-2021-23440 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-2.0.0.tgz</b>, <b>set-value-0.4.3.tgz</b></p></summary> <p> <details><summary><b>set-value-2.0.0.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/set-value</p> <p> Dependency Hierarchy: - lint-staged-7.2.0.tgz (Root Library) - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>set-value-0.4.3.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/set-value</p> <p> Dependency Hierarchy: - lint-staged-7.2.0.tgz (Root Library) - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - union-value-1.0.0.tgz - :x: **set-value-0.4.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/dmyers87/headerstrip/commit/8442716a2567dd4d8cfb426df6ecf4129181aacc">8442716a2567dd4d8cfb426df6ecf4129181aacc</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package set-value before <2.0.1, >=3.0.0 <4.0.1. A type confusion vulnerability can lead to a bypass of CVE-2019-10747 when the user-provided keys used in the path parameter are arrays. <p>Publish Date: 2021-09-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23440>CVE-2021-23440</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23440">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23440</a></p> <p>Release Date: 2021-09-12</p> <p>Fix Resolution (set-value): 2.0.1</p> <p>Direct dependency fix Resolution (lint-staged): 7.2.1</p><p>Fix Resolution (set-value): 2.0.1</p> <p>Direct dependency fix Resolution (lint-staged): 7.2.1</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
True
CVE-2021-23440 (High) detected in set-value-2.0.0.tgz, set-value-0.4.3.tgz - ## CVE-2021-23440 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-2.0.0.tgz</b>, <b>set-value-0.4.3.tgz</b></p></summary> <p> <details><summary><b>set-value-2.0.0.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/set-value</p> <p> Dependency Hierarchy: - lint-staged-7.2.0.tgz (Root Library) - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>set-value-0.4.3.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/set-value</p> <p> Dependency Hierarchy: - lint-staged-7.2.0.tgz (Root Library) - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - union-value-1.0.0.tgz - :x: **set-value-0.4.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/dmyers87/headerstrip/commit/8442716a2567dd4d8cfb426df6ecf4129181aacc">8442716a2567dd4d8cfb426df6ecf4129181aacc</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package set-value before <2.0.1, >=3.0.0 <4.0.1. A type confusion vulnerability can lead to a bypass of CVE-2019-10747 when the user-provided keys used in the path parameter are arrays. <p>Publish Date: 2021-09-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23440>CVE-2021-23440</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23440">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23440</a></p> <p>Release Date: 2021-09-12</p> <p>Fix Resolution (set-value): 2.0.1</p> <p>Direct dependency fix Resolution (lint-staged): 7.2.1</p><p>Fix Resolution (set-value): 2.0.1</p> <p>Direct dependency fix Resolution (lint-staged): 7.2.1</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
non_main
cve high detected in set value tgz set value tgz cve high severity vulnerability vulnerable libraries set value tgz set value tgz set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file package json path to vulnerable library node modules set value dependency hierarchy lint staged tgz root library micromatch tgz snapdragon tgz base tgz cache base tgz x set value tgz vulnerable library set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file package json path to vulnerable library node modules set value dependency hierarchy lint staged tgz root library micromatch tgz snapdragon tgz base tgz cache base tgz union value tgz x set value tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects the package set value before a type confusion vulnerability can lead to a bypass of cve when the user provided keys used in the path parameter are arrays publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution set value direct dependency fix resolution lint staged fix resolution set value direct dependency fix resolution lint staged check this box to open an automated fix pr
0
2,646
9,040,018,074
IssuesEvent
2019-02-10 12:48:58
beefproject/beef
https://api.github.com/repos/beefproject/beef
closed
GeoLite Legacy databases are now discontinued
Core Maintainability
[GeoLite Legacy databases are discontinued](https://support.maxmind.com/geolite-legacy-discontinuation-notice/) as of January 2, 2019. The GeoIP gem and associated GeoIP functionality will need to be replaced with https://github.com/maxmind/MaxMind-DB-Reader-ruby ``` # wget 'https://geolite.maxmind.com/download/geoip/database/GeoLite2-City.tar.gz' --2019-02-10 02:32:42-- https://geolite.maxmind.com/download/geoip/database/GeoLite2-City.tar.gz Resolving geolite.maxmind.com (geolite.maxmind.com)... 104.16.38.47, 104.16.37.47, 2606:4700::6810:252f, ... Connecting to geolite.maxmind.com (geolite.maxmind.com)|104.16.38.47|:443... connected. HTTP request sent, awaiting response... 200 OK Length: 27849181 (27M) [application/gzip] Saving to: β€˜GeoLite2-City.tar.gz’ GeoLite2-City.tar.gz 100%[===========================================>] 26.56M 10.3MB/s in 2.6s 2019-02-10 02:32:45 (10.3 MB/s) - β€˜GeoLite2-City.tar.gz’ saved [27849181/27849181] # gunzip GeoLite2-City.tar.gz # tar xvf GeoLite2-City.tar GeoLite2-City_20190205/ GeoLite2-City_20190205/COPYRIGHT.txt GeoLite2-City_20190205/README.txt GeoLite2-City_20190205/LICENSE.txt GeoLite2-City_20190205/GeoLite2-City.mmdb root@kali-2018:~/Desktop/beef# ./asdf.rb {"continent"=>{"code"=>"NA", "geoname_id"=>6255149, "names"=>{"de"=>"Nordamerika", "en"=>"North America", "es"=>"NorteamΓ©rica", "fr"=>"AmΓ©rique du Nord", "ja"=>"εŒ—γ‚’γƒ‘γƒͺγ‚«", "pt-BR"=>"AmΓ©rica do Norte", "ru"=>"БСвСрная АмСрика", "zh-CN"=>"εŒ—ηΎŽζ΄²"}}, "country"=>{"geoname_id"=>6252001, "iso_code"=>"US", "names"=>{"de"=>"USA", "en"=>"United States", "es"=>"Estados Unidos", "fr"=>"Γ‰tats-Unis", "ja"=>"をパγƒͺγ‚«εˆθ‘†ε›½", "pt-BR"=>"Estados Unidos", "ru"=>"БША", "zh-CN"=>"ηΎŽε›½"}}, "location"=>{"accuracy_radius"=>1000, "latitude"=>37.751, "longitude"=>-97.822}, "registered_country"=>{"geoname_id"=>6252001, "iso_code"=>"US", "names"=>{"de"=>"USA", "en"=>"United States", "es"=>"Estados Unidos", "fr"=>"Γ‰tats-Unis", "ja"=>"をパγƒͺγ‚«εˆθ‘†ε›½", "pt-BR"=>"Estados Unidos", "ru"=>"БША", "zh-CN"=>"ηΎŽε›½"}}} # cat asdf.rb #!/usr/bin/env ruby # require 'maxmind/db' reader = MaxMind::DB.new('GeoLite2-City_20190205/GeoLite2-City.mmdb', mode: MaxMind::DB::MODE_MEMORY) record = reader.get('8.8.8.8') if record.nil? puts '8.8.8.8 not found' else puts record.inspect end ```
True
GeoLite Legacy databases are now discontinued - [GeoLite Legacy databases are discontinued](https://support.maxmind.com/geolite-legacy-discontinuation-notice/) as of January 2, 2019. The GeoIP gem and associated GeoIP functionality will need to be replaced with https://github.com/maxmind/MaxMind-DB-Reader-ruby ``` # wget 'https://geolite.maxmind.com/download/geoip/database/GeoLite2-City.tar.gz' --2019-02-10 02:32:42-- https://geolite.maxmind.com/download/geoip/database/GeoLite2-City.tar.gz Resolving geolite.maxmind.com (geolite.maxmind.com)... 104.16.38.47, 104.16.37.47, 2606:4700::6810:252f, ... Connecting to geolite.maxmind.com (geolite.maxmind.com)|104.16.38.47|:443... connected. HTTP request sent, awaiting response... 200 OK Length: 27849181 (27M) [application/gzip] Saving to: β€˜GeoLite2-City.tar.gz’ GeoLite2-City.tar.gz 100%[===========================================>] 26.56M 10.3MB/s in 2.6s 2019-02-10 02:32:45 (10.3 MB/s) - β€˜GeoLite2-City.tar.gz’ saved [27849181/27849181] # gunzip GeoLite2-City.tar.gz # tar xvf GeoLite2-City.tar GeoLite2-City_20190205/ GeoLite2-City_20190205/COPYRIGHT.txt GeoLite2-City_20190205/README.txt GeoLite2-City_20190205/LICENSE.txt GeoLite2-City_20190205/GeoLite2-City.mmdb root@kali-2018:~/Desktop/beef# ./asdf.rb {"continent"=>{"code"=>"NA", "geoname_id"=>6255149, "names"=>{"de"=>"Nordamerika", "en"=>"North America", "es"=>"NorteamΓ©rica", "fr"=>"AmΓ©rique du Nord", "ja"=>"εŒ—γ‚’γƒ‘γƒͺγ‚«", "pt-BR"=>"AmΓ©rica do Norte", "ru"=>"БСвСрная АмСрика", "zh-CN"=>"εŒ—ηΎŽζ΄²"}}, "country"=>{"geoname_id"=>6252001, "iso_code"=>"US", "names"=>{"de"=>"USA", "en"=>"United States", "es"=>"Estados Unidos", "fr"=>"Γ‰tats-Unis", "ja"=>"をパγƒͺγ‚«εˆθ‘†ε›½", "pt-BR"=>"Estados Unidos", "ru"=>"БША", "zh-CN"=>"ηΎŽε›½"}}, "location"=>{"accuracy_radius"=>1000, "latitude"=>37.751, "longitude"=>-97.822}, "registered_country"=>{"geoname_id"=>6252001, "iso_code"=>"US", "names"=>{"de"=>"USA", "en"=>"United States", "es"=>"Estados Unidos", "fr"=>"Γ‰tats-Unis", "ja"=>"をパγƒͺγ‚«εˆθ‘†ε›½", "pt-BR"=>"Estados Unidos", "ru"=>"БША", "zh-CN"=>"ηΎŽε›½"}}} # cat asdf.rb #!/usr/bin/env ruby # require 'maxmind/db' reader = MaxMind::DB.new('GeoLite2-City_20190205/GeoLite2-City.mmdb', mode: MaxMind::DB::MODE_MEMORY) record = reader.get('8.8.8.8') if record.nil? puts '8.8.8.8 not found' else puts record.inspect end ```
main
geolite legacy databases are now discontinued as of january the geoip gem and associated geoip functionality will need to be replaced with wget resolving geolite maxmind com geolite maxmind com connecting to geolite maxmind com geolite maxmind com connected http request sent awaiting response ok length saving to β€˜ city tar gz’ city tar gz s in mb s β€˜ city tar gz’ saved gunzip city tar gz tar xvf city tar city city copyright txt city readme txt city license txt city city mmdb root kali desktop beef asdf rb continent code na geoname id names de nordamerika en north america es norteamΓ©rica fr amΓ©rique du nord ja εŒ—γ‚’γƒ‘γƒͺγ‚« pt br amΓ©rica do norte ru сСвСрная Π°ΠΌΠ΅Ρ€ΠΈΠΊΠ° zh cn εŒ—ηΎŽζ΄² country geoname id iso code us names de usa en united states es estados unidos fr Γ©tats unis ja をパγƒͺγ‚«εˆθ‘†ε›½ pt br estados unidos ru сша zh cn ηΎŽε›½ location accuracy radius latitude longitude registered country geoname id iso code us names de usa en united states es estados unidos fr Γ©tats unis ja をパγƒͺγ‚«εˆθ‘†ε›½ pt br estados unidos ru сша zh cn ηΎŽε›½ cat asdf rb usr bin env ruby require maxmind db reader maxmind db new city city mmdb mode maxmind db mode memory record reader get if record nil puts not found else puts record inspect end
1
44,642
5,637,696,353
IssuesEvent
2017-04-06 09:50:15
Promact/trappist
https://api.github.com/repos/Promact/trappist
opened
Test Question Selection page
Test Creation and Management
we are supposing that The Test name , Edit button and preview button will be bound throughout the test page? is our assumption right?
1.0
Test Question Selection page - we are supposing that The Test name , Edit button and preview button will be bound throughout the test page? is our assumption right?
non_main
test question selection page we are supposing that the test name edit button and preview button will be bound throughout the test page is our assumption right
0
4,258
21,186,110,221
IssuesEvent
2022-04-08 12:55:26
jesus2099/konami-command
https://api.github.com/repos/jesus2099/konami-command
opened
Duplicate JASRAC work domain selection feature
ninja jasrac-mb-minc_WORK-IMPORT-CROSS-LINKING jasrac_DIRECT-LINK maintainability
[selectDomain()](https://github.com/jesus2099/konami-command/blob/f8c0f3af88395c7ba05cd6ab1e6dd4644f25d9e2/jasrac_DIRECT-LINK.user.js#L69) and [Select music release rights](https://github.com/jesus2099/konami-command/blob/40c4ef3565548d829d1db1ea77e398a95ddefb7a/jasrac-mb-minc_WORK-IMPORT-CROSS-LINKING.user.js#L417) want to do the same thing, concurrently.
True
Duplicate JASRAC work domain selection feature - [selectDomain()](https://github.com/jesus2099/konami-command/blob/f8c0f3af88395c7ba05cd6ab1e6dd4644f25d9e2/jasrac_DIRECT-LINK.user.js#L69) and [Select music release rights](https://github.com/jesus2099/konami-command/blob/40c4ef3565548d829d1db1ea77e398a95ddefb7a/jasrac-mb-minc_WORK-IMPORT-CROSS-LINKING.user.js#L417) want to do the same thing, concurrently.
main
duplicate jasrac work domain selection feature and want to do the same thing concurrently
1
3,255
12,402,316,546
IssuesEvent
2020-05-21 11:43:31
ocaml/opam-repository
https://api.github.com/repos/ocaml/opam-repository
closed
lambda-term.1.12.0 doesn't compile on 4.02.3
Stale needs maintainer action
```#=== ERROR while installing lambda-term.1.12.0 ================================# # opam-version 1.2.2 # os linux # command jbuilder build -p lambda-term -j 4 # path /home/jonathandav/.opam/4.02.3/build/lambda-term.1.12.0 # compiler 4.02.3 # exit-code 1 # env-file /home/jonathandav/.opam/4.02.3/build/lambda-term.1.12.0/lambda-term-23227-c266f8.env # stdout-file /home/jonathandav/.opam/4.02.3/build/lambda-term.1.12.0/lambda-term-23227-c266f8.out # stderr-file /home/jonathandav/.opam/4.02.3/build/lambda-term.1.12.0/lambda-term-23227-c266f8.err ### stderr ### # Warning 3: deprecated: Lwt_unix.execute_job # [...] # ocamlopt src/lTerm_resource_lexer.{cmx,o} # ocamlopt src/lTerm_mouse.{cmx,o} # ocamlc src/lTerm.{cmi,cmti} # ocamlc src/lTerm_draw.{cmo,cmt} # ocamlc src/lTerm_unix.{cmo,cmt} (exit 2) # (cd _build/default && /home/jonathandav/.opam/4.02.3/bin/ocamlc.opt -w -40 -safe-string -g -bin-annot -I /home/jonathandav/.opam/4.02.3/lib/bytes -I /home/jonathandav/.opam/4.02.3/lib/camomile -I /home/jonathandav/.opam/4.02.3/lib/lwt -I /home/jonathandav/.opam/4.02.3/lib/lwt_react -I /home/jonathandav/.opam/4.02.3/lib/ocaml -I /home/jonathandav/.opam/4.02.3/lib/react -I /home/jonathandav/.opam/4.02.3/lib/result -I /home/jonathandav/.opam/4.02.3/lib/zed -no-alias-deps -I src -o src/lTerm_unix.cmo -c -impl src/lTerm_unix.ml) # File "src/lTerm_unix.ml", line 342, characters 32-51: # Error: This expression has type bytes but an expression was expected of type # string ```
True
lambda-term.1.12.0 doesn't compile on 4.02.3 - ```#=== ERROR while installing lambda-term.1.12.0 ================================# # opam-version 1.2.2 # os linux # command jbuilder build -p lambda-term -j 4 # path /home/jonathandav/.opam/4.02.3/build/lambda-term.1.12.0 # compiler 4.02.3 # exit-code 1 # env-file /home/jonathandav/.opam/4.02.3/build/lambda-term.1.12.0/lambda-term-23227-c266f8.env # stdout-file /home/jonathandav/.opam/4.02.3/build/lambda-term.1.12.0/lambda-term-23227-c266f8.out # stderr-file /home/jonathandav/.opam/4.02.3/build/lambda-term.1.12.0/lambda-term-23227-c266f8.err ### stderr ### # Warning 3: deprecated: Lwt_unix.execute_job # [...] # ocamlopt src/lTerm_resource_lexer.{cmx,o} # ocamlopt src/lTerm_mouse.{cmx,o} # ocamlc src/lTerm.{cmi,cmti} # ocamlc src/lTerm_draw.{cmo,cmt} # ocamlc src/lTerm_unix.{cmo,cmt} (exit 2) # (cd _build/default && /home/jonathandav/.opam/4.02.3/bin/ocamlc.opt -w -40 -safe-string -g -bin-annot -I /home/jonathandav/.opam/4.02.3/lib/bytes -I /home/jonathandav/.opam/4.02.3/lib/camomile -I /home/jonathandav/.opam/4.02.3/lib/lwt -I /home/jonathandav/.opam/4.02.3/lib/lwt_react -I /home/jonathandav/.opam/4.02.3/lib/ocaml -I /home/jonathandav/.opam/4.02.3/lib/react -I /home/jonathandav/.opam/4.02.3/lib/result -I /home/jonathandav/.opam/4.02.3/lib/zed -no-alias-deps -I src -o src/lTerm_unix.cmo -c -impl src/lTerm_unix.ml) # File "src/lTerm_unix.ml", line 342, characters 32-51: # Error: This expression has type bytes but an expression was expected of type # string ```
main
lambda term doesn t compile on error while installing lambda term opam version os linux command jbuilder build p lambda term j path home jonathandav opam build lambda term compiler exit code env file home jonathandav opam build lambda term lambda term env stdout file home jonathandav opam build lambda term lambda term out stderr file home jonathandav opam build lambda term lambda term err stderr warning deprecated lwt unix execute job ocamlopt src lterm resource lexer cmx o ocamlopt src lterm mouse cmx o ocamlc src lterm cmi cmti ocamlc src lterm draw cmo cmt ocamlc src lterm unix cmo cmt exit cd build default home jonathandav opam bin ocamlc opt w safe string g bin annot i home jonathandav opam lib bytes i home jonathandav opam lib camomile i home jonathandav opam lib lwt i home jonathandav opam lib lwt react i home jonathandav opam lib ocaml i home jonathandav opam lib react i home jonathandav opam lib result i home jonathandav opam lib zed no alias deps i src o src lterm unix cmo c impl src lterm unix ml file src lterm unix ml line characters error this expression has type bytes but an expression was expected of type string
1
909
4,579,007,012
IssuesEvent
2016-09-18 01:40:42
daemonraco/toobasic
https://api.github.com/repos/daemonraco/toobasic
opened
JSON Validator for Database Specs
Core Logic Database Database Structure Maintainer JSONValidator
## What to do Implement validations for database specifications using _JSON Validator_, both version 1 and 2.
True
JSON Validator for Database Specs - ## What to do Implement validations for database specifications using _JSON Validator_, both version 1 and 2.
main
json validator for database specs what to do implement validations for database specifications using json validator both version and
1
6,564
9,550,086,414
IssuesEvent
2019-05-02 11:03:45
adaptlearning/adapt_authoring
https://api.github.com/repos/adaptlearning/adapt_authoring
opened
Tenant management alternatives
T: requirements
The purpose to this issue is to explore alternatives to tenant management. Please list any requirements you may have of tenant management.
1.0
Tenant management alternatives - The purpose to this issue is to explore alternatives to tenant management. Please list any requirements you may have of tenant management.
non_main
tenant management alternatives the purpose to this issue is to explore alternatives to tenant management please list any requirements you may have of tenant management
0
2,770
27,590,655,135
IssuesEvent
2023-03-08 23:59:00
Azure/azure-functions-host
https://api.github.com/repos/Azure/azure-functions-host
opened
Add heartbeat process to detect thread pool exhaustion
reliability
#### What problem would the feature you're requesting solve? Please describe. Currently, there is no automated diagnostic in place that would detect deadlock/thread pool exhaustion in the Host. The only mode we have right now to detect thread pool exhaustion is by manually capturing and examining a memory dump. #### Describe the solution you'd like Add a heartbeat monitor process that pings the host on a regular schedule, and detects thread pool exhaustion scenarios. #### Additional context Once the heartbeat is in place, we will add one of two options for mitigation of the condition detected: 1. Auto-restart the host or 2. Report this information through the status API's to the Antares platform for auto-mitigation/ memory dump collection.
True
Add heartbeat process to detect thread pool exhaustion - #### What problem would the feature you're requesting solve? Please describe. Currently, there is no automated diagnostic in place that would detect deadlock/thread pool exhaustion in the Host. The only mode we have right now to detect thread pool exhaustion is by manually capturing and examining a memory dump. #### Describe the solution you'd like Add a heartbeat monitor process that pings the host on a regular schedule, and detects thread pool exhaustion scenarios. #### Additional context Once the heartbeat is in place, we will add one of two options for mitigation of the condition detected: 1. Auto-restart the host or 2. Report this information through the status API's to the Antares platform for auto-mitigation/ memory dump collection.
non_main
add heartbeat process to detect thread pool exhaustion what problem would the feature you re requesting solve please describe currently there is no automated diagnostic in place that would detect deadlock thread pool exhaustion in the host the only mode we have right now to detect thread pool exhaustion is by manually capturing and examining a memory dump describe the solution you d like add a heartbeat monitor process that pings the host on a regular schedule and detects thread pool exhaustion scenarios additional context once the heartbeat is in place we will add one of two options for mitigation of the condition detected auto restart the host or report this information through the status api s to the antares platform for auto mitigation memory dump collection
0
105,666
13,205,237,148
IssuesEvent
2020-08-14 17:31:18
toggl/mobileapp
https://api.github.com/repos/toggl/mobileapp
opened
Visual changes in CalendarSettingsViewController
estimate:M ios needs-design
## Description Implement the visual changes in CalendarSettingsViewController ## Dependencies Depends on #7713 ## Definition of done - [ ] The visual changes are implemented
1.0
Visual changes in CalendarSettingsViewController - ## Description Implement the visual changes in CalendarSettingsViewController ## Dependencies Depends on #7713 ## Definition of done - [ ] The visual changes are implemented
non_main
visual changes in calendarsettingsviewcontroller description implement the visual changes in calendarsettingsviewcontroller dependencies depends on definition of done the visual changes are implemented
0
96,929
3,975,632,230
IssuesEvent
2016-05-05 06:54:51
Sententiaregum/Sententiaregum
https://api.github.com/repos/Sententiaregum/Sententiaregum
opened
improve configuration handler of capistrano
Deployment Improvement Low priority
### Description the capistrano deployment handler is responsible for setting up the whole configuration. Instead multiple files should be merged, so the heavy config loader can be simplified.
1.0
improve configuration handler of capistrano - ### Description the capistrano deployment handler is responsible for setting up the whole configuration. Instead multiple files should be merged, so the heavy config loader can be simplified.
non_main
improve configuration handler of capistrano description the capistrano deployment handler is responsible for setting up the whole configuration instead multiple files should be merged so the heavy config loader can be simplified
0
14,577
2,829,609,333
IssuesEvent
2015-05-23 02:05:30
awesomebing1/fuzzdb
https://api.github.com/repos/awesomebing1/fuzzdb
closed
http://www.rf-dimension.com/forum/entry.php?70656-Football%28Chicago-Bears-vs-Green-Bay-Packers%29-Live-Streaming
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. 2. 3. http://www.rf-dimension.com/forum/entry.php?70656-Football%28Chicago-Bears-vs-Gr een-Bay-Packers%29-Live-Streaming http://www.rf-dimension.com/forum/entry.php?70656-Football%28Chicago-Bears-vs-Gr een-Bay-Packers%29-Live-Streaming What is the expected output? What do you see instead? What version of the product are you using? On what operating system? Please provide any additional information below. ``` Original issue reported on code.google.com by `sabujhos...@gmail.com` on 9 Nov 2014 at 2:47
1.0
http://www.rf-dimension.com/forum/entry.php?70656-Football%28Chicago-Bears-vs-Green-Bay-Packers%29-Live-Streaming - ``` What steps will reproduce the problem? 1. 2. 3. http://www.rf-dimension.com/forum/entry.php?70656-Football%28Chicago-Bears-vs-Gr een-Bay-Packers%29-Live-Streaming http://www.rf-dimension.com/forum/entry.php?70656-Football%28Chicago-Bears-vs-Gr een-Bay-Packers%29-Live-Streaming What is the expected output? What do you see instead? What version of the product are you using? On what operating system? Please provide any additional information below. ``` Original issue reported on code.google.com by `sabujhos...@gmail.com` on 9 Nov 2014 at 2:47
non_main
what steps will reproduce the problem een bay packers live streaming een bay packers live streaming what is the expected output what do you see instead what version of the product are you using on what operating system please provide any additional information below original issue reported on code google com by sabujhos gmail com on nov at
0
321,843
27,560,601,127
IssuesEvent
2023-03-07 21:39:27
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
DISABLED test_dynamic_shapes_right_side_dynamic_shapes (torch._dynamo.testing.DynamicShapesReproTests)
triaged module: flaky-tests skipped module: dynamo
Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_dynamic_shapes_right_side_dynamic_shapes&suite=torch._dynamo.testing.DynamicShapesReproTests) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/11828458327). Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_dynamic_shapes_right_side_dynamic_shapes` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/testing.py` or `/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/testing.py`
1.0
DISABLED test_dynamic_shapes_right_side_dynamic_shapes (torch._dynamo.testing.DynamicShapesReproTests) - Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_dynamic_shapes_right_side_dynamic_shapes&suite=torch._dynamo.testing.DynamicShapesReproTests) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/11828458327). Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_dynamic_shapes_right_side_dynamic_shapes` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/testing.py` or `/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/testing.py`
non_main
disabled test dynamic shapes right side dynamic shapes torch dynamo testing dynamicshapesreprotests platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test dynamic shapes right side dynamic shapes there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path opt conda envs py lib site packages torch dynamo testing py or opt conda envs py lib site packages torch dynamo testing py
0
5,235
26,551,134,376
IssuesEvent
2023-01-20 07:51:49
AhmadWaleed/laravel-blanket
https://api.github.com/repos/AhmadWaleed/laravel-blanket
opened
Looking for maintainer.
help wanted Looking For Maintainer
It's been a while since i worked in php/laravel, I don't have to time to maintain this package anymore. If anybody interested volunteering himself for becoming active maintainer for this project. Feel free to reach out to me. If i could not find any maintainer in few weeks, unfortunately i will archive this repository.
True
Looking for maintainer. - It's been a while since i worked in php/laravel, I don't have to time to maintain this package anymore. If anybody interested volunteering himself for becoming active maintainer for this project. Feel free to reach out to me. If i could not find any maintainer in few weeks, unfortunately i will archive this repository.
main
looking for maintainer it s been a while since i worked in php laravel i don t have to time to maintain this package anymore if anybody interested volunteering himself for becoming active maintainer for this project feel free to reach out to me if i could not find any maintainer in few weeks unfortunately i will archive this repository
1
349,254
24,939,968,298
IssuesEvent
2022-10-31 18:04:32
music-encoding/music-encoding.github.io
https://api.github.com/repos/music-encoding/music-encoding.github.io
closed
Link Checker Report
documentation bug
## Summary | Status | Count | |---------------|-------| | πŸ” Total | 432 | | βœ… Successful | 199 | | ⏳ Timeouts | 0 | | πŸ”€ Redirected | 0 | | πŸ‘» Excluded | 232 | | ❓ Unknown | 0 | | 🚫 Errors | 1 | ## Errors per input ### Errors in ./resources/tools.md * [http://custom.music-encoding.org/](http://custom.music-encoding.org/): Failed: Network error (status code: 404) [Full Github Actions output](https://github.com/music-encoding/music-encoding.github.io/actions/runs/3355658640?check_suite_focus=true)
1.0
Link Checker Report - ## Summary | Status | Count | |---------------|-------| | πŸ” Total | 432 | | βœ… Successful | 199 | | ⏳ Timeouts | 0 | | πŸ”€ Redirected | 0 | | πŸ‘» Excluded | 232 | | ❓ Unknown | 0 | | 🚫 Errors | 1 | ## Errors per input ### Errors in ./resources/tools.md * [http://custom.music-encoding.org/](http://custom.music-encoding.org/): Failed: Network error (status code: 404) [Full Github Actions output](https://github.com/music-encoding/music-encoding.github.io/actions/runs/3355658640?check_suite_focus=true)
non_main
link checker report summary status count πŸ” total βœ… successful ⏳ timeouts πŸ”€ redirected πŸ‘» excluded ❓ unknown 🚫 errors errors per input errors in resources tools md failed network error status code
0
627,699
19,912,356,818
IssuesEvent
2022-01-25 18:29:25
exalearn/EXARL
https://api.github.com/repos/exalearn/EXARL
closed
Do we need to save current_state and next_state?
Low Priority
learner_base.py: 158 train_writer.writerow([current_state,action,reward,next_state,total_reward,done]) Basically, each state is saved twice...
1.0
Do we need to save current_state and next_state? - learner_base.py: 158 train_writer.writerow([current_state,action,reward,next_state,total_reward,done]) Basically, each state is saved twice...
non_main
do we need to save current state and next state learner base py train writer writerow basically each state is saved twice
0
315,739
9,631,612,191
IssuesEvent
2019-05-15 14:33:35
canonical-web-and-design/snapcraft.io
https://api.github.com/repos/canonical-web-and-design/snapcraft.io
closed
Automate notifications in snapcraft.io
Priority: Medium
Recently we have been including banners to promote live stream notifications on youtube but would be good to find a way to automate this piece of work, so we dont need to (a) add and remove code manually, (b) external people like advocacy can schedule without our intervention and (c) #winning.
1.0
Automate notifications in snapcraft.io - Recently we have been including banners to promote live stream notifications on youtube but would be good to find a way to automate this piece of work, so we dont need to (a) add and remove code manually, (b) external people like advocacy can schedule without our intervention and (c) #winning.
non_main
automate notifications in snapcraft io recently we have been including banners to promote live stream notifications on youtube but would be good to find a way to automate this piece of work so we dont need to a add and remove code manually b external people like advocacy can schedule without our intervention and c winning
0
199,996
6,996,998,843
IssuesEvent
2017-12-16 08:51:00
giangm9/enduel
https://api.github.com/repos/giangm9/enduel
closed
Sα»­a lαΊ‘i thời gian vΓ  hiển thα»‹ khi accept trαΊ­n
High priority
- [x] Thời gian đợi αΊ₯n Accept quΓ‘ ngαΊ―n, bα»‹ bug ![capture1](https://user-images.githubusercontent.com/25172258/34034113-b5ac128e-e1af-11e7-8efa-402f8f3cfd42.PNG) - [x] Found match cαΊ§n hiện thα»‹ tgian Δ‘αΊΏm ngược trΖ°α»›c khi time out. VD: Found match 10->9->8..... ![capture](https://user-images.githubusercontent.com/25172258/34034117-b733b242-e1af-11e7-87ad-3d32ff3adde1.PNG)
1.0
Sα»­a lαΊ‘i thời gian vΓ  hiển thα»‹ khi accept trαΊ­n - - [x] Thời gian đợi αΊ₯n Accept quΓ‘ ngαΊ―n, bα»‹ bug ![capture1](https://user-images.githubusercontent.com/25172258/34034113-b5ac128e-e1af-11e7-8efa-402f8f3cfd42.PNG) - [x] Found match cαΊ§n hiện thα»‹ tgian Δ‘αΊΏm ngược trΖ°α»›c khi time out. VD: Found match 10->9->8..... ![capture](https://user-images.githubusercontent.com/25172258/34034117-b733b242-e1af-11e7-87ad-3d32ff3adde1.PNG)
non_main
sα»­a lαΊ‘i thời gian vΓ  hiển thα»‹ khi accept trαΊ­n thời gian đợi αΊ₯n accept quΓ‘ ngαΊ―n bα»‹ bug found match cαΊ§n hiện thα»‹ tgian Δ‘αΊΏm ngược trΖ°α»›c khi time out vd found match
0
220,928
7,372,235,202
IssuesEvent
2018-03-13 14:15:18
mlibrary/heliotrope
https://api.github.com/repos/mlibrary/heliotrope
opened
Stop Unity's pop-up alert for every JS error
bug gabii high priority
- [ ] Add a NOP error handler (see link) to the 2 places in Heliotrope where the Unity engine is instantiated - [ ] verify this stops unrelated JS errors from giving a pop-up alert https://forum.unity.com/threads/make-javascript-errors-not-alert-in-5-6.466772/
1.0
Stop Unity's pop-up alert for every JS error - - [ ] Add a NOP error handler (see link) to the 2 places in Heliotrope where the Unity engine is instantiated - [ ] verify this stops unrelated JS errors from giving a pop-up alert https://forum.unity.com/threads/make-javascript-errors-not-alert-in-5-6.466772/
non_main
stop unity s pop up alert for every js error add a nop error handler see link to the places in heliotrope where the unity engine is instantiated verify this stops unrelated js errors from giving a pop up alert
0
148,342
11,848,162,831
IssuesEvent
2020-03-24 13:21:02
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI][master] DocsClientYamlTestSuiteIT put-auto-follow-pattern
:Distributed/CCR >test-failure
Log : https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+multijob-unix-compatibility/os=centos-8&&immutable/627/console Build Scans: https://gradle-enterprise.elastic.co/s/t7pibrqbzjvsm REPRODUCE WITH: ./gradlew ':docs:integTestRunner' --tests "org.elasticsearch.smoketest.DocsClientYamlTestSuiteIT.test {yaml=reference/ccr/apis/auto-follow/put-auto-follow-pattern/line_15}" \ -Dtests.seed=1FB43FBC4EBFC245 \ -Dtests.security.manager=true \ -Dtests.locale=da \ -Dtests.timezone=America/Cuiaba \ -Dcompiler.java=13 Doesn't reproduce for me ``` Caused by: java.lang.AssertionError: expected [2xx] status code but api [raw[method=PUT path=_ccr/auto_follow/auto_follow_pattern_name]] returned [400 Bad Request] [{"error":{"root_cause":[{"type":"resource_not_found_exception","reason":"license info is missing for cluster [remote_cluster]","stack_trace":"org.elasticsearch.ResourceNotFoundException: license info is missing for cluster [remote_cluster]\n\tat org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:168)\n\tat org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:162)\n\tat org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43)\n\tat org.elasticsearch.action.ActionListenerResponseHandler.handleResponse(ActionListenerResponseHandler.java:54)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1057)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1057)\n\tat org.elasticsearch.transport.InboundHandler$1.doRun(InboundHandler.java:220)\n\tat org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37)\n\tat org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:177)\n\tat org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:212)\n\tat org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:138)\n\tat org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:102)\n\tat org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:667)\n\tat org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:321)\n\tat io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:295)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:227)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)\n\tat io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:615)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:578)\n\tat io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493)\n\tat io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)\n\tat io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)\n\tat java.base/java.lang.Thread.run(Thread.java:834)\n"}],"type":"status_exception","reason":"can not fetch remote cluster state as the license state of the remote cluster [remote_cluster] could not be determined","caused_by":{"type":"resource_not_found_exception","reason":"license info is missing for cluster [remote_cluster]","stack_trace":"org.elasticsearch.ResourceNotFoundException: license info is missing for cluster [remote_cluster]\n\tat org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:168)\n\tat org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:162)\n\tat org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43)\n\tat org.elasticsearch.action.ActionListenerResponseHandler.handleResponse(ActionListenerResponseHandler.java:54)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1057)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1057)\n\tat org.elasticsearch.transport.InboundHandler$1.doRun(InboundHandler.java:220)\n\tat org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37)\n\tat org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:177)\n\tat org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:212)\n\tat org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:138)\n\tat org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:102)\n\tat org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:667)\n\tat org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:321)\n\tat io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:295)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:227)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)\n\tat io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:615)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:578)\n\tat io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493)\n\tat io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)\n\tat io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)\n\tat java.base/java.lang.Thread.run(Thread.java:834)\n"},"stack_trace":"org.elasticsearch.ElasticsearchStatusException: can not fetch remote cluster state as the license state of the remote cluster [remote_cluster] could not be determined\n\tat org.elasticsearch.xpack.ccr.CcrLicenseChecker.clusterStateUnknownRemoteLicense(CcrLicenseChecker.java:448)\n\tat org.elasticsearch.xpack.ccr.CcrLicenseChecker.lambda$checkRemoteClusterLicenseAndFetchClusterState$5(CcrLicenseChecker.java:180)\n\tat org.elasticsearch.xpack.ccr.CcrLicenseChecker$1.onFailure(CcrLicenseChecker.java:231)\n\tat org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:168)\n\tat org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:162)\n\tat org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43)\n\tat org.elasticsearch.action.ActionListenerResponseHandler.handleResponse(ActionListenerResponseHandler.java:54)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1057)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1057)\n\tat org.elasticsearch.transport.InboundHandler$1.doRun(InboundHandler.java:220)\n\tat org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37)\n\tat org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:177)\n\tat org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:212)\n\tat org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:138)\n\tat org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:102)\n\tat org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:667)\n\tat org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:321)\n\tat io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:295)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:227)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)\n\tat io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:615)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:578)\n\tat io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493)\n\tat io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)\n\tat io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)\n\tat java.base/java.lang.Thread.run(Thread.java:834)\nCaused by: org.elasticsearch.ResourceNotFoundException: license info is missing for cluster [remote_cluster]\n\t... 38 more\n"},"status":400}]Open stacktrace [2020-03-11T14:07:09,441][INFO ][o.e.s.DocsClientYamlTestSuiteIT] [test] [yaml=reference/ccr/apis/auto-follow/put-auto-follow-pattern/line_15] before test [2020-03-11T14:07:09,674][INFO ][o.e.s.DocsClientYamlTestSuiteIT] [test] initializing REST clients against [http://[::1]:34583, http://127.0.0.1:36961] [2020-03-11T14:07:10,741][INFO ][o.e.s.DocsClientYamlTestSuiteIT] [test] initializing client, minimum es version [8.0.0], master version, [8.0.0], hosts [http://[::1]:34583, http://127.0.0.1:36961] [2020-03-11T14:07:11,405][INFO ][o.e.s.DocsClientYamlTestSuiteIT] [test] Stash dump on test failure [{ "stash" : { "host" : "127.0.0.1:36961", "body" : { "error" : { "root_cause" : [ { "type" : "resource_not_found_exception", "reason" : "license info is missing for cluster [remote_cluster]", "stack_trace" : "org.elasticsearch.ResourceNotFoundException: license info is missing for cluster [remote_cluster] at org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:168) at org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:162) at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) ``` Possibly related to : https://github.com/elastic/elasticsearch/issues/36815
1.0
[CI][master] DocsClientYamlTestSuiteIT put-auto-follow-pattern - Log : https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+multijob-unix-compatibility/os=centos-8&&immutable/627/console Build Scans: https://gradle-enterprise.elastic.co/s/t7pibrqbzjvsm REPRODUCE WITH: ./gradlew ':docs:integTestRunner' --tests "org.elasticsearch.smoketest.DocsClientYamlTestSuiteIT.test {yaml=reference/ccr/apis/auto-follow/put-auto-follow-pattern/line_15}" \ -Dtests.seed=1FB43FBC4EBFC245 \ -Dtests.security.manager=true \ -Dtests.locale=da \ -Dtests.timezone=America/Cuiaba \ -Dcompiler.java=13 Doesn't reproduce for me ``` Caused by: java.lang.AssertionError: expected [2xx] status code but api [raw[method=PUT path=_ccr/auto_follow/auto_follow_pattern_name]] returned [400 Bad Request] [{"error":{"root_cause":[{"type":"resource_not_found_exception","reason":"license info is missing for cluster [remote_cluster]","stack_trace":"org.elasticsearch.ResourceNotFoundException: license info is missing for cluster [remote_cluster]\n\tat org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:168)\n\tat org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:162)\n\tat org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43)\n\tat org.elasticsearch.action.ActionListenerResponseHandler.handleResponse(ActionListenerResponseHandler.java:54)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1057)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1057)\n\tat org.elasticsearch.transport.InboundHandler$1.doRun(InboundHandler.java:220)\n\tat org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37)\n\tat org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:177)\n\tat org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:212)\n\tat org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:138)\n\tat org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:102)\n\tat org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:667)\n\tat org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:321)\n\tat io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:295)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:227)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)\n\tat io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:615)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:578)\n\tat io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493)\n\tat io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)\n\tat io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)\n\tat java.base/java.lang.Thread.run(Thread.java:834)\n"}],"type":"status_exception","reason":"can not fetch remote cluster state as the license state of the remote cluster [remote_cluster] could not be determined","caused_by":{"type":"resource_not_found_exception","reason":"license info is missing for cluster [remote_cluster]","stack_trace":"org.elasticsearch.ResourceNotFoundException: license info is missing for cluster [remote_cluster]\n\tat org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:168)\n\tat org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:162)\n\tat org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43)\n\tat org.elasticsearch.action.ActionListenerResponseHandler.handleResponse(ActionListenerResponseHandler.java:54)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1057)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1057)\n\tat org.elasticsearch.transport.InboundHandler$1.doRun(InboundHandler.java:220)\n\tat org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37)\n\tat org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:177)\n\tat org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:212)\n\tat org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:138)\n\tat org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:102)\n\tat org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:667)\n\tat org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:321)\n\tat io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:295)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:227)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)\n\tat io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:615)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:578)\n\tat io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493)\n\tat io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)\n\tat io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)\n\tat java.base/java.lang.Thread.run(Thread.java:834)\n"},"stack_trace":"org.elasticsearch.ElasticsearchStatusException: can not fetch remote cluster state as the license state of the remote cluster [remote_cluster] could not be determined\n\tat org.elasticsearch.xpack.ccr.CcrLicenseChecker.clusterStateUnknownRemoteLicense(CcrLicenseChecker.java:448)\n\tat org.elasticsearch.xpack.ccr.CcrLicenseChecker.lambda$checkRemoteClusterLicenseAndFetchClusterState$5(CcrLicenseChecker.java:180)\n\tat org.elasticsearch.xpack.ccr.CcrLicenseChecker$1.onFailure(CcrLicenseChecker.java:231)\n\tat org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:168)\n\tat org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:162)\n\tat org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43)\n\tat org.elasticsearch.action.ActionListenerResponseHandler.handleResponse(ActionListenerResponseHandler.java:54)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1057)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1057)\n\tat org.elasticsearch.transport.InboundHandler$1.doRun(InboundHandler.java:220)\n\tat org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37)\n\tat org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:177)\n\tat org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:212)\n\tat org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:138)\n\tat org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:102)\n\tat org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:667)\n\tat org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:321)\n\tat io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:295)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:227)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:355)\n\tat io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:377)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:363)\n\tat io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)\n\tat io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:615)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:578)\n\tat io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493)\n\tat io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)\n\tat io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)\n\tat java.base/java.lang.Thread.run(Thread.java:834)\nCaused by: org.elasticsearch.ResourceNotFoundException: license info is missing for cluster [remote_cluster]\n\t... 38 more\n"},"status":400}]Open stacktrace [2020-03-11T14:07:09,441][INFO ][o.e.s.DocsClientYamlTestSuiteIT] [test] [yaml=reference/ccr/apis/auto-follow/put-auto-follow-pattern/line_15] before test [2020-03-11T14:07:09,674][INFO ][o.e.s.DocsClientYamlTestSuiteIT] [test] initializing REST clients against [http://[::1]:34583, http://127.0.0.1:36961] [2020-03-11T14:07:10,741][INFO ][o.e.s.DocsClientYamlTestSuiteIT] [test] initializing client, minimum es version [8.0.0], master version, [8.0.0], hosts [http://[::1]:34583, http://127.0.0.1:36961] [2020-03-11T14:07:11,405][INFO ][o.e.s.DocsClientYamlTestSuiteIT] [test] Stash dump on test failure [{ "stash" : { "host" : "127.0.0.1:36961", "body" : { "error" : { "root_cause" : [ { "type" : "resource_not_found_exception", "reason" : "license info is missing for cluster [remote_cluster]", "stack_trace" : "org.elasticsearch.ResourceNotFoundException: license info is missing for cluster [remote_cluster] at org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:168) at org.elasticsearch.license.RemoteClusterLicenseChecker$1.onResponse(RemoteClusterLicenseChecker.java:162) at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) ``` Possibly related to : https://github.com/elastic/elasticsearch/issues/36815
non_main
docsclientyamltestsuiteit put auto follow pattern log build scans reproduce with gradlew docs integtestrunner tests org elasticsearch smoketest docsclientyamltestsuiteit test yaml reference ccr apis auto follow put auto follow pattern line dtests seed dtests security manager true dtests locale da dtests timezone america cuiaba dcompiler java doesn t reproduce for me caused by java lang assertionerror expected status code but api returned stack trace org elasticsearch resourcenotfoundexception license info is missing for cluster n tat org elasticsearch license remoteclusterlicensechecker onresponse remoteclusterlicensechecker java n tat org elasticsearch license remoteclusterlicensechecker onresponse remoteclusterlicensechecker java n tat org elasticsearch action support contextpreservingactionlistener onresponse contextpreservingactionlistener java n tat org elasticsearch action actionlistenerresponsehandler handleresponse actionlistenerresponsehandler java n tat org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch transport inboundhandler dorun inboundhandler java n tat org elasticsearch common util concurrent abstractrunnable run abstractrunnable java n tat org elasticsearch common util concurrent esexecutors directexecutorservice execute esexecutors java n tat org elasticsearch transport inboundhandler handleresponse inboundhandler java n tat org elasticsearch transport inboundhandler messagereceived inboundhandler java n tat org elasticsearch transport inboundhandler inboundmessage inboundhandler java n tat org elasticsearch transport tcptransport inboundmessage tcptransport java n tat org elasticsearch transport channelread java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java n tat io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty handler logging logginghandler channelread logginghandler java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java n tat io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java n tat io netty channel nio nioeventloop processselectedkey nioeventloop java n tat io netty channel nio nioeventloop processselectedkeysplain nioeventloop java n tat io netty channel nio nioeventloop processselectedkeys nioeventloop java n tat io netty channel nio nioeventloop run nioeventloop java n tat io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java n tat io netty util internal threadexecutormap run threadexecutormap java n tat java base java lang thread run thread java n type status exception reason can not fetch remote cluster state as the license state of the remote cluster could not be determined caused by type resource not found exception reason license info is missing for cluster stack trace org elasticsearch resourcenotfoundexception license info is missing for cluster n tat org elasticsearch license remoteclusterlicensechecker onresponse remoteclusterlicensechecker java n tat org elasticsearch license remoteclusterlicensechecker onresponse remoteclusterlicensechecker java n tat org elasticsearch action support contextpreservingactionlistener onresponse contextpreservingactionlistener java n tat org elasticsearch action actionlistenerresponsehandler handleresponse actionlistenerresponsehandler java n tat org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch transport inboundhandler dorun inboundhandler java n tat org elasticsearch common util concurrent abstractrunnable run abstractrunnable java n tat org elasticsearch common util concurrent esexecutors directexecutorservice execute esexecutors java n tat org elasticsearch transport inboundhandler handleresponse inboundhandler java n tat org elasticsearch transport inboundhandler messagereceived inboundhandler java n tat org elasticsearch transport inboundhandler inboundmessage inboundhandler java n tat org elasticsearch transport tcptransport inboundmessage tcptransport java n tat org elasticsearch transport channelread java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java n tat io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty handler logging logginghandler channelread logginghandler java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java n tat io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java n tat io netty channel nio nioeventloop processselectedkey nioeventloop java n tat io netty channel nio nioeventloop processselectedkeysplain nioeventloop java n tat io netty channel nio nioeventloop processselectedkeys nioeventloop java n tat io netty channel nio nioeventloop run nioeventloop java n tat io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java n tat io netty util internal threadexecutormap run threadexecutormap java n tat java base java lang thread run thread java n stack trace org elasticsearch elasticsearchstatusexception can not fetch remote cluster state as the license state of the remote cluster could not be determined n tat org elasticsearch xpack ccr ccrlicensechecker clusterstateunknownremotelicense ccrlicensechecker java n tat org elasticsearch xpack ccr ccrlicensechecker lambda checkremoteclusterlicenseandfetchclusterstate ccrlicensechecker java n tat org elasticsearch xpack ccr ccrlicensechecker onfailure ccrlicensechecker java n tat org elasticsearch license remoteclusterlicensechecker onresponse remoteclusterlicensechecker java n tat org elasticsearch license remoteclusterlicensechecker onresponse remoteclusterlicensechecker java n tat org elasticsearch action support contextpreservingactionlistener onresponse contextpreservingactionlistener java n tat org elasticsearch action actionlistenerresponsehandler handleresponse actionlistenerresponsehandler java n tat org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch transport inboundhandler dorun inboundhandler java n tat org elasticsearch common util concurrent abstractrunnable run abstractrunnable java n tat org elasticsearch common util concurrent esexecutors directexecutorservice execute esexecutors java n tat org elasticsearch transport inboundhandler handleresponse inboundhandler java n tat org elasticsearch transport inboundhandler messagereceived inboundhandler java n tat org elasticsearch transport inboundhandler inboundmessage inboundhandler java n tat org elasticsearch transport tcptransport inboundmessage tcptransport java n tat org elasticsearch transport channelread java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java n tat io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty handler logging logginghandler channelread logginghandler java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java n tat io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java n tat io netty channel nio nioeventloop processselectedkey nioeventloop java n tat io netty channel nio nioeventloop processselectedkeysplain nioeventloop java n tat io netty channel nio nioeventloop processselectedkeys nioeventloop java n tat io netty channel nio nioeventloop run nioeventloop java n tat io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java n tat io netty util internal threadexecutormap run threadexecutormap java n tat java base java lang thread run thread java ncaused by org elasticsearch resourcenotfoundexception license info is missing for cluster n t more n status open stacktrace before test initializing rest clients against initializing client minimum es version master version hosts stash dump on test failure stash host body error root cause type resource not found exception reason license info is missing for cluster stack trace org elasticsearch resourcenotfoundexception license info is missing for cluster at org elasticsearch license remoteclusterlicensechecker onresponse remoteclusterlicensechecker java at org elasticsearch license remoteclusterlicensechecker onresponse remoteclusterlicensechecker java at org elasticsearch action support contextpreservingactionlistener onresponse contextpreservingactionlistener java possibly related to
0
4,264
21,263,934,782
IssuesEvent
2022-04-13 08:06:19
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Dependency error when running `mvn clean install`
kind/toil area/maintainability
**Description** When checking out a recent main (e.g. 16d0e6811228fa51e36ea96bc8fac4af6a78fb20) and running `mvn clean install` the build fails with a dependency error. ``` [INFO] --- maven-dependency-plugin:3.3.0:analyze-only (analyze-dependencies) @ zeebe-atomix-cluster --- [WARNING] Unused declared dependencies found: [WARNING] org.junit.jupiter:junit-jupiter-engine:jar:5.8.2:test [WARNING] org.junit.vintage:junit-vintage-engine:jar:5.8.2:test ``` However, the same commit passes on Jenkins.
True
Dependency error when running `mvn clean install` - **Description** When checking out a recent main (e.g. 16d0e6811228fa51e36ea96bc8fac4af6a78fb20) and running `mvn clean install` the build fails with a dependency error. ``` [INFO] --- maven-dependency-plugin:3.3.0:analyze-only (analyze-dependencies) @ zeebe-atomix-cluster --- [WARNING] Unused declared dependencies found: [WARNING] org.junit.jupiter:junit-jupiter-engine:jar:5.8.2:test [WARNING] org.junit.vintage:junit-vintage-engine:jar:5.8.2:test ``` However, the same commit passes on Jenkins.
main
dependency error when running mvn clean install description when checking out a recent main e g and running mvn clean install the build fails with a dependency error maven dependency plugin analyze only analyze dependencies zeebe atomix cluster unused declared dependencies found org junit jupiter junit jupiter engine jar test org junit vintage junit vintage engine jar test however the same commit passes on jenkins
1
142,549
21,784,077,633
IssuesEvent
2022-05-13 23:15:18
dotnet/efcore
https://api.github.com/repos/dotnet/efcore
closed
Support FromSql with table splitting
type-enhancement needs-design area-query
Need to investigate in what scenarios it makes sense to use them together and decide whether we'll support them or throw preemptively.
1.0
Support FromSql with table splitting - Need to investigate in what scenarios it makes sense to use them together and decide whether we'll support them or throw preemptively.
non_main
support fromsql with table splitting need to investigate in what scenarios it makes sense to use them together and decide whether we ll support them or throw preemptively
0
5,566
27,840,805,694
IssuesEvent
2023-03-20 12:38:38
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Critical regression: Multi cell/column selection no longer works in Data Explorer
type: bug work: frontend status: ready restricted: maintainers
## Description * Open Data Explorer, add a few columns. * Attempt to select multiple cells. * Notice that we're no longer able to do this. * This regression was introduced by https://github.com/centerofci/mathesar/pull/2213.
True
Critical regression: Multi cell/column selection no longer works in Data Explorer - ## Description * Open Data Explorer, add a few columns. * Attempt to select multiple cells. * Notice that we're no longer able to do this. * This regression was introduced by https://github.com/centerofci/mathesar/pull/2213.
main
critical regression multi cell column selection no longer works in data explorer description open data explorer add a few columns attempt to select multiple cells notice that we re no longer able to do this this regression was introduced by
1
2,648
9,081,678,442
IssuesEvent
2019-02-17 04:07:52
zendframework/zend-mvc
https://api.github.com/repos/zendframework/zend-mvc
closed
What's the purpose of dispatching requests with missing controller or action in the route match?
awaiting maintainer response question
I don't really understand the purpose of dispatching requests without a controller or an action in the route match. I think Zend\Mvc\RouteListener (line 47.) should check the existence of the controller and action parameters too. If any of them are missing, it should trigger an error (MvcEvent::EVENT_DISPATCH_ERROR). The name of the error could be "Application::ERROR_ROUTER_INVALID_MATCH" or something similar to that. It should express that we have a route match, but it cannot be used to start the dispatch process. If we do so, we can remove the default values from the getParam functions in the following files: **Zend\Mvc\DispatchListener** ``` $controllerName = $routeMatch instanceof RouteMatch ? $routeMatch->getParam('controller', 'not-found') : 'not-found'; ``` ==> ``` $controllerName = $routeMatch instanceof RouteMatch ? $routeMatch->getParam('controller') : 'not-found'; ``` **Zend\Mvc\AbstractActionController** $action = $routeMatch->getParam('action', 'not-found'); ==> `$action = $routeMatch->getParam('action')` These small things (and there are much more) make the life harder for people trying to understand the flow of MVC appliactions. We are in the middle of a dispatch event and the framework is checking whether controller or actions paramters are defined or not... It's weird. At least for me. Opinions?
True
What's the purpose of dispatching requests with missing controller or action in the route match? - I don't really understand the purpose of dispatching requests without a controller or an action in the route match. I think Zend\Mvc\RouteListener (line 47.) should check the existence of the controller and action parameters too. If any of them are missing, it should trigger an error (MvcEvent::EVENT_DISPATCH_ERROR). The name of the error could be "Application::ERROR_ROUTER_INVALID_MATCH" or something similar to that. It should express that we have a route match, but it cannot be used to start the dispatch process. If we do so, we can remove the default values from the getParam functions in the following files: **Zend\Mvc\DispatchListener** ``` $controllerName = $routeMatch instanceof RouteMatch ? $routeMatch->getParam('controller', 'not-found') : 'not-found'; ``` ==> ``` $controllerName = $routeMatch instanceof RouteMatch ? $routeMatch->getParam('controller') : 'not-found'; ``` **Zend\Mvc\AbstractActionController** $action = $routeMatch->getParam('action', 'not-found'); ==> `$action = $routeMatch->getParam('action')` These small things (and there are much more) make the life harder for people trying to understand the flow of MVC appliactions. We are in the middle of a dispatch event and the framework is checking whether controller or actions paramters are defined or not... It's weird. At least for me. Opinions?
main
what s the purpose of dispatching requests with missing controller or action in the route match i don t really understand the purpose of dispatching requests without a controller or an action in the route match i think zend mvc routelistener line should check the existence of the controller and action parameters too if any of them are missing it should trigger an error mvcevent event dispatch error the name of the error could be application error router invalid match or something similar to that it should express that we have a route match but it cannot be used to start the dispatch process if we do so we can remove the default values from the getparam functions in the following files zend mvc dispatchlistener controllername routematch instanceof routematch routematch getparam controller not found not found controllername routematch instanceof routematch routematch getparam controller not found zend mvc abstractactioncontroller action routematch getparam action not found action routematch getparam action these small things and there are much more make the life harder for people trying to understand the flow of mvc appliactions we are in the middle of a dispatch event and the framework is checking whether controller or actions paramters are defined or not it s weird at least for me opinions
1
1,071
4,890,687,163
IssuesEvent
2016-11-18 14:39:10
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
opened
HTML Entity Decoder: failing on entities with a digit at end
Maintainer Input Requested
\&sup; works but \&sup2; doesn't. Looks like they are being kicked out here - https://github.com/duckduckgo/zeroclickinfo-goodies/blob/master/lib/DDG/Goodie/HTMLEntitiesDecode.pm#L23. Not sure if it's intentional but we've received some feedback about it. ------ IA Page: http://duck.co/ia/view/htmlentities_decode [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @nishanths
True
HTML Entity Decoder: failing on entities with a digit at end - \&sup; works but \&sup2; doesn't. Looks like they are being kicked out here - https://github.com/duckduckgo/zeroclickinfo-goodies/blob/master/lib/DDG/Goodie/HTMLEntitiesDecode.pm#L23. Not sure if it's intentional but we've received some feedback about it. ------ IA Page: http://duck.co/ia/view/htmlentities_decode [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @nishanths
main
html entity decoder failing on entities with a digit at end sup works but doesn t looks like they are being kicked out here not sure if it s intentional but we ve received some feedback about it ia page nishanths
1
5,612
28,091,718,159
IssuesEvent
2023-03-30 13:27:57
oele-isis-vanderbilt/ChimeraPy
https://api.github.com/repos/oele-isis-vanderbilt/ChimeraPy
closed
Tests in CI only run in ubuntu
maintainence
The windows and Linux dependencies installation even use apt for them. It looks like the following line should be `${{ matrix.os }}` https://github.com/oele-isis-vanderbilt/ChimeraPy/blob/6df0c7d37b14f89f0ecc614e3e48e9313568dde7/.github/workflows/test.yml#L23-L28
True
Tests in CI only run in ubuntu - The windows and Linux dependencies installation even use apt for them. It looks like the following line should be `${{ matrix.os }}` https://github.com/oele-isis-vanderbilt/ChimeraPy/blob/6df0c7d37b14f89f0ecc614e3e48e9313568dde7/.github/workflows/test.yml#L23-L28
main
tests in ci only run in ubuntu the windows and linux dependencies installation even use apt for them it looks like the following line should be matrix os
1
1,066
4,889,234,103
IssuesEvent
2016-11-18 09:31:30
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
include_role : troubleshooting passing variable
affects_2.2 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME include_role ##### ANSIBLE VERSION ``` ansible 2.2.0.0 config file = /home/userdev/Documents/dlp-ansible/deploy_aqz/ansible.cfg configured module search path = ['library'] ``` (branch stable-2.2) ##### CONFIGURATION local roles and libraries location ##### OS / ENVIRONMENT Master: Ubuntu 16.04.2 Managed: Rhel 6.6 ##### SUMMARY role vars are not visible in vars attribute of include_role ##### STEPS TO REPRODUCE ``` - hosts: all gather_facts: True tasks: - include_role: name: "role_test_a" vars: a: "test" ``` role_test_a/tasks/main.yml : ``` --- - debug: var=a - debug: var=c - include_role: name: "role_test_b" vars: b: "{{ c }}" ``` role_test_a/vars/main.yml ``` --- c: "{{ a }} dummy" ``` role_test_b/tasks/main.yml ``` --- - debug: var=b ``` ##### EXPECTED RESULTS ``` PLAY [all] ********************************************************************* TASK [setup] ******************************************************************* ok: [host] TASK [role_test_a : debug] ***************************************************** ok: [host] => { "a": "test" } TASK [role_test_a : debug] ***************************************************** ok: [host] => { "c": "test dummy" } TASK [role_test_b : debug] ***************************************************** ok: [host] => { "b": "test dummy" } PLAY RECAP ********************************************************************* host : ok=4 changed=0 unreachable=0 failed=0 ``` ##### ACTUAL RESULTS ``` PLAY [all] ********************************************************************* TASK [setup] ******************************************************************* ok: [host] TASK [role_test_a : debug] ***************************************************** ok: [host] => { "a": "test" } TASK [role_test_a : debug] ***************************************************** ok: [host] => { "c": "test dummy" } TASK [role_test_b : debug] ***************************************************** ok: [host] => { "b": "VARIABLE IS NOT DEFINED!" } PLAY RECAP ********************************************************************* host : ok=4 changed=0 unreachable=0 failed=0 ```
True
include_role : troubleshooting passing variable - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME include_role ##### ANSIBLE VERSION ``` ansible 2.2.0.0 config file = /home/userdev/Documents/dlp-ansible/deploy_aqz/ansible.cfg configured module search path = ['library'] ``` (branch stable-2.2) ##### CONFIGURATION local roles and libraries location ##### OS / ENVIRONMENT Master: Ubuntu 16.04.2 Managed: Rhel 6.6 ##### SUMMARY role vars are not visible in vars attribute of include_role ##### STEPS TO REPRODUCE ``` - hosts: all gather_facts: True tasks: - include_role: name: "role_test_a" vars: a: "test" ``` role_test_a/tasks/main.yml : ``` --- - debug: var=a - debug: var=c - include_role: name: "role_test_b" vars: b: "{{ c }}" ``` role_test_a/vars/main.yml ``` --- c: "{{ a }} dummy" ``` role_test_b/tasks/main.yml ``` --- - debug: var=b ``` ##### EXPECTED RESULTS ``` PLAY [all] ********************************************************************* TASK [setup] ******************************************************************* ok: [host] TASK [role_test_a : debug] ***************************************************** ok: [host] => { "a": "test" } TASK [role_test_a : debug] ***************************************************** ok: [host] => { "c": "test dummy" } TASK [role_test_b : debug] ***************************************************** ok: [host] => { "b": "test dummy" } PLAY RECAP ********************************************************************* host : ok=4 changed=0 unreachable=0 failed=0 ``` ##### ACTUAL RESULTS ``` PLAY [all] ********************************************************************* TASK [setup] ******************************************************************* ok: [host] TASK [role_test_a : debug] ***************************************************** ok: [host] => { "a": "test" } TASK [role_test_a : debug] ***************************************************** ok: [host] => { "c": "test dummy" } TASK [role_test_b : debug] ***************************************************** ok: [host] => { "b": "VARIABLE IS NOT DEFINED!" } PLAY RECAP ********************************************************************* host : ok=4 changed=0 unreachable=0 failed=0 ```
main
include role troubleshooting passing variable issue type bug report component name include role ansible version ansible config file home userdev documents dlp ansible deploy aqz ansible cfg configured module search path branch stable configuration local roles and libraries location os environment master ubuntu managed rhel summary role vars are not visible in vars attribute of include role steps to reproduce hosts all gather facts true tasks include role name role test a vars a test role test a tasks main yml debug var a debug var c include role name role test b vars b c role test a vars main yml c a dummy role test b tasks main yml debug var b expected results play task ok task ok a test task ok c test dummy task ok b test dummy play recap host ok changed unreachable failed actual results play task ok task ok a test task ok c test dummy task ok b variable is not defined play recap host ok changed unreachable failed
1
298,068
25,784,493,780
IssuesEvent
2022-12-09 18:59:17
deislabs/spiderlightning
https://api.github.com/repos/deislabs/spiderlightning
opened
run redis implementation test on Windows
πŸ§ͺ tests
**Describe the solution you'd like** I've setup the pipelines to install Redis on Windows. We should be setup to to run the unit tests there. **Additional context** n/a
1.0
run redis implementation test on Windows - **Describe the solution you'd like** I've setup the pipelines to install Redis on Windows. We should be setup to to run the unit tests there. **Additional context** n/a
non_main
run redis implementation test on windows describe the solution you d like i ve setup the pipelines to install redis on windows we should be setup to to run the unit tests there additional context n a
0
2,514
8,655,460,059
IssuesEvent
2018-11-27 16:00:32
codestation/qcma
https://api.github.com/repos/codestation/qcma
closed
QCMA stop connecting over WiFi with time
unmaintained
Ubuntu 16.04 QCMA 0.4.1 Ubuntu linked to N56U router w/Padavan (OpenWRT) by cable, the same router serves wifi to VITA Starts with the system and work flawlessly several hours. Then stops connecting at all - PSVITA gives "PC is not connected" error What can be done about this?
True
QCMA stop connecting over WiFi with time - Ubuntu 16.04 QCMA 0.4.1 Ubuntu linked to N56U router w/Padavan (OpenWRT) by cable, the same router serves wifi to VITA Starts with the system and work flawlessly several hours. Then stops connecting at all - PSVITA gives "PC is not connected" error What can be done about this?
main
qcma stop connecting over wifi with time ubuntu qcma ubuntu linked to router w padavan openwrt by cable the same router serves wifi to vita starts with the system and work flawlessly several hours then stops connecting at all psvita gives pc is not connected error what can be done about this
1
4,457
23,211,519,746
IssuesEvent
2022-08-02 10:33:04
mozilla/foundation.mozilla.org
https://api.github.com/repos/mozilla/foundation.mozilla.org
closed
Research Google Analytics 4
engineering Maintain
We will need to update to [Google Analytics 4](https://blog.google/products/marketingplatform/analytics/prepare-for-future-with-google-analytics-4/) however I have no idea how much or little effort this will be. - [ ] Get confirmation from from Jackie that it is ok to upgrade (I have pinged Jackie to see if it needs to go through privacy and legal review) - [ ] Look into how it is done and what the effects are - [ ] Open tickets to do the work - [ ] Inform staff that we will be doing this cc @bheasman in case you have more context!
True
Research Google Analytics 4 - We will need to update to [Google Analytics 4](https://blog.google/products/marketingplatform/analytics/prepare-for-future-with-google-analytics-4/) however I have no idea how much or little effort this will be. - [ ] Get confirmation from from Jackie that it is ok to upgrade (I have pinged Jackie to see if it needs to go through privacy and legal review) - [ ] Look into how it is done and what the effects are - [ ] Open tickets to do the work - [ ] Inform staff that we will be doing this cc @bheasman in case you have more context!
main
research google analytics we will need to update to however i have no idea how much or little effort this will be get confirmation from from jackie that it is ok to upgrade i have pinged jackie to see if it needs to go through privacy and legal review look into how it is done and what the effects are open tickets to do the work inform staff that we will be doing this cc bheasman in case you have more context
1
631
4,147,294,130
IssuesEvent
2016-06-15 06:10:55
Particular/ServiceControl
https://api.github.com/repos/Particular/ServiceControl
closed
Support "NT AUTHORITY\LocalService" in installer
Tag: Installer Tag: Maintainer Prio Type: Improvement
I almost always install SC using the *NT AUTHORITY\LocalService* account and have to search for the exact string to paste into the box. it would help if this account could be added in the list: ![image](https://cloud.githubusercontent.com/assets/152998/14665735/908021d2-06d5-11e6-9700-aad5f0c6d1ae.png) This makes it possible to install SC for SQL with integrated security.
True
Support "NT AUTHORITY\LocalService" in installer - I almost always install SC using the *NT AUTHORITY\LocalService* account and have to search for the exact string to paste into the box. it would help if this account could be added in the list: ![image](https://cloud.githubusercontent.com/assets/152998/14665735/908021d2-06d5-11e6-9700-aad5f0c6d1ae.png) This makes it possible to install SC for SQL with integrated security.
main
support nt authority localservice in installer i almost always install sc using the nt authority localservice account and have to search for the exact string to paste into the box it would help if this account could be added in the list this makes it possible to install sc for sql with integrated security
1
2,288
8,142,152,473
IssuesEvent
2018-08-21 06:26:20
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Crayons color the full tile then just draw the drawing
Maintainability/Hinders improvements
It really fucks up my drawing if i write a black D and the whole background is an D we had paint for tile colouring crayon shouldnt do this
True
Crayons color the full tile then just draw the drawing - It really fucks up my drawing if i write a black D and the whole background is an D we had paint for tile colouring crayon shouldnt do this
main
crayons color the full tile then just draw the drawing it really fucks up my drawing if i write a black d and the whole background is an d we had paint for tile colouring crayon shouldnt do this
1
4,015
18,750,147,654
IssuesEvent
2021-11-05 00:06:06
portage-brew/portage-brew-staging-and-evolution
https://api.github.com/repos/portage-brew/portage-brew-staging-and-evolution
closed
Bikeshed the Name
enhancement help wanted Needs Discussion Needs Maintainer Feedback
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;Placing an assumption that whatever keys off of what's already here ends up being a Homebrew fork is, admittedly, rather presumptuous of me. (I can move it to a personal repository and just let this repository serve as a staging ground for something that only takes _some_ points from it if desired.) Names I've considered, in any case, include: - 'Sourcebrew' (This was already taken as an organization/user name.) - 'Brew-From-Source' and variations on that with different capitalization and hyphenation conventions (Also taken, I think?) - 'ports-brew' (_Might_ also be taken, IIRC…?) Additional suggestions are welcome.
True
Bikeshed the Name - &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;Placing an assumption that whatever keys off of what's already here ends up being a Homebrew fork is, admittedly, rather presumptuous of me. (I can move it to a personal repository and just let this repository serve as a staging ground for something that only takes _some_ points from it if desired.) Names I've considered, in any case, include: - 'Sourcebrew' (This was already taken as an organization/user name.) - 'Brew-From-Source' and variations on that with different capitalization and hyphenation conventions (Also taken, I think?) - 'ports-brew' (_Might_ also be taken, IIRC…?) Additional suggestions are welcome.
main
bikeshed the name nbsp nbsp nbsp nbsp nbsp placing an assumption that whatever keys off of what s already here ends up being a homebrew fork is admittedly rather presumptuous of me i can move it to a personal repository and just let this repository serve as a staging ground for something that only takes some points from it if desired names i ve considered in any case include sourcebrew this was already taken as an organization user name brew from source and variations on that with different capitalization and hyphenation conventions also taken i think ports brew might also be taken iirc… additional suggestions are welcome
1
806,148
29,803,017,788
IssuesEvent
2023-06-16 09:31:21
smol-dot/smoldot
https://api.github.com/repos/smol-dot/smoldot
opened
No longer use the `log` crate in `light-base`
priority-low
Instead, the `PlatformRef` trait should get a `log` function.
1.0
No longer use the `log` crate in `light-base` - Instead, the `PlatformRef` trait should get a `log` function.
non_main
no longer use the log crate in light base instead the platformref trait should get a log function
0
1,331
5,714,470,665
IssuesEvent
2017-04-19 10:29:04
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
Cannot uninstall adobe-air
awaiting maintainer feedback
#### General troubleshooting steps - [X] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue. - [X] None of the templates was appropriate for my issue, or I’m not sure. - [X] I ran `brew update-reset && brew update` and retried my command. - [X] I ran `brew doctor`, fixed as many issues as possible and retried my command. - [X] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md). #### Description of issue Uninstalling adobe-air exits with an error #### Output of your command with `--verbose --debug` <details> <summary>$ brew cask remove adobe-air --verbose --debug</summary> ```shell ==> Uninstalling Cask adobe-air ==> Hbc::Installer#uninstall ==> Un-installing artifacts ==> Determining which artifacts are present in Cask adobe-air ==> 3 artifact/s defined #<Hbc::Artifact::Uninstall:0x007feacc09fa10> #<Hbc::Artifact::Installer:0x007feacc09fc68> #<Hbc::Artifact::Zap:0x007feacc09f8a8> ==> Un-installing artifact of class Hbc::Artifact::Uninstall ==> Running uninstall process for adobe-air; your password may be necessary ==> Running uninstall script Adobe AIR Installer.app/Contents/MacOS/Adobe AIR Installer ==> Executing: ["/bin/chmod", "--", "+x", "/usr/local/Caskroom/adobe-air/23.0/Adobe AIR Installer.app/Contents/MacOS/Adobe AIR Installer"] ==> Executing: ["/usr/local/Caskroom/adobe-air/23.0/Adobe AIR Installer.app/Contents/MacOS/Adobe AIR Installer", "-uninstall"] ==> Uninstalling Adobe AIR (all versions) ==> failed (consult log) Error: Command failed to execute! ==> Failed command: #<Pathname:/usr/local/Caskroom/adobe-air/23.0/Adobe AIR Installer.app/Contents/MacOS/Adobe AIR Installer> -uninstall ==> Standard Output of failed command: Uninstalling Adobe AIR (all versions) failed (consult log) ==> Standard Error of failed command: ==> Exit status of failed command: #<Process::Status: pid 93676 exit 7> /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/system_command.rb:65:in `assert_success' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/system_command.rb:36:in `run!' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/system_command.rb:14:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall_base.rb:182:in `uninstall_script' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall_base.rb:34:in `block (2 levels) in dispatch_uninstall_directives' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall_base.rb:32:in `each' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall_base.rb:32:in `block in dispatch_uninstall_directives' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall_base.rb:31:in `each' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall_base.rb:31:in `dispatch_uninstall_directives' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall.rb:7:in `uninstall_phase' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/installer.rb:330:in `block in uninstall_artifacts' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/installer.rb:327:in `each' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/installer.rb:327:in `uninstall_artifacts' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/installer.rb:312:in `uninstall' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/uninstall.rb:26:in `block in run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/uninstall.rb:9:in `each' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/uninstall.rb:9:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:115:in `run_command' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:158:in `process' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:91:in `<main>' Error: Kernel.exit /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:163:in `exit' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:163:in `rescue in process' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:149:in `process' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:91:in `<main>' ``` </details> #### Output of `brew cask doctor` <details> <summary>$ brew cask doctor</summary> ```shell ==> Homebrew-Cask Version Homebrew-Cask 1.1.13-19-g55c02ae774 caskroom/homebrew-cask (git revision c24a6; last commit 2017-04-18) ==> Homebrew-Cask Install Location <NONE> ==> Homebrew-Cask Staging Location /usr/local/Caskroom ==> Homebrew-Cask Cached Downloads ~/Library/Caches/Homebrew/Cask (2 files, 47.2MB) ==> Homebrew-Cask Taps: /usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3632 casks) /usr/local/Homebrew/Library/Taps/argon/homebrew-mas (0 casks) /usr/local/Homebrew/Library/Taps/blackfireio/homebrew-blackfire (0 casks) /usr/local/Homebrew/Library/Taps/dart-lang/homebrew-dart (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-binary (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-bundle (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-core (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-dupes (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-php (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-python (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-science (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-services (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-x11 (0 casks) /usr/local/Homebrew/Library/Taps/ravenac95/homebrew-sudolikeaboss (0 casks) /usr/local/Homebrew/Library/Taps/thoughtbot/homebrew-formulae (0 casks) /usr/local/Homebrew/Library/Taps/tobli/homebrew-browsertime (0 casks) ==> Contents of $LOAD_PATH /usr/local/Homebrew/Library/Homebrew/cask/lib /usr/local/Homebrew/Library/Homebrew /Library/Ruby/Gems/2.0.0/gems/json-2.0.3/lib /Library/Ruby/Gems/2.0.0/extensions/universal-darwin-16/2.0.0/json-2.0.3 /Library/Ruby/Gems/2.0.0/gems/psych-2.2.4/lib /Library/Ruby/Gems/2.0.0/extensions/universal-darwin-16/2.0.0/psych-2.2.4 /Library/Ruby/Site/2.0.0 /Library/Ruby/Site/2.0.0/x86_64-darwin16 /Library/Ruby/Site/2.0.0/universal-darwin16 /Library/Ruby/Site /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16 ==> Environment Variables LANG="fr_FR.UTF-8" PATH="/usr/local/bin:/usr/local/sbin:/usr/local/share/npm/bin:/usr/local/opt/go/libexec/bin:~/.composer/vendor/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/MacGPG2/bin:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-bundle/cmd:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-services/cmd:/usr/local/Homebrew/Library/Homebrew/shims/scm" SHELL="/bin/zsh" ``` </details> <details> <summary>$ brew doctor</summary> ```shell Please note that these warnings are just used to help the Homebrew maintainers with debugging if you file an issue. If everything you use Homebrew for is working fine: please don't worry and just ignore them. Thanks! Warning: Some keg-only formula are linked into the Cellar. Linking a keg-only formula, such as gettext, into the cellar with `brew link <formula>` will cause other formulae to detect them during the `./configure` step. This may cause problems when compiling those other formulae. Binaries provided by keg-only formulae may override system binaries with other strange results. You may wish to `brew unlink` these brews: curl ``` </details>
True
Cannot uninstall adobe-air - #### General troubleshooting steps - [X] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue. - [X] None of the templates was appropriate for my issue, or I’m not sure. - [X] I ran `brew update-reset && brew update` and retried my command. - [X] I ran `brew doctor`, fixed as many issues as possible and retried my command. - [X] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md). #### Description of issue Uninstalling adobe-air exits with an error #### Output of your command with `--verbose --debug` <details> <summary>$ brew cask remove adobe-air --verbose --debug</summary> ```shell ==> Uninstalling Cask adobe-air ==> Hbc::Installer#uninstall ==> Un-installing artifacts ==> Determining which artifacts are present in Cask adobe-air ==> 3 artifact/s defined #<Hbc::Artifact::Uninstall:0x007feacc09fa10> #<Hbc::Artifact::Installer:0x007feacc09fc68> #<Hbc::Artifact::Zap:0x007feacc09f8a8> ==> Un-installing artifact of class Hbc::Artifact::Uninstall ==> Running uninstall process for adobe-air; your password may be necessary ==> Running uninstall script Adobe AIR Installer.app/Contents/MacOS/Adobe AIR Installer ==> Executing: ["/bin/chmod", "--", "+x", "/usr/local/Caskroom/adobe-air/23.0/Adobe AIR Installer.app/Contents/MacOS/Adobe AIR Installer"] ==> Executing: ["/usr/local/Caskroom/adobe-air/23.0/Adobe AIR Installer.app/Contents/MacOS/Adobe AIR Installer", "-uninstall"] ==> Uninstalling Adobe AIR (all versions) ==> failed (consult log) Error: Command failed to execute! ==> Failed command: #<Pathname:/usr/local/Caskroom/adobe-air/23.0/Adobe AIR Installer.app/Contents/MacOS/Adobe AIR Installer> -uninstall ==> Standard Output of failed command: Uninstalling Adobe AIR (all versions) failed (consult log) ==> Standard Error of failed command: ==> Exit status of failed command: #<Process::Status: pid 93676 exit 7> /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/system_command.rb:65:in `assert_success' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/system_command.rb:36:in `run!' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/system_command.rb:14:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall_base.rb:182:in `uninstall_script' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall_base.rb:34:in `block (2 levels) in dispatch_uninstall_directives' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall_base.rb:32:in `each' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall_base.rb:32:in `block in dispatch_uninstall_directives' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall_base.rb:31:in `each' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall_base.rb:31:in `dispatch_uninstall_directives' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/artifact/uninstall.rb:7:in `uninstall_phase' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/installer.rb:330:in `block in uninstall_artifacts' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/installer.rb:327:in `each' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/installer.rb:327:in `uninstall_artifacts' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/installer.rb:312:in `uninstall' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/uninstall.rb:26:in `block in run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/uninstall.rb:9:in `each' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/uninstall.rb:9:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:115:in `run_command' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:158:in `process' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:91:in `<main>' Error: Kernel.exit /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:163:in `exit' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:163:in `rescue in process' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:149:in `process' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:91:in `<main>' ``` </details> #### Output of `brew cask doctor` <details> <summary>$ brew cask doctor</summary> ```shell ==> Homebrew-Cask Version Homebrew-Cask 1.1.13-19-g55c02ae774 caskroom/homebrew-cask (git revision c24a6; last commit 2017-04-18) ==> Homebrew-Cask Install Location <NONE> ==> Homebrew-Cask Staging Location /usr/local/Caskroom ==> Homebrew-Cask Cached Downloads ~/Library/Caches/Homebrew/Cask (2 files, 47.2MB) ==> Homebrew-Cask Taps: /usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3632 casks) /usr/local/Homebrew/Library/Taps/argon/homebrew-mas (0 casks) /usr/local/Homebrew/Library/Taps/blackfireio/homebrew-blackfire (0 casks) /usr/local/Homebrew/Library/Taps/dart-lang/homebrew-dart (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-binary (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-bundle (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-core (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-dupes (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-php (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-python (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-science (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-services (0 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-x11 (0 casks) /usr/local/Homebrew/Library/Taps/ravenac95/homebrew-sudolikeaboss (0 casks) /usr/local/Homebrew/Library/Taps/thoughtbot/homebrew-formulae (0 casks) /usr/local/Homebrew/Library/Taps/tobli/homebrew-browsertime (0 casks) ==> Contents of $LOAD_PATH /usr/local/Homebrew/Library/Homebrew/cask/lib /usr/local/Homebrew/Library/Homebrew /Library/Ruby/Gems/2.0.0/gems/json-2.0.3/lib /Library/Ruby/Gems/2.0.0/extensions/universal-darwin-16/2.0.0/json-2.0.3 /Library/Ruby/Gems/2.0.0/gems/psych-2.2.4/lib /Library/Ruby/Gems/2.0.0/extensions/universal-darwin-16/2.0.0/psych-2.2.4 /Library/Ruby/Site/2.0.0 /Library/Ruby/Site/2.0.0/x86_64-darwin16 /Library/Ruby/Site/2.0.0/universal-darwin16 /Library/Ruby/Site /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16 ==> Environment Variables LANG="fr_FR.UTF-8" PATH="/usr/local/bin:/usr/local/sbin:/usr/local/share/npm/bin:/usr/local/opt/go/libexec/bin:~/.composer/vendor/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/MacGPG2/bin:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-bundle/cmd:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-services/cmd:/usr/local/Homebrew/Library/Homebrew/shims/scm" SHELL="/bin/zsh" ``` </details> <details> <summary>$ brew doctor</summary> ```shell Please note that these warnings are just used to help the Homebrew maintainers with debugging if you file an issue. If everything you use Homebrew for is working fine: please don't worry and just ignore them. Thanks! Warning: Some keg-only formula are linked into the Cellar. Linking a keg-only formula, such as gettext, into the cellar with `brew link <formula>` will cause other formulae to detect them during the `./configure` step. This may cause problems when compiling those other formulae. Binaries provided by keg-only formulae may override system binaries with other strange results. You may wish to `brew unlink` these brews: curl ``` </details>
main
cannot uninstall adobe air general troubleshooting steps i have checked the instructions for or before opening the issue none of the templates was appropriate for my issue or i’m not sure i ran brew update reset brew update and retried my command i ran brew doctor fixed as many issues as possible and retried my command i understand that description of issue uninstalling adobe air exits with an error output of your command with verbose debug brew cask remove adobe air verbose debug shell uninstalling cask adobe air hbc installer uninstall un installing artifacts determining which artifacts are present in cask adobe air artifact s defined un installing artifact of class hbc artifact uninstall running uninstall process for adobe air your password may be necessary running uninstall script adobe air installer app contents macos adobe air installer executing executing uninstalling adobe air all versions failed consult log error command failed to execute failed command uninstall standard output of failed command uninstalling adobe air all versions failed consult log standard error of failed command exit status of failed command usr local homebrew library homebrew cask lib hbc system command rb in assert success usr local homebrew library homebrew cask lib hbc system command rb in run usr local homebrew library homebrew cask lib hbc system command rb in run usr local homebrew library homebrew cask lib hbc artifact uninstall base rb in uninstall script usr local homebrew library homebrew cask lib hbc artifact uninstall base rb in block levels in dispatch uninstall directives usr local homebrew library homebrew cask lib hbc artifact uninstall base rb in each usr local homebrew library homebrew cask lib hbc artifact uninstall base rb in block in dispatch uninstall directives usr local homebrew library homebrew cask lib hbc artifact uninstall base rb in each usr local homebrew library homebrew cask lib hbc artifact uninstall base rb in dispatch uninstall directives usr local homebrew library homebrew cask lib hbc artifact uninstall rb in uninstall phase usr local homebrew library homebrew cask lib hbc installer rb in block in uninstall artifacts usr local homebrew library homebrew cask lib hbc installer rb in each usr local homebrew library homebrew cask lib hbc installer rb in uninstall artifacts usr local homebrew library homebrew cask lib hbc installer rb in uninstall usr local homebrew library homebrew cask lib hbc cli uninstall rb in block in run usr local homebrew library homebrew cask lib hbc cli uninstall rb in each usr local homebrew library homebrew cask lib hbc cli uninstall rb in run usr local homebrew library homebrew cask lib hbc cli rb in run command usr local homebrew library homebrew cask lib hbc cli rb in process usr local homebrew library homebrew cmd cask rb in cask usr local homebrew library homebrew brew rb in error kernel exit usr local homebrew library homebrew cask lib hbc cli rb in exit usr local homebrew library homebrew cask lib hbc cli rb in rescue in process usr local homebrew library homebrew cask lib hbc cli rb in process usr local homebrew library homebrew cmd cask rb in cask usr local homebrew library homebrew brew rb in output of brew cask doctor brew cask doctor shell homebrew cask version homebrew cask caskroom homebrew cask git revision last commit homebrew cask install location homebrew cask staging location usr local caskroom homebrew cask cached downloads library caches homebrew cask files homebrew cask taps usr local homebrew library taps caskroom homebrew cask casks usr local homebrew library taps argon homebrew mas casks usr local homebrew library taps blackfireio homebrew blackfire casks usr local homebrew library taps dart lang homebrew dart casks usr local homebrew library taps homebrew homebrew binary casks usr local homebrew library taps homebrew homebrew bundle casks usr local homebrew library taps homebrew homebrew core casks usr local homebrew library taps homebrew homebrew dupes casks usr local homebrew library taps homebrew homebrew php casks usr local homebrew library taps homebrew homebrew python casks usr local homebrew library taps homebrew homebrew science casks usr local homebrew library taps homebrew homebrew services casks usr local homebrew library taps homebrew homebrew casks usr local homebrew library taps homebrew sudolikeaboss casks usr local homebrew library taps thoughtbot homebrew formulae casks usr local homebrew library taps tobli homebrew browsertime casks contents of load path usr local homebrew library homebrew cask lib usr local homebrew library homebrew library ruby gems gems json lib library ruby gems extensions universal darwin json library ruby gems gems psych lib library ruby gems extensions universal darwin psych library ruby site library ruby site library ruby site universal library ruby site system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby universal system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby universal environment variables lang fr fr utf path usr local bin usr local sbin usr local share npm bin usr local opt go libexec bin composer vendor bin usr bin bin usr sbin sbin opt bin usr local bin usr local homebrew library taps homebrew homebrew bundle cmd usr local homebrew library taps homebrew homebrew services cmd usr local homebrew library homebrew shims scm shell bin zsh brew doctor shell please note that these warnings are just used to help the homebrew maintainers with debugging if you file an issue if everything you use homebrew for is working fine please don t worry and just ignore them thanks warning some keg only formula are linked into the cellar linking a keg only formula such as gettext into the cellar with brew link will cause other formulae to detect them during the configure step this may cause problems when compiling those other formulae binaries provided by keg only formulae may override system binaries with other strange results you may wish to brew unlink these brews curl
1
777,868
27,296,263,031
IssuesEvent
2023-02-23 20:38:10
NuGet/Home
https://api.github.com/repos/NuGet/Home
closed
WebSite projects opened from IIS fail to install packages
Priority:1 Product:VS.Client Type:Bug Functionality:Install Style:Packages.Config RegressionFromPreviousRTM
### NuGet Product Used Visual Studio Package Management UI, Visual Studio Package Manager Console ### Product Version VS 17.4 ### Worked before? 17.3 ### Impact I'm unable to use this version ### Repro Steps & Context 1. In IIS Manager, create a virtual directory as an application. 2. In VS, use "open website", select "local IIS", and then select the directory 3. Try to install a package that has `lib/<tfm>/*.dll` files. An obscure error message is reported: > Failed to add reference to β€˜Newtonsoft.Json’. > > Invalid URI: The Authority/Host could not be parsed. ### Verbose Logs _No response_
1.0
WebSite projects opened from IIS fail to install packages - ### NuGet Product Used Visual Studio Package Management UI, Visual Studio Package Manager Console ### Product Version VS 17.4 ### Worked before? 17.3 ### Impact I'm unable to use this version ### Repro Steps & Context 1. In IIS Manager, create a virtual directory as an application. 2. In VS, use "open website", select "local IIS", and then select the directory 3. Try to install a package that has `lib/<tfm>/*.dll` files. An obscure error message is reported: > Failed to add reference to β€˜Newtonsoft.Json’. > > Invalid URI: The Authority/Host could not be parsed. ### Verbose Logs _No response_
non_main
website projects opened from iis fail to install packages nuget product used visual studio package management ui visual studio package manager console product version vs worked before impact i m unable to use this version repro steps context in iis manager create a virtual directory as an application in vs use open website select local iis and then select the directory try to install a package that has lib dll files an obscure error message is reported failed to add reference to β€˜newtonsoft json’ invalid uri the authority host could not be parsed verbose logs no response
0
7,016
10,166,810,409
IssuesEvent
2019-08-07 16:39:56
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
closed
Bigtable: 'test_bigtable_create_table' snippet flakes with '504 Deadline Exceeded'.
api: bigtable flaky testing type: process
From [this Kokoro failure](https://source.cloud.google.com/results/invocations/2322354e-a3c8-4e30-8cac-2dbf3a814cad/targets/cloud-devrel%2Fclient-libraries%2Fgoogle-cloud-python%2Fpresubmit%2Fbigtable/log): ```python __________________________ test_bigtable_create_table __________________________ args = (parent: "projects/precise-truck-742/instances/snippet-tests-1561572579521" table_id: "table_my" table { column_families { key: "cf1" value { gc_rule { max_num_versions: 2 } } } } ,) kwargs = {'metadata': [('x-goog-request-params', 'parent=projects/precise-truck-742/instances/snippet-tests-1561572579521'), ('x-goog-api-client', 'gl-python/3.7.0b3 grpc/1.21.1 gax/1.13.0 gapic/0.33.0 gccl/0.33.0')], 'timeout': 20.0} @six.wraps(callable_) def error_remapped_callable(*args, **kwargs): try: > return callable_(*args, **kwargs) ../api_core/google/api_core/grpc_helpers.py:57: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7f9df35a5828> request = parent: "projects/precise-truck-742/instances/snippet-tests-1561572579521" table_id: "table_my" table { column_families { key: "cf1" value { gc_rule { max_num_versions: 2 } } } } timeout = 20.0 metadata = [('x-goog-request-params', 'parent=projects/precise-truck-742/instances/snippet-tests-1561572579521'), ('x-goog-api-client', 'gl-python/3.7.0b3 grpc/1.21.1 gax/1.13.0 gapic/0.33.0 gccl/0.33.0')] credentials = None, wait_for_ready = None, compression = None def __call__(self, request, timeout=None, metadata=None, credentials=None, wait_for_ready=None, compression=None): state, call, = self._blocking(request, timeout, metadata, credentials, wait_for_ready, compression) > return _end_unary_response_blocking(state, call, False, None) .nox/snippets-3-7/lib/python3.7/site-packages/grpc/_channel.py:565: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ state = <grpc._channel._RPCState object at 0x7f9df35c54e0> call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f9df35b1e88> with_call = False, deadline = None def _end_unary_response_blocking(state, call, with_call, deadline): if state.code is grpc.StatusCode.OK: if with_call: rendezvous = _Rendezvous(state, call, None, deadline) return state.response, rendezvous else: return state.response else: > raise _Rendezvous(state, None, None, deadline) E grpc._channel._Rendezvous: <_Rendezvous of RPC that terminated with: E status = StatusCode.DEADLINE_EXCEEDED E details = "Deadline Exceeded" E debug_error_string = "{"created":"@1561572698.043685871","description":"Error received from peer ipv4:74.125.142.95:443","file":"src/core/lib/surface/call.cc","file_line":1046,"grpc_message":"Deadline Exceeded","grpc_status":4}" E > .nox/snippets-3-7/lib/python3.7/site-packages/grpc/_channel.py:467: _Rendezvous The above exception was the direct cause of the following exception: def test_bigtable_create_table(): # [START bigtable_create_table] from google.cloud.bigtable import Client from google.cloud.bigtable import column_family client = Client(admin=True) instance = client.instance(INSTANCE_ID) table = instance.table("table_my") # Define the GC policy to retain only the most recent 2 versions. max_versions_rule = column_family.MaxVersionsGCRule(2) > table.create(column_families={"cf1": max_versions_rule}) docs/snippets.py:341: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/bigtable/table.py:251: in create initial_splits=splits, google/cloud/bigtable_admin_v2/gapic/bigtable_table_admin_client.py:340: in create_table request, retry=retry, timeout=timeout, metadata=metadata ../api_core/google/api_core/gapic_v1/method.py:143: in __call__ return wrapped_func(*args, **kwargs) ../api_core/google/api_core/retry.py:273: in retry_wrapped_func on_error=on_error, ../api_core/google/api_core/retry.py:182: in retry_target return target() ../api_core/google/api_core/timeout.py:214: in func_with_timeout return func(*args, **kwargs) ../api_core/google/api_core/grpc_helpers.py:59: in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None from_value = <_Rendezvous of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug...2.95:443","file":"src/core/lib/surface/call.cc","file_line":1046,"grpc_message":"Deadline Exceeded","grpc_status":4}" > > ??? E google.api_core.exceptions.DeadlineExceeded: 504 Deadline Exceeded ``
1.0
Bigtable: 'test_bigtable_create_table' snippet flakes with '504 Deadline Exceeded'. - From [this Kokoro failure](https://source.cloud.google.com/results/invocations/2322354e-a3c8-4e30-8cac-2dbf3a814cad/targets/cloud-devrel%2Fclient-libraries%2Fgoogle-cloud-python%2Fpresubmit%2Fbigtable/log): ```python __________________________ test_bigtable_create_table __________________________ args = (parent: "projects/precise-truck-742/instances/snippet-tests-1561572579521" table_id: "table_my" table { column_families { key: "cf1" value { gc_rule { max_num_versions: 2 } } } } ,) kwargs = {'metadata': [('x-goog-request-params', 'parent=projects/precise-truck-742/instances/snippet-tests-1561572579521'), ('x-goog-api-client', 'gl-python/3.7.0b3 grpc/1.21.1 gax/1.13.0 gapic/0.33.0 gccl/0.33.0')], 'timeout': 20.0} @six.wraps(callable_) def error_remapped_callable(*args, **kwargs): try: > return callable_(*args, **kwargs) ../api_core/google/api_core/grpc_helpers.py:57: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7f9df35a5828> request = parent: "projects/precise-truck-742/instances/snippet-tests-1561572579521" table_id: "table_my" table { column_families { key: "cf1" value { gc_rule { max_num_versions: 2 } } } } timeout = 20.0 metadata = [('x-goog-request-params', 'parent=projects/precise-truck-742/instances/snippet-tests-1561572579521'), ('x-goog-api-client', 'gl-python/3.7.0b3 grpc/1.21.1 gax/1.13.0 gapic/0.33.0 gccl/0.33.0')] credentials = None, wait_for_ready = None, compression = None def __call__(self, request, timeout=None, metadata=None, credentials=None, wait_for_ready=None, compression=None): state, call, = self._blocking(request, timeout, metadata, credentials, wait_for_ready, compression) > return _end_unary_response_blocking(state, call, False, None) .nox/snippets-3-7/lib/python3.7/site-packages/grpc/_channel.py:565: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ state = <grpc._channel._RPCState object at 0x7f9df35c54e0> call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f9df35b1e88> with_call = False, deadline = None def _end_unary_response_blocking(state, call, with_call, deadline): if state.code is grpc.StatusCode.OK: if with_call: rendezvous = _Rendezvous(state, call, None, deadline) return state.response, rendezvous else: return state.response else: > raise _Rendezvous(state, None, None, deadline) E grpc._channel._Rendezvous: <_Rendezvous of RPC that terminated with: E status = StatusCode.DEADLINE_EXCEEDED E details = "Deadline Exceeded" E debug_error_string = "{"created":"@1561572698.043685871","description":"Error received from peer ipv4:74.125.142.95:443","file":"src/core/lib/surface/call.cc","file_line":1046,"grpc_message":"Deadline Exceeded","grpc_status":4}" E > .nox/snippets-3-7/lib/python3.7/site-packages/grpc/_channel.py:467: _Rendezvous The above exception was the direct cause of the following exception: def test_bigtable_create_table(): # [START bigtable_create_table] from google.cloud.bigtable import Client from google.cloud.bigtable import column_family client = Client(admin=True) instance = client.instance(INSTANCE_ID) table = instance.table("table_my") # Define the GC policy to retain only the most recent 2 versions. max_versions_rule = column_family.MaxVersionsGCRule(2) > table.create(column_families={"cf1": max_versions_rule}) docs/snippets.py:341: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/bigtable/table.py:251: in create initial_splits=splits, google/cloud/bigtable_admin_v2/gapic/bigtable_table_admin_client.py:340: in create_table request, retry=retry, timeout=timeout, metadata=metadata ../api_core/google/api_core/gapic_v1/method.py:143: in __call__ return wrapped_func(*args, **kwargs) ../api_core/google/api_core/retry.py:273: in retry_wrapped_func on_error=on_error, ../api_core/google/api_core/retry.py:182: in retry_target return target() ../api_core/google/api_core/timeout.py:214: in func_with_timeout return func(*args, **kwargs) ../api_core/google/api_core/grpc_helpers.py:59: in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None from_value = <_Rendezvous of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug...2.95:443","file":"src/core/lib/surface/call.cc","file_line":1046,"grpc_message":"Deadline Exceeded","grpc_status":4}" > > ??? E google.api_core.exceptions.DeadlineExceeded: 504 Deadline Exceeded ``
non_main
bigtable test bigtable create table snippet flakes with deadline exceeded from python test bigtable create table args parent projects precise truck instances snippet tests table id table my table column families key value gc rule max num versions kwargs metadata timeout six wraps callable def error remapped callable args kwargs try return callable args kwargs api core google api core grpc helpers py self request parent projects precise truck instances snippet tests table id table my table column families key value gc rule max num versions timeout metadata credentials none wait for ready none compression none def call self request timeout none metadata none credentials none wait for ready none compression none state call self blocking request timeout metadata credentials wait for ready compression return end unary response blocking state call false none nox snippets lib site packages grpc channel py state call with call false deadline none def end unary response blocking state call with call deadline if state code is grpc statuscode ok if with call rendezvous rendezvous state call none deadline return state response rendezvous else return state response else raise rendezvous state none none deadline e grpc channel rendezvous rendezvous of rpc that terminated with e status statuscode deadline exceeded e details deadline exceeded e debug error string created description error received from peer file src core lib surface call cc file line grpc message deadline exceeded grpc status e nox snippets lib site packages grpc channel py rendezvous the above exception was the direct cause of the following exception def test bigtable create table from google cloud bigtable import client from google cloud bigtable import column family client client admin true instance client instance instance id table instance table table my define the gc policy to retain only the most recent versions max versions rule column family maxversionsgcrule table create column families max versions rule docs snippets py google cloud bigtable table py in create initial splits splits google cloud bigtable admin gapic bigtable table admin client py in create table request retry retry timeout timeout metadata metadata api core google api core gapic method py in call return wrapped func args kwargs api core google api core retry py in retry wrapped func on error on error api core google api core retry py in retry target return target api core google api core timeout py in func with timeout return func args kwargs api core google api core grpc helpers py in error remapped callable six raise from exceptions from grpc error exc exc value none from value rendezvous of rpc that terminated with status statuscode deadline exceeded details deadline exceeded debug file src core lib surface call cc file line grpc message deadline exceeded grpc status e google api core exceptions deadlineexceeded deadline exceeded
0
4,673
24,160,487,637
IssuesEvent
2022-09-22 11:13:28
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Hide selection background when only one cell is selected
type: enhancement work: frontend restricted: maintainers status: review
## Current behavior - A blue background displays on all cells that are selected -- even when only one cell is selected. - The background displays during edit mode too, which is particularly weird. ## Desired behavior - The selection cell background is hidden when any of the following conditions are true: - only one cell is selected - The cell is in edit mode (Note that a cell can be in edit mode even when multiple cells are selected, so it's important to use both the above criteria) CC @rajatvijay
True
Hide selection background when only one cell is selected - ## Current behavior - A blue background displays on all cells that are selected -- even when only one cell is selected. - The background displays during edit mode too, which is particularly weird. ## Desired behavior - The selection cell background is hidden when any of the following conditions are true: - only one cell is selected - The cell is in edit mode (Note that a cell can be in edit mode even when multiple cells are selected, so it's important to use both the above criteria) CC @rajatvijay
main
hide selection background when only one cell is selected current behavior a blue background displays on all cells that are selected even when only one cell is selected the background displays during edit mode too which is particularly weird desired behavior the selection cell background is hidden when any of the following conditions are true only one cell is selected the cell is in edit mode note that a cell can be in edit mode even when multiple cells are selected so it s important to use both the above criteria cc rajatvijay
1
616,714
19,318,748,482
IssuesEvent
2021-12-14 01:20:27
bounswe/2021SpringGroup8
https://api.github.com/repos/bounswe/2021SpringGroup8
closed
Choose City API is changed
enhancement Priority - Normal Revision mobile improvement
When a user chooses city in sighnup page, backend stored only name of the city but now, backend wants langitude and lattitude points as well. In android it should be fixed.
1.0
Choose City API is changed - When a user chooses city in sighnup page, backend stored only name of the city but now, backend wants langitude and lattitude points as well. In android it should be fixed.
non_main
choose city api is changed when a user chooses city in sighnup page backend stored only name of the city but now backend wants langitude and lattitude points as well in android it should be fixed
0
384,553
26,592,630,225
IssuesEvent
2023-01-23 09:58:32
starlite-api/starlite
https://api.github.com/repos/starlite-api/starlite
opened
Documentation: How do I log to a file?
documentation
Hi all, apologies if this is answered else where in the documentation or if it is obvious. But, how do I configure starlite logging to log to a file instead of console? I would like to be able to make the logging middleware to log to the file as well. Moreover, how do I use the logger configured by starlite outside of the request function? In fastapi I used to configure the logger manually and do ```python import logging logger = logging.getLogger(__name__) ``` and be done with it. What do I do in starlite? Are there any examples I can refer to? TIA!
1.0
Documentation: How do I log to a file? - Hi all, apologies if this is answered else where in the documentation or if it is obvious. But, how do I configure starlite logging to log to a file instead of console? I would like to be able to make the logging middleware to log to the file as well. Moreover, how do I use the logger configured by starlite outside of the request function? In fastapi I used to configure the logger manually and do ```python import logging logger = logging.getLogger(__name__) ``` and be done with it. What do I do in starlite? Are there any examples I can refer to? TIA!
non_main
documentation how do i log to a file hi all apologies if this is answered else where in the documentation or if it is obvious but how do i configure starlite logging to log to a file instead of console i would like to be able to make the logging middleware to log to the file as well moreover how do i use the logger configured by starlite outside of the request function in fastapi i used to configure the logger manually and do python import logging logger logging getlogger name and be done with it what do i do in starlite are there any examples i can refer to tia
0
354,304
10,565,304,843
IssuesEvent
2019-10-05 10:12:22
EUCweb/BIS-F
https://api.github.com/repos/EUCweb/BIS-F
closed
Citrix BrokerAgent Service set to manual
Priority: Medium wontfix
If ADMX VDA Configuration is enabled, the Citrix BrokerAgent Service is disabled during preparation, it's possible to stop and set to manual. Sometimes is see the BrokerAgent service is not started on BIS-F personalization.
1.0
Citrix BrokerAgent Service set to manual - If ADMX VDA Configuration is enabled, the Citrix BrokerAgent Service is disabled during preparation, it's possible to stop and set to manual. Sometimes is see the BrokerAgent service is not started on BIS-F personalization.
non_main
citrix brokeragent service set to manual if admx vda configuration is enabled the citrix brokeragent service is disabled during preparation it s possible to stop and set to manual sometimes is see the brokeragent service is not started on bis f personalization
0
15,454
19,717,645,004
IssuesEvent
2022-01-13 12:38:41
Creators-of-Create/Create
https://api.github.com/repos/Creators-of-Create/Create
closed
[mc1.16.5_v0.3.1c] Cannot extract fluids from mixer
needs more info from author compatibility stale
I created chocolate from a mixer but as my basin still have milk, I have to filter to import only chocolate. But when I filter on chocolate liquid nothing is imported (works without filter) I tried with a lot of different pipes with filters (Refined Storage, Applied Energistics, Pipez...) but nothing is extracted. Tried with Milk with blacklist but the result is the same. Only works when no filter is applied....
True
[mc1.16.5_v0.3.1c] Cannot extract fluids from mixer - I created chocolate from a mixer but as my basin still have milk, I have to filter to import only chocolate. But when I filter on chocolate liquid nothing is imported (works without filter) I tried with a lot of different pipes with filters (Refined Storage, Applied Energistics, Pipez...) but nothing is extracted. Tried with Milk with blacklist but the result is the same. Only works when no filter is applied....
non_main
cannot extract fluids from mixer i created chocolate from a mixer but as my basin still have milk i have to filter to import only chocolate but when i filter on chocolate liquid nothing is imported works without filter i tried with a lot of different pipes with filters refined storage applied energistics pipez but nothing is extracted tried with milk with blacklist but the result is the same only works when no filter is applied
0
798,399
28,262,986,501
IssuesEvent
2023-04-07 02:15:26
AY2223S2-CS2103-W17-3/tp
https://api.github.com/repos/AY2223S2-CS2103-W17-3/tp
closed
[PE-D][Tester D] Find functionality inconsistent with UserGuide (partial vs full word search)
bug docs priority.Medium
## Bug Report - ezSchedule App ### Issue Description There is an issue with find, where finding strings that are partial in a word will return the event. For example, `find e` will return any events which names' contain the letter e. This contradicts the userguide, which says that find finds based on **keyword**. ![Screenshot 2023-03-31 at 8.12.42 PM.png](https://raw.githubusercontent.com/wengkit1/CS2103-T-PE-Dry-run/main/files/de97d23a-7dda-44ad-b6d7-10e4245666ee.png) ### Expected Behavior I expected the app to find events based on word alone(non-case sensitive). As some words may be part of another word for example, event `eating fish` and event `fishing`, I may only want to find events that contain fish, and not care about events like fishing. ### Actual Behavior Currently, the app's find returns any string as long as they contain any letter passed as a parameter to find. ### Steps to Reproduce 1. Open the ezSchedule app. 2. Add a tennis event 3. Use the find command `find e` ### Environment - OS: Mac OS 13.3 Ventura - App Version: 1.3.1 ### Priority We consider this to be a very low priority bug, as it can cause confusion for users and make it slightly unintuitive to use the app, but not a huge problem. <!--session: 1680242278740-fb3aa958-ae37-4e91-b3e1-1f77cd8ac339--><!--Version: Web v3.4.7--> ------------- Labels: `type.FunctionalityBug` `severity.VeryLow` original: wengkit1/CS2103-T-PE-Dry-run#12
1.0
[PE-D][Tester D] Find functionality inconsistent with UserGuide (partial vs full word search) - ## Bug Report - ezSchedule App ### Issue Description There is an issue with find, where finding strings that are partial in a word will return the event. For example, `find e` will return any events which names' contain the letter e. This contradicts the userguide, which says that find finds based on **keyword**. ![Screenshot 2023-03-31 at 8.12.42 PM.png](https://raw.githubusercontent.com/wengkit1/CS2103-T-PE-Dry-run/main/files/de97d23a-7dda-44ad-b6d7-10e4245666ee.png) ### Expected Behavior I expected the app to find events based on word alone(non-case sensitive). As some words may be part of another word for example, event `eating fish` and event `fishing`, I may only want to find events that contain fish, and not care about events like fishing. ### Actual Behavior Currently, the app's find returns any string as long as they contain any letter passed as a parameter to find. ### Steps to Reproduce 1. Open the ezSchedule app. 2. Add a tennis event 3. Use the find command `find e` ### Environment - OS: Mac OS 13.3 Ventura - App Version: 1.3.1 ### Priority We consider this to be a very low priority bug, as it can cause confusion for users and make it slightly unintuitive to use the app, but not a huge problem. <!--session: 1680242278740-fb3aa958-ae37-4e91-b3e1-1f77cd8ac339--><!--Version: Web v3.4.7--> ------------- Labels: `type.FunctionalityBug` `severity.VeryLow` original: wengkit1/CS2103-T-PE-Dry-run#12
non_main
find functionality inconsistent with userguide partial vs full word search bug report ezschedule app issue description there is an issue with find where finding strings that are partial in a word will return the event for example find e will return any events which names contain the letter e this contradicts the userguide which says that find finds based on keyword expected behavior i expected the app to find events based on word alone non case sensitive as some words may be part of another word for example event eating fish and event fishing i may only want to find events that contain fish and not care about events like fishing actual behavior currently the app s find returns any string as long as they contain any letter passed as a parameter to find steps to reproduce open the ezschedule app add a tennis event use the find command find e environment os mac os ventura app version priority we consider this to be a very low priority bug as it can cause confusion for users and make it slightly unintuitive to use the app but not a huge problem labels type functionalitybug severity verylow original t pe dry run
0
1,164
5,071,492,845
IssuesEvent
2016-12-26 13:53:56
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Docker troubles with ansible hostname module
affects_2.3 bug_report waiting_on_maintainer
Hi, I prepared a usecase that will fail with ansible https://github.com/xrow/docker.ansible.bug1. The output will be like this. I believe the reason why hostnamectl doesn't exists is that the systemd package was replaced by systemd-container. Maybe the solution would be to skip that task if the binary doesn't exist. ``` -bash-4.2# docker build -f Dockerfile -t bash . Sending build context to Docker daemon 7.68 kB Sending build context to Docker daemon Step 0 : FROM centos:7 ---> ce20c473cd8a Step 1 : MAINTAINER "Bjârn Dieding" <bjoern@xrow.de> ---> Using cache ---> d3eec8a3bf48 Step 2 : VOLUME /run /tmp /sys/fs/cgroup ---> Using cache ---> bb410b68dadb Step 3 : ADD RPM-GPG-KEY-EPEL-7 /etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-7 ---> Using cache ---> ac7e3e116967 Step 4 : ADD epel.repo /etc/yum.repos.d/epel.repo ---> Using cache ---> de7ff6cb9763 Step 5 : RUN yum install -y ansible ---> Using cache ---> 32686ff89319 Step 6 : RUN yum -y install python-pip redis ---> Using cache ---> 4bbd888b3814 Step 7 : RUN pip install redis ---> Using cache ---> b1b1aad0c28e Step 8 : ADD test.yml test.yml ---> Using cache ---> 76747563e5da Step 9 : RUN echo "[ezcluster]" > /etc/ansible/hosts ---> Using cache ---> b5c1c7fa2430 Step 10 : RUN echo "localhost" >> /etc/ansible/hosts ---> Using cache ---> a0916c3c8ee2 Step 11 : RUN ansible-playbook -c local test.yml -vvvv ---> Running in 338a9e2b6ec4 PLAY [ezcluster] ************************************************************** GATHERING FACTS *************************************************************** <localhost> REMOTE_MODULE setup <localhost> EXEC ['/bin/sh', '-c', 'mkdir -p $HOME/.ansible/tmp/ansible-tmp-1445501553.12-214209363520399 && echo $HOME/.ansible/tmp/ansible-tmp-1445501553.12-214209363520399'] <localhost> PUT /tmp/tmpoBWhBT TO /root/.ansible/tmp/ansible-tmp-1445501553.12-214209363520399/setup <localhost> EXEC ['/bin/sh', '-c', u'LANG=C LC_CTYPE=C /usr/bin/python /root/.ansible/tmp/ansible-tmp-1445501553.12-214209363520399/setup; rm -rf /root/.ansible/tmp/ansible-tmp-1445501553.12-214209363520399/ >/dev/null 2>&1'] ok: [localhost] TASK: [hostname name=myhostname] ********************************************** <localhost> REMOTE_MODULE hostname name=myhostname <localhost> EXEC ['/bin/sh', '-c', 'mkdir -p $HOME/.ansible/tmp/ansible-tmp-1445501554.51-46884008533714 && echo $HOME/.ansible/tmp/ansible-tmp-1445501554.51-46884008533714'] <localhost> PUT /tmp/tmpNbuHLK TO /root/.ansible/tmp/ansible-tmp-1445501554.51-46884008533714/hostname <localhost> EXEC ['/bin/sh', '-c', u'LANG=C LC_CTYPE=C /usr/bin/python /root/.ansible/tmp/ansible-tmp-1445501554.51-46884008533714/hostname; rm -rf /root/.ansible/tmp/ansible-tmp-1445501554.51-46884008533714/ >/dev/null 2>&1'] failed: [localhost] => {"cmd": "hostnamectl --transient set-hostname myhostname", "failed": true, "rc": 2} msg: [Errno 2] No such file or directory FATAL: all hosts have already failed -- aborting PLAY RECAP ******************************************************************** to retry, use: --limit @/root/test.retry localhost : ok=1 changed=0 unreachable=0 failed=1 The command '/bin/sh -c ansible-playbook -c local test.yml -vvvv' returned a non-zero code: 2 ```
True
Docker troubles with ansible hostname module - Hi, I prepared a usecase that will fail with ansible https://github.com/xrow/docker.ansible.bug1. The output will be like this. I believe the reason why hostnamectl doesn't exists is that the systemd package was replaced by systemd-container. Maybe the solution would be to skip that task if the binary doesn't exist. ``` -bash-4.2# docker build -f Dockerfile -t bash . Sending build context to Docker daemon 7.68 kB Sending build context to Docker daemon Step 0 : FROM centos:7 ---> ce20c473cd8a Step 1 : MAINTAINER "Bjârn Dieding" <bjoern@xrow.de> ---> Using cache ---> d3eec8a3bf48 Step 2 : VOLUME /run /tmp /sys/fs/cgroup ---> Using cache ---> bb410b68dadb Step 3 : ADD RPM-GPG-KEY-EPEL-7 /etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-7 ---> Using cache ---> ac7e3e116967 Step 4 : ADD epel.repo /etc/yum.repos.d/epel.repo ---> Using cache ---> de7ff6cb9763 Step 5 : RUN yum install -y ansible ---> Using cache ---> 32686ff89319 Step 6 : RUN yum -y install python-pip redis ---> Using cache ---> 4bbd888b3814 Step 7 : RUN pip install redis ---> Using cache ---> b1b1aad0c28e Step 8 : ADD test.yml test.yml ---> Using cache ---> 76747563e5da Step 9 : RUN echo "[ezcluster]" > /etc/ansible/hosts ---> Using cache ---> b5c1c7fa2430 Step 10 : RUN echo "localhost" >> /etc/ansible/hosts ---> Using cache ---> a0916c3c8ee2 Step 11 : RUN ansible-playbook -c local test.yml -vvvv ---> Running in 338a9e2b6ec4 PLAY [ezcluster] ************************************************************** GATHERING FACTS *************************************************************** <localhost> REMOTE_MODULE setup <localhost> EXEC ['/bin/sh', '-c', 'mkdir -p $HOME/.ansible/tmp/ansible-tmp-1445501553.12-214209363520399 && echo $HOME/.ansible/tmp/ansible-tmp-1445501553.12-214209363520399'] <localhost> PUT /tmp/tmpoBWhBT TO /root/.ansible/tmp/ansible-tmp-1445501553.12-214209363520399/setup <localhost> EXEC ['/bin/sh', '-c', u'LANG=C LC_CTYPE=C /usr/bin/python /root/.ansible/tmp/ansible-tmp-1445501553.12-214209363520399/setup; rm -rf /root/.ansible/tmp/ansible-tmp-1445501553.12-214209363520399/ >/dev/null 2>&1'] ok: [localhost] TASK: [hostname name=myhostname] ********************************************** <localhost> REMOTE_MODULE hostname name=myhostname <localhost> EXEC ['/bin/sh', '-c', 'mkdir -p $HOME/.ansible/tmp/ansible-tmp-1445501554.51-46884008533714 && echo $HOME/.ansible/tmp/ansible-tmp-1445501554.51-46884008533714'] <localhost> PUT /tmp/tmpNbuHLK TO /root/.ansible/tmp/ansible-tmp-1445501554.51-46884008533714/hostname <localhost> EXEC ['/bin/sh', '-c', u'LANG=C LC_CTYPE=C /usr/bin/python /root/.ansible/tmp/ansible-tmp-1445501554.51-46884008533714/hostname; rm -rf /root/.ansible/tmp/ansible-tmp-1445501554.51-46884008533714/ >/dev/null 2>&1'] failed: [localhost] => {"cmd": "hostnamectl --transient set-hostname myhostname", "failed": true, "rc": 2} msg: [Errno 2] No such file or directory FATAL: all hosts have already failed -- aborting PLAY RECAP ******************************************************************** to retry, use: --limit @/root/test.retry localhost : ok=1 changed=0 unreachable=0 failed=1 The command '/bin/sh -c ansible-playbook -c local test.yml -vvvv' returned a non-zero code: 2 ```
main
docker troubles with ansible hostname module hi i prepared a usecase that will fail with ansible the output will be like this i believe the reason why hostnamectl doesn t exists is that the systemd package was replaced by systemd container maybe the solution would be to skip that task if the binary doesn t exist bash docker build f dockerfile t bash sending build context to docker daemon kb sending build context to docker daemon step from centos step maintainer bjã¢rn dieding using cache step volume run tmp sys fs cgroup using cache step add rpm gpg key epel etc pki rpm gpg rpm gpg key epel using cache step add epel repo etc yum repos d epel repo using cache step run yum install y ansible using cache step run yum y install python pip redis using cache step run pip install redis using cache step add test yml test yml using cache step run echo etc ansible hosts using cache step run echo localhost etc ansible hosts using cache step run ansible playbook c local test yml vvvv running in play gathering facts remote module setup exec put tmp tmpobwhbt to root ansible tmp ansible tmp setup exec ok task remote module hostname name myhostname exec put tmp tmpnbuhlk to root ansible tmp ansible tmp hostname exec failed cmd hostnamectl transient set hostname myhostname failed true rc msg no such file or directory fatal all hosts have already failed aborting play recap to retry use limit root test retry localhost ok changed unreachable failed the command bin sh c ansible playbook c local test yml vvvv returned a non zero code
1
11,029
4,868,275,598
IssuesEvent
2016-11-15 08:51:45
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
Cat tests expect indices to be returned in a particular order
build
This [build](https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+java9-periodic/656/console) failed because the `_cat` API returned index `twitter2` before `twitter`, while the regular expression expected them to occur in the opposite order. I think we should either fix the test to not care about the order or the cat API to return indices in a specific order (eg. sorted alphabetically?).
1.0
Cat tests expect indices to be returned in a particular order - This [build](https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+java9-periodic/656/console) failed because the `_cat` API returned index `twitter2` before `twitter`, while the regular expression expected them to occur in the opposite order. I think we should either fix the test to not care about the order or the cat API to return indices in a specific order (eg. sorted alphabetically?).
non_main
cat tests expect indices to be returned in a particular order this failed because the cat api returned index before twitter while the regular expression expected them to occur in the opposite order i think we should either fix the test to not care about the order or the cat api to return indices in a specific order eg sorted alphabetically
0
2,535
8,657,435,098
IssuesEvent
2018-11-27 21:19:14
Kapeli/Dash-User-Contributions
https://api.github.com/repos/Kapeli/Dash-User-Contributions
closed
Flow Docset maintainer needed
needs maintainer
I can no longer have time to maintain this docset and I am looking for additional contributors to assist. My repo is located at [https://github.com/epitaphmike/flow-dash](https://github.com/epitaphmike/flow-dash). If this is something you are interested in helping with please reach out. Thank you.
True
Flow Docset maintainer needed - I can no longer have time to maintain this docset and I am looking for additional contributors to assist. My repo is located at [https://github.com/epitaphmike/flow-dash](https://github.com/epitaphmike/flow-dash). If this is something you are interested in helping with please reach out. Thank you.
main
flow docset maintainer needed i can no longer have time to maintain this docset and i am looking for additional contributors to assist my repo is located at if this is something you are interested in helping with please reach out thank you
1
5,407
27,144,114,236
IssuesEvent
2023-02-16 18:30:23
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Access control in Table page: general -
work: frontend status: ready restricted: maintainers
- Editing the spreadsheet - Adding new columns - Operations in right-click context menus
True
Access control in Table page: general - - - Editing the spreadsheet - Adding new columns - Operations in right-click context menus
main
access control in table page general editing the spreadsheet adding new columns operations in right click context menus
1
5,140
26,201,103,346
IssuesEvent
2023-01-03 17:32:25
omigroup/omigroup
https://api.github.com/repos/omigroup/omigroup
closed
Let's learn from Sindhu
Make the metaverse more human Cultivate Resiliency Maintain sustainable innovation
Let's learn from Sindhu, who has extensive experience as both a Scrum Master, and Project Manager. Sindhu has identified a few gaps in our communication approach to help people actually get involved and get work done. What can we learn? How can we channel those learnings into our working groups who want to learn? Maybe use this ticket as a example ticket to track the journey per @fire 's recommendation. Shirley mentioned we have systems in place it may just not be obvious where those are. RedEagle has been advocating for a welcome process for onboarding newcomers. Jin has been an advocate of posting known tasks/tickets on a site like dework.xyz to incentivize known to-do's. Many people shared concerns that we may be losing people who don't come back because they can't easily see how to get involved. _Originally posted by @mrmetaverse in https://github.com/omigroup/omigroup/discussions/302#discussioncomment-3862513_
True
Let's learn from Sindhu - Let's learn from Sindhu, who has extensive experience as both a Scrum Master, and Project Manager. Sindhu has identified a few gaps in our communication approach to help people actually get involved and get work done. What can we learn? How can we channel those learnings into our working groups who want to learn? Maybe use this ticket as a example ticket to track the journey per @fire 's recommendation. Shirley mentioned we have systems in place it may just not be obvious where those are. RedEagle has been advocating for a welcome process for onboarding newcomers. Jin has been an advocate of posting known tasks/tickets on a site like dework.xyz to incentivize known to-do's. Many people shared concerns that we may be losing people who don't come back because they can't easily see how to get involved. _Originally posted by @mrmetaverse in https://github.com/omigroup/omigroup/discussions/302#discussioncomment-3862513_
main
let s learn from sindhu let s learn from sindhu who has extensive experience as both a scrum master and project manager sindhu has identified a few gaps in our communication approach to help people actually get involved and get work done what can we learn how can we channel those learnings into our working groups who want to learn maybe use this ticket as a example ticket to track the journey per fire s recommendation shirley mentioned we have systems in place it may just not be obvious where those are redeagle has been advocating for a welcome process for onboarding newcomers jin has been an advocate of posting known tasks tickets on a site like dework xyz to incentivize known to do s many people shared concerns that we may be losing people who don t come back because they can t easily see how to get involved originally posted by mrmetaverse in
1
5,546
27,769,433,836
IssuesEvent
2023-03-16 13:37:22
oele-isis-vanderbilt/ChimeraPy
https://api.github.com/repos/oele-isis-vanderbilt/ChimeraPy
closed
Segment Test Execution by Markers
enhancement maintainence
It might be nice to segment test execution by `slow` and `not slow` markers introduced in #130.
True
Segment Test Execution by Markers - It might be nice to segment test execution by `slow` and `not slow` markers introduced in #130.
main
segment test execution by markers it might be nice to segment test execution by slow and not slow markers introduced in
1
5,135
26,184,079,321
IssuesEvent
2023-01-02 20:18:03
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
opened
Remove backend commands exposed for each operation
enhancement maintainability
So far, each operation that the user can run on a project comes with the following Java classes in the backend: * an Operation class, which holds the metadata for the operation and is responsible for its JSON serialization (which is exposed in the history tab, among others) * a Change class (often reused by different operations), which is responsible for actually applying the operation to the project (carrying out the corresponding transformation) * a Command class, which exposes an HTTP API to initiate the operation on a project. Therefore, each operation comes with its own HTTP endpoint to apply it, and the frontend can call that endpoint when the user clicks on some menu item or validates some dialog, for instance. At the same time, we have a generic endpoint to apply a series of operations, represented by their JSON serialization, on a given project. This is used in the history tab, to let the user apply a workflow exported in JSON. This endpoint (called `apply-operations`) makes all the other operation-specific commands redundant: we should be able to use it to apply any single operation, hence removing the need for a fairly large number of other commands. This would have the following benefits: * remove a lot of Java classes * refactor the frontend so that making any change to the project goes through the same code, making it easier to implement logic that should apply to all those changes (such as warning the user that they are discarding a part of the history if there are "future" history entries: #3184) * make it easier to implement new operations, by removing the need to implement a corresponding operation (but the move can be done without any breaking change with respect to the extension interface - extensions are still free to define their own commands for each operation they define). This will likely require improving the error handling in the `apply-operations` endpoint, since it is likely that operation-specific endpoints currently offer a better handling. When an operation's JSON representation is incomplete or invalid, we want to be able to return an appropriate response which should be replicated in the frontend accordingly. This will benefit the "Apply operations" UX as well.
True
Remove backend commands exposed for each operation - So far, each operation that the user can run on a project comes with the following Java classes in the backend: * an Operation class, which holds the metadata for the operation and is responsible for its JSON serialization (which is exposed in the history tab, among others) * a Change class (often reused by different operations), which is responsible for actually applying the operation to the project (carrying out the corresponding transformation) * a Command class, which exposes an HTTP API to initiate the operation on a project. Therefore, each operation comes with its own HTTP endpoint to apply it, and the frontend can call that endpoint when the user clicks on some menu item or validates some dialog, for instance. At the same time, we have a generic endpoint to apply a series of operations, represented by their JSON serialization, on a given project. This is used in the history tab, to let the user apply a workflow exported in JSON. This endpoint (called `apply-operations`) makes all the other operation-specific commands redundant: we should be able to use it to apply any single operation, hence removing the need for a fairly large number of other commands. This would have the following benefits: * remove a lot of Java classes * refactor the frontend so that making any change to the project goes through the same code, making it easier to implement logic that should apply to all those changes (such as warning the user that they are discarding a part of the history if there are "future" history entries: #3184) * make it easier to implement new operations, by removing the need to implement a corresponding operation (but the move can be done without any breaking change with respect to the extension interface - extensions are still free to define their own commands for each operation they define). This will likely require improving the error handling in the `apply-operations` endpoint, since it is likely that operation-specific endpoints currently offer a better handling. When an operation's JSON representation is incomplete or invalid, we want to be able to return an appropriate response which should be replicated in the frontend accordingly. This will benefit the "Apply operations" UX as well.
main
remove backend commands exposed for each operation so far each operation that the user can run on a project comes with the following java classes in the backend an operation class which holds the metadata for the operation and is responsible for its json serialization which is exposed in the history tab among others a change class often reused by different operations which is responsible for actually applying the operation to the project carrying out the corresponding transformation a command class which exposes an http api to initiate the operation on a project therefore each operation comes with its own http endpoint to apply it and the frontend can call that endpoint when the user clicks on some menu item or validates some dialog for instance at the same time we have a generic endpoint to apply a series of operations represented by their json serialization on a given project this is used in the history tab to let the user apply a workflow exported in json this endpoint called apply operations makes all the other operation specific commands redundant we should be able to use it to apply any single operation hence removing the need for a fairly large number of other commands this would have the following benefits remove a lot of java classes refactor the frontend so that making any change to the project goes through the same code making it easier to implement logic that should apply to all those changes such as warning the user that they are discarding a part of the history if there are future history entries make it easier to implement new operations by removing the need to implement a corresponding operation but the move can be done without any breaking change with respect to the extension interface extensions are still free to define their own commands for each operation they define this will likely require improving the error handling in the apply operations endpoint since it is likely that operation specific endpoints currently offer a better handling when an operation s json representation is incomplete or invalid we want to be able to return an appropriate response which should be replicated in the frontend accordingly this will benefit the apply operations ux as well
1
504,516
14,620,016,818
IssuesEvent
2020-12-22 18:55:21
metal3-io/baremetal-operator
https://api.github.com/repos/metal3-io/baremetal-operator
closed
On bmo pod restart unable to provision ready state BMH's
kind/bug lifecycle/stale priority/backlog
There appears to be a bug where after creating a number of bmh that go through the registering -> inspecting -> ready phases, after which the baremetal operator pod is restarted, you are no longer able to provision any kube machines on these bmhs, and the following error is thrown: `host validation error: Node 3a0e38e3-6d5f-424e-b4cb-cb4758506da5 does not have any port associated with it.; Node 3a0e38e3-6d5f-424e-b4cb-cb4758506da5 does not have any port associated with it.` Here's the full output and and logs from bmo: ``` kubectl get bmh ξ‚² βœ” ξ‚² 11342 ξ‚² 11:06:45 NAME STATUS PROVISIONING STATUS CONSUMER BMC HARDWARE PROFILE ONLINE ERROR baremetal0003d1mdw1.sendgrid.net OK inspecting ipmi://10.16.5.211/ true baremetal0004d1mdw1.sendgrid.net OK ready ipmi://10.16.6.14/ unknown true baremetal0005d1mdw1.sendgrid.net OK ready ipmi://10.16.5.227/ unknown true baremetal0006d1mdw1.sendgrid.net OK ready ipmi://10.16.5.241/ unknown true baremetal0007d1mdw1.sendgrid.net OK ready ipmi://10.16.5.239/ unknown true baremetal0008d1mdw1.sendgrid.net OK ready ipmi://10.16.5.235/ unknown true baremetal0009d1mdw1.sendgrid.net OK ready ipmi://10.16.5.207/ unknown true baremetal0010d1mdw1.sendgrid.net OK ready ipmi://10.16.5.170/ unknown true baremetal0011d1mdw1.sendgrid.net OK ready ipmi://10.16.5.179/ unknown true baremetal0012d1mdw1.sendgrid.net OK ready ipmi://10.16.5.237/ unknown true baremetal0013d1mdw1.sendgrid.net OK ready ipmi://10.16.6.18/ unknown true baremetal0014d1mdw1.sendgrid.net OK ready ipmi://10.16.6.60/ unknown true baremetal0015d1mdw1.sendgrid.net OK ready ipmi://10.16.5.185/ unknown true baremetal0016d1mdw1.sendgrid.net OK ready ipmi://10.16.5.164/ unknown true baremetal0017d1mdw1.sendgrid.net OK ready ipmi://10.16.5.183/ unknown true baremetal0018d1mdw1.sendgrid.net OK ready ipmi://10.16.5.242/ unknown true baremetal0019d1mdw1.sendgrid.net OK ready ipmi://10.16.5.174/ unknown true lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° ξ‚² βœ” ξ‚² 11343 ξ‚² 11:06:46 lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° kubectl apply -f _out/ ξ‚² 1 ↡ ξ‚² 11344 ξ‚² 11:06:54 cluster.cluster.x-k8s.io/lukasz unchanged baremetalcluster.infrastructure.cluster.x-k8s.io/lukasz unchanged kubeadmconfig.bootstrap.cluster.x-k8s.io/lukasz-controlplane-0 created kubeadmconfig.bootstrap.cluster.x-k8s.io/lukasz-controlplane-1 created kubeadmconfig.bootstrap.cluster.x-k8s.io/lukasz-controlplane-2 created machine.cluster.x-k8s.io/lukasz-controlplane-0 created machine.cluster.x-k8s.io/lukasz-controlplane-1 created machine.cluster.x-k8s.io/lukasz-controlplane-2 created baremetalmachine.infrastructure.cluster.x-k8s.io/lukasz-controlplane-0 created baremetalmachine.infrastructure.cluster.x-k8s.io/lukasz-controlplane-1 created baremetalmachine.infrastructure.cluster.x-k8s.io/lukasz-controlplane-2 created kubeadmconfigtemplate.bootstrap.cluster.x-k8s.io/lukasz-md-0 created kubeadmconfigtemplate.bootstrap.cluster.x-k8s.io/lukasz-old-md-0 created machinedeployment.cluster.x-k8s.io/lukasz-md-0 created machinedeployment.cluster.x-k8s.io/lukasz-old-md-0 created baremetalmachinetemplate.infrastructure.cluster.x-k8s.io/lukasz-md-0 created baremetalmachinetemplate.infrastructure.cluster.x-k8s.io/lukasz-old-md-0 created lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° ξ‚² βœ” ξ‚² 11345 ξ‚² 11:06:59 lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° kubectl get machine ξ‚² βœ” ξ‚² 11345 ξ‚² 11:06:59 NAME PROVIDERID PHASE lukasz-controlplane-0 provisioning lukasz-controlplane-1 pending lukasz-controlplane-2 pending lukasz-md-0-cf48d76f6-jjrv7 pending lukasz-old-md-0-5d997d4449-r4s5r pending lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° kubectl get bmh ξ‚² βœ” ξ‚² 11346 ξ‚² 11:07:07 NAME STATUS PROVISIONING STATUS CONSUMER BMC HARDWARE PROFILE ONLINE ERROR baremetal0003d1mdw1.sendgrid.net OK inspecting ipmi://10.16.5.211/ true baremetal0004d1mdw1.sendgrid.net OK ready ipmi://10.16.6.14/ unknown true baremetal0005d1mdw1.sendgrid.net OK ready ipmi://10.16.5.227/ unknown true baremetal0006d1mdw1.sendgrid.net OK ready ipmi://10.16.5.241/ unknown true baremetal0007d1mdw1.sendgrid.net OK ready ipmi://10.16.5.239/ unknown true baremetal0008d1mdw1.sendgrid.net OK ready ipmi://10.16.5.235/ unknown true baremetal0009d1mdw1.sendgrid.net OK ready ipmi://10.16.5.207/ unknown true baremetal0010d1mdw1.sendgrid.net OK ready ipmi://10.16.5.170/ unknown true baremetal0011d1mdw1.sendgrid.net OK ready ipmi://10.16.5.179/ unknown true baremetal0012d1mdw1.sendgrid.net OK ready ipmi://10.16.5.237/ unknown true baremetal0013d1mdw1.sendgrid.net OK ready ipmi://10.16.6.18/ unknown true baremetal0014d1mdw1.sendgrid.net OK ready ipmi://10.16.6.60/ unknown true baremetal0015d1mdw1.sendgrid.net error provisioning lukasz-controlplane-0 ipmi://10.16.5.185/ unknown true host validation error: Node 3a0e38e3-6d5f-424e-b4cb-cb4758506da5 does not have any port associated with it.; Node 3a0e38e3-6d5f-424e-b4cb-cb4758506da5 does not have any port associated with it. baremetal0016d1mdw1.sendgrid.net OK ready ipmi://10.16.5.164/ unknown true baremetal0017d1mdw1.sendgrid.net OK ready ipmi://10.16.5.183/ unknown true baremetal0018d1mdw1.sendgrid.net OK ready ipmi://10.16.5.242/ unknown true baremetal0019d1mdw1.sendgrid.net OK ready ipmi://10.16.5.174/ unknown true lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° ξ‚² βœ” ξ‚² 11347 ξ‚² 11:07:20 lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° {"level":"info","ts":1581703726.561492,"logger":"baremetalhost","msg":"Reconciling BareMetalHost","Request.Namespace":"metal3","Request.Name":"baremetal0003d1mdw1.sendgrid.net"} {"level":"info","ts":1581703726.5615559,"logger":"baremetalhost","msg":"inspecting hardware","Request.Namespace":"metal3","Request.Name":"baremetal0003d1mdw1.sendgrid.net","provisioningState":"inspecting"} {"level":"info","ts":1581703726.5615826,"logger":"baremetalhost_ironic","msg":"inspecting hardware","host":"baremetal0003d1mdw1.sendgrid.net","status":"OK"} {"level":"info","ts":1581703726.5742512,"logger":"baremetalhost_ironic","msg":"looking for existing node by name","host":"baremetal0003d1mdw1.sendgrid.net","name":"baremetal0003d1mdw1.sendgrid.net"} {"level":"error","ts":1581703726.5862474,"logger":"controller-runtime.controller","msg":"Reconciler error","controller":"metal3-baremetalhost-controller","request":"metal3/baremetal0003d1mdw1.sendgrid.net","error":"action \"inspecting\" failed: hardware inspection failed: no ironic node for host","errorVerbose":"no ironic node for host\nhardware inspection failed\ngithub.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).actionInspecting\n\t/go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:491\ngithub.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).handleInspecting\n\t/go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:197\ngithub.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).(github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.handleInspecting)-fm\n\t/go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:41\ngithub.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).ReconcileState\n\t/go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:109\ngithub.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).Reconcile\n\t/go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:281\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:256\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:232\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:211\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).(github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.worker)-fm\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:2361\naction \"inspecting\" failed\ngithub.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).Reconcile\n\t/go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:285\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:256\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:232\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:211\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).(github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.worker)-fm\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:2361","stacktrace":"github.com/metal3-io/baremetal-operator/vendor/github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/github.com/go-logr/zapr/zapr.go:128\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:258\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:232\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:211\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).(github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.worker)-fm\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} {"level":"info","ts":1581703727.5868793,"logger":"baremetalhost","msg":"Reconciling BareMetalHost","Request.Namespace":"metal3","Request.Name":"baremetal0004d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.5870068,"logger":"baremetalhost_ironic","msg":"validating management access","host":"baremetal0004d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.6093876,"logger":"baremetalhost_ironic","msg":"found existing node by ID","host":"baremetal0004d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.6094184,"logger":"baremetalhost_ironic","msg":"current provision state","host":"baremetal0004d1mdw1.sendgrid.net","lastError":"","current":"manageable","target":""} {"level":"info","ts":1581703727.60943,"logger":"baremetalhost_ironic","msg":"have manageable host","host":"baremetal0004d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.609438,"logger":"baremetalhost_ironic","msg":"updating hardware state","host":"baremetal0004d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.6298566,"logger":"baremetalhost_ironic","msg":"found existing node by ID","host":"baremetal0004d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.6298933,"logger":"baremetalhost","msg":"saving host status","Request.Namespace":"metal3","Request.Name":"baremetal0004d1mdw1.sendgrid.net","provisioningState":"ready","operational status":"OK","provisioning state":"ready"} {"level":"info","ts":1581703727.6373432,"logger":"baremetalhost","msg":"done","Request.Namespace":"metal3","Request.Name":"baremetal0004d1mdw1.sendgrid.net","provisioningState":"ready","requeue":true,"after":60} {"level":"info","ts":1581703727.637434,"logger":"baremetalhost","msg":"Reconciling BareMetalHost","Request.Namespace":"metal3","Request.Name":"baremetal0015d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.6375093,"logger":"baremetalhost","msg":"provisioning","Request.Namespace":"metal3","Request.Name":"baremetal0015d1mdw1.sendgrid.net","provisioningState":"provisioning"} {"level":"info","ts":1581703727.6551394,"logger":"baremetalhost_ironic","msg":"found existing node by ID","host":"baremetal0015d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.6551712,"logger":"baremetalhost_ironic","msg":"provisioning image to host","host":"baremetal0015d1mdw1.sendgrid.net","state":"manageable"} E0214 18:08:47.655283 1 runtime.go:78] Observed a panic: "invalid memory address or nil pointer dereference" (runtime error: invalid memory address or nil pointer dereference) goroutine 270 [running]: github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/runtime.logPanic(0x137efe0, 0x20309c0) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:74 +0xaa github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/runtime.HandleCrash(0x0, 0x0, 0x0) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:48 +0x82 panic(0x137efe0, 0x20309c0) /usr/local/go/src/runtime/panic.go:502 +0x229 github.com/metal3-io/baremetal-operator/pkg/provisioner/ironic.(*ironicProvisioner).Provision(0xc420169030, 0x1649da0, 0xc4207c7a40, 0x0, 0x0, 0x0, 0xa858015299444fb4, 0x4122f8, 0x10) /go/src/github.com/metal3-io/baremetal-operator/pkg/provisioner/ironic/ironic.go:924 +0x156 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).actionProvisioning(0xc4208b9ac0, 0x1672860, 0xc420169030, 0xc4207f7400, 0x9e0000c420787260, 0xc420787240) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:563 +0x117 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).handleProvisioning(0xc4207c79e0, 0xc4207f7400, 0xc4207871f0, 0xc4200e1790) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:246 +0xb8 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).(github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.handleProvisioning)-fm(0xc4207f7400, 0xc4207c7a10, 0xc42047ac90) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:45 +0x34 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).ReconcileState(0xc4207c79e0, 0xc4207f7400, 0x0, 0x0) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:109 +0x327 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).Reconcile(0xc4208b9ac0, 0xc42047a8da, 0x6, 0xc4206b17e0, 0x20, 0x204ac00, 0x0, 0x0, 0x0) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:281 +0x987 github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler(0xc4205623c0, 0x13c9ea0, 0xc420916ee0, 0x13c9e00) /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:256 +0x100 github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem(0xc4205623c0, 0xc4206be300) /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:232 +0xb7 github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker(0xc4205623c0) /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:211 +0x2b github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).(github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.worker)-fm() /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193 +0x2a github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1(0xc42065a490) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152 +0x54 github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil(0xc42065a490, 0x3b9aca00, 0x0, 0x1, 0xc420023a40) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153 +0xbd github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.Until(0xc42065a490, 0x3b9aca00, 0xc420023a40) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88 +0x4d created by github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1 /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193 +0x63b panic: runtime error: invalid memory address or nil pointer dereference [recovered] panic: runtime error: invalid memory address or nil pointer dereference [signal SIGSEGV: segmentation violation code=0x1 addr=0x18 pc=0xb40e76] goroutine 270 [running]: github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/runtime.HandleCrash(0x0, 0x0, 0x0) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:55 +0x107 panic(0x137efe0, 0x20309c0) /usr/local/go/src/runtime/panic.go:502 +0x229 github.com/metal3-io/baremetal-operator/pkg/provisioner/ironic.(*ironicProvisioner).Provision(0xc420169030, 0x1649da0, 0xc4207c7a40, 0x0, 0x0, 0x0, 0xa858015299444fb4, 0x4122f8, 0x10) /go/src/github.com/metal3-io/baremetal-operator/pkg/provisioner/ironic/ironic.go:924 +0x156 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).actionProvisioning(0xc4208b9ac0, 0x1672860, 0xc420169030, 0xc4207f7400, 0x9e0000c420787260, 0xc420787240) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:563 +0x117 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).handleProvisioning(0xc4207c79e0, 0xc4207f7400, 0xc4207871f0, 0xc4200e1790) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:246 +0xb8 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).(github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.handleProvisioning)-fm(0xc4207f7400, 0xc4207c7a10, 0xc42047ac90) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:45 +0x34 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).ReconcileState(0xc4207c79e0, 0xc4207f7400, 0x0, 0x0) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:109 +0x327 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).Reconcile(0xc4208b9ac0, 0xc42047a8da, 0x6, 0xc4206b17e0, 0x20, 0x204ac00, 0x0, 0x0, 0x0) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:281 +0x987 github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler(0xc4205623c0, 0x13c9ea0, 0xc420916ee0, 0x13c9e00) /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:256 +0x100 github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem(0xc4205623c0, 0xc4206be300) /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:232 +0xb7 github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker(0xc4205623c0) /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:211 +0x2b github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).(github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.worker)-fm() /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193 +0x2a github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1(0xc42065a490) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152 +0x54 github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil(0xc42065a490, 0x3b9aca00, 0x0, 0x1, 0xc420023a40) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153 +0xbd github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.Until(0xc42065a490, 0x3b9aca00, 0xc420023a40) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88 +0x4d created by github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1 /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193 +0x63b ``` Steps to recreate: Use bmo as normal and get some bmh into the ready state. Bounce the pod. After the operator comes back up attempt to provision a new cluster the above error should be thrown. We have also encountered a similar error, although haven't been able to reproduce it 100% of the time as the one above, where a bmh previous to the pod bounce was in the provisioned state, in the example provided we provisioned the bmh by adding the image to the spec it wasn't provisioned as part of a cluster, but we have seen the exact same issue below when the bmhs were provisioned as part of a cluster, after the pod is bounced do not get adopted properly, i.e.: ``` baremetal0012d1mdw1.sendgrid.net error registration error ipmi://10.16.5.237/ unknown true Host adoption failed: Error while attempting to adopt node 60895ea0-419a-4138-aac3-f4fd72f320a4: Node 60895ea0-419a-4138-aac3-f4fd72f320a4 does not have any port associated with it.. ``` And the logs for the event: ``` {"level":"info","ts":1582047700.842536,"logger":"baremetalhost","msg":"Reconciling BareMetalHost","Request.Namespace":"metal3","Request.Name":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047700.856129,"logger":"baremetalhost_ironic","msg":"looking for existing node by name","host":"baremetal0012d1mdw1.sendgrid.net","name":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047700.8713014,"logger":"baremetalhost_ironic","msg":"re-registering host","host":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047700.8713233,"logger":"baremetalhost_ironic","msg":"validating management access","host":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047700.886314,"logger":"baremetalhost_ironic","msg":"looking for existing node by name","host":"baremetal0012d1mdw1.sendgrid.net","name":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047700.8993719,"logger":"baremetalhost_ironic","msg":"registering host in ironic","host":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047700.9414284,"logger":"baremetalhost_ironic","msg":"setting provisioning id","host":"baremetal0012d1mdw1.sendgrid.net","ID":"60895ea0-419a-4138-aac3-f4fd72f320a4"} {"level":"info","ts":1582047700.941458,"logger":"baremetalhost_ironic","msg":"setting instance info","host":"baremetal0012d1mdw1.sendgrid.net","image_source":"https://filestore-staging.sendgrid.net/met al3-image-build/fedora-31-metal3-0.0.20200213231225.qcow2","checksum":"https://filestore-staging.sendgrid.net/metal3-image-build/fedora-31-metal3-0.0.20200213231225.qcow2.md5sum"} {"level":"info","ts":1582047701.0355387,"logger":"baremetalhost_ironic","msg":"current provision state","host":"baremetal0012d1mdw1.sendgrid.net","lastError":"","current":"enroll","target":""} {"level":"info","ts":1582047701.0355792,"logger":"baremetalhost_ironic","msg":"changing provisioning state","host":"baremetal0012d1mdw1.sendgrid.net","current":"enroll","existing target":"","new target ":"manage"} {"level":"info","ts":1582047701.1334321,"logger":"baremetalhost","msg":"saving host status","Request.Namespace":"metal3","Request.Name":"baremetal0012d1mdw1.sendgrid.net","provisioningState":"provision ed","operational status":"OK","provisioning state":"provisioned"} {"level":"info","ts":1582047701.1406105,"logger":"baremetalhost","msg":"publishing event","reason":"Registered","message":"Registered new host"} {"level":"info","ts":1582047701.144376,"logger":"baremetalhost","msg":"done","Request.Namespace":"metal3","Request.Name":"baremetal0012d1mdw1.sendgrid.net","provisioningState":"provisioned","requeue":t rue,"after":10} A little while later {"level":"info","ts":1582047704.92819,"logger":"baremetalhost","msg":"Reconciling BareMetalHost","Request.Namespace":"metal3","Request.Name":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047704.9464087,"logger":"baremetalhost_ironic","msg":"found existing node by ID","host":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047704.9464579,"logger":"baremetalhost","msg":"saving host status","Request.Namespace":"metal3","Request.Name":"baremetal0012d1mdw1.sendgrid.net","provisioningState":"provisioned","operational status":"error","provisioning state":"registration error"} {"level":"info","ts":1582047704.9539585,"logger":"baremetalhost","msg":"publishing event","reason":"RegistrationError","message":"Host adoption failed: Error while attempting to adopt node 60895ea0-419a-4138-aac3-f4fd72f320a4: Node 60895ea0-419a-4138-aac3-f4fd72f320a4 does not have any port associated with it.."} {"level":"info","ts":1582047704.9578178,"logger":"baremetalhost","msg":"stopping on host error","Request.Namespace":"metal3","Request.Name":"baremetal0012d1mdw1.sendgrid.net","provisioningState":"provisioned","message":"Host adoption failed: Error while attempting to adopt node 60895ea0-419a-4138-aac3-f4fd72f320a4: Node 60895ea0-419a-4138-aac3-f4fd72f320a4 does not have any port associated with it.."} ```
1.0
On bmo pod restart unable to provision ready state BMH's - There appears to be a bug where after creating a number of bmh that go through the registering -> inspecting -> ready phases, after which the baremetal operator pod is restarted, you are no longer able to provision any kube machines on these bmhs, and the following error is thrown: `host validation error: Node 3a0e38e3-6d5f-424e-b4cb-cb4758506da5 does not have any port associated with it.; Node 3a0e38e3-6d5f-424e-b4cb-cb4758506da5 does not have any port associated with it.` Here's the full output and and logs from bmo: ``` kubectl get bmh ξ‚² βœ” ξ‚² 11342 ξ‚² 11:06:45 NAME STATUS PROVISIONING STATUS CONSUMER BMC HARDWARE PROFILE ONLINE ERROR baremetal0003d1mdw1.sendgrid.net OK inspecting ipmi://10.16.5.211/ true baremetal0004d1mdw1.sendgrid.net OK ready ipmi://10.16.6.14/ unknown true baremetal0005d1mdw1.sendgrid.net OK ready ipmi://10.16.5.227/ unknown true baremetal0006d1mdw1.sendgrid.net OK ready ipmi://10.16.5.241/ unknown true baremetal0007d1mdw1.sendgrid.net OK ready ipmi://10.16.5.239/ unknown true baremetal0008d1mdw1.sendgrid.net OK ready ipmi://10.16.5.235/ unknown true baremetal0009d1mdw1.sendgrid.net OK ready ipmi://10.16.5.207/ unknown true baremetal0010d1mdw1.sendgrid.net OK ready ipmi://10.16.5.170/ unknown true baremetal0011d1mdw1.sendgrid.net OK ready ipmi://10.16.5.179/ unknown true baremetal0012d1mdw1.sendgrid.net OK ready ipmi://10.16.5.237/ unknown true baremetal0013d1mdw1.sendgrid.net OK ready ipmi://10.16.6.18/ unknown true baremetal0014d1mdw1.sendgrid.net OK ready ipmi://10.16.6.60/ unknown true baremetal0015d1mdw1.sendgrid.net OK ready ipmi://10.16.5.185/ unknown true baremetal0016d1mdw1.sendgrid.net OK ready ipmi://10.16.5.164/ unknown true baremetal0017d1mdw1.sendgrid.net OK ready ipmi://10.16.5.183/ unknown true baremetal0018d1mdw1.sendgrid.net OK ready ipmi://10.16.5.242/ unknown true baremetal0019d1mdw1.sendgrid.net OK ready ipmi://10.16.5.174/ unknown true lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° ξ‚² βœ” ξ‚² 11343 ξ‚² 11:06:46 lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° kubectl apply -f _out/ ξ‚² 1 ↡ ξ‚² 11344 ξ‚² 11:06:54 cluster.cluster.x-k8s.io/lukasz unchanged baremetalcluster.infrastructure.cluster.x-k8s.io/lukasz unchanged kubeadmconfig.bootstrap.cluster.x-k8s.io/lukasz-controlplane-0 created kubeadmconfig.bootstrap.cluster.x-k8s.io/lukasz-controlplane-1 created kubeadmconfig.bootstrap.cluster.x-k8s.io/lukasz-controlplane-2 created machine.cluster.x-k8s.io/lukasz-controlplane-0 created machine.cluster.x-k8s.io/lukasz-controlplane-1 created machine.cluster.x-k8s.io/lukasz-controlplane-2 created baremetalmachine.infrastructure.cluster.x-k8s.io/lukasz-controlplane-0 created baremetalmachine.infrastructure.cluster.x-k8s.io/lukasz-controlplane-1 created baremetalmachine.infrastructure.cluster.x-k8s.io/lukasz-controlplane-2 created kubeadmconfigtemplate.bootstrap.cluster.x-k8s.io/lukasz-md-0 created kubeadmconfigtemplate.bootstrap.cluster.x-k8s.io/lukasz-old-md-0 created machinedeployment.cluster.x-k8s.io/lukasz-md-0 created machinedeployment.cluster.x-k8s.io/lukasz-old-md-0 created baremetalmachinetemplate.infrastructure.cluster.x-k8s.io/lukasz-md-0 created baremetalmachinetemplate.infrastructure.cluster.x-k8s.io/lukasz-old-md-0 created lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° ξ‚² βœ” ξ‚² 11345 ξ‚² 11:06:59 lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° kubectl get machine ξ‚² βœ” ξ‚² 11345 ξ‚² 11:06:59 NAME PROVIDERID PHASE lukasz-controlplane-0 provisioning lukasz-controlplane-1 pending lukasz-controlplane-2 pending lukasz-md-0-cf48d76f6-jjrv7 pending lukasz-old-md-0-5d997d4449-r4s5r pending lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° kubectl get bmh ξ‚² βœ” ξ‚² 11346 ξ‚² 11:07:07 NAME STATUS PROVISIONING STATUS CONSUMER BMC HARDWARE PROFILE ONLINE ERROR baremetal0003d1mdw1.sendgrid.net OK inspecting ipmi://10.16.5.211/ true baremetal0004d1mdw1.sendgrid.net OK ready ipmi://10.16.6.14/ unknown true baremetal0005d1mdw1.sendgrid.net OK ready ipmi://10.16.5.227/ unknown true baremetal0006d1mdw1.sendgrid.net OK ready ipmi://10.16.5.241/ unknown true baremetal0007d1mdw1.sendgrid.net OK ready ipmi://10.16.5.239/ unknown true baremetal0008d1mdw1.sendgrid.net OK ready ipmi://10.16.5.235/ unknown true baremetal0009d1mdw1.sendgrid.net OK ready ipmi://10.16.5.207/ unknown true baremetal0010d1mdw1.sendgrid.net OK ready ipmi://10.16.5.170/ unknown true baremetal0011d1mdw1.sendgrid.net OK ready ipmi://10.16.5.179/ unknown true baremetal0012d1mdw1.sendgrid.net OK ready ipmi://10.16.5.237/ unknown true baremetal0013d1mdw1.sendgrid.net OK ready ipmi://10.16.6.18/ unknown true baremetal0014d1mdw1.sendgrid.net OK ready ipmi://10.16.6.60/ unknown true baremetal0015d1mdw1.sendgrid.net error provisioning lukasz-controlplane-0 ipmi://10.16.5.185/ unknown true host validation error: Node 3a0e38e3-6d5f-424e-b4cb-cb4758506da5 does not have any port associated with it.; Node 3a0e38e3-6d5f-424e-b4cb-cb4758506da5 does not have any port associated with it. baremetal0016d1mdw1.sendgrid.net OK ready ipmi://10.16.5.164/ unknown true baremetal0017d1mdw1.sendgrid.net OK ready ipmi://10.16.5.183/ unknown true baremetal0018d1mdw1.sendgrid.net OK ready ipmi://10.16.5.242/ unknown true baremetal0019d1mdw1.sendgrid.net OK ready ipmi://10.16.5.174/ unknown true lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° ξ‚² βœ” ξ‚² 11347 ξ‚² 11:07:20 lukasz@lukasz ξ‚° ~/go/src/github.com/sendgrid/cluster-api-provider-bmo/armada ξ‚° ξ‚  lukaszbranch ● ? ξ‚° {"level":"info","ts":1581703726.561492,"logger":"baremetalhost","msg":"Reconciling BareMetalHost","Request.Namespace":"metal3","Request.Name":"baremetal0003d1mdw1.sendgrid.net"} {"level":"info","ts":1581703726.5615559,"logger":"baremetalhost","msg":"inspecting hardware","Request.Namespace":"metal3","Request.Name":"baremetal0003d1mdw1.sendgrid.net","provisioningState":"inspecting"} {"level":"info","ts":1581703726.5615826,"logger":"baremetalhost_ironic","msg":"inspecting hardware","host":"baremetal0003d1mdw1.sendgrid.net","status":"OK"} {"level":"info","ts":1581703726.5742512,"logger":"baremetalhost_ironic","msg":"looking for existing node by name","host":"baremetal0003d1mdw1.sendgrid.net","name":"baremetal0003d1mdw1.sendgrid.net"} {"level":"error","ts":1581703726.5862474,"logger":"controller-runtime.controller","msg":"Reconciler error","controller":"metal3-baremetalhost-controller","request":"metal3/baremetal0003d1mdw1.sendgrid.net","error":"action \"inspecting\" failed: hardware inspection failed: no ironic node for host","errorVerbose":"no ironic node for host\nhardware inspection failed\ngithub.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).actionInspecting\n\t/go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:491\ngithub.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).handleInspecting\n\t/go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:197\ngithub.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).(github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.handleInspecting)-fm\n\t/go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:41\ngithub.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).ReconcileState\n\t/go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:109\ngithub.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).Reconcile\n\t/go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:281\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:256\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:232\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:211\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).(github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.worker)-fm\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:2361\naction \"inspecting\" failed\ngithub.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).Reconcile\n\t/go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:285\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:256\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:232\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:211\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).(github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.worker)-fm\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:2361","stacktrace":"github.com/metal3-io/baremetal-operator/vendor/github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/github.com/go-logr/zapr/zapr.go:128\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:258\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:232\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:211\ngithub.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).(github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.worker)-fm\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153\ngithub.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} {"level":"info","ts":1581703727.5868793,"logger":"baremetalhost","msg":"Reconciling BareMetalHost","Request.Namespace":"metal3","Request.Name":"baremetal0004d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.5870068,"logger":"baremetalhost_ironic","msg":"validating management access","host":"baremetal0004d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.6093876,"logger":"baremetalhost_ironic","msg":"found existing node by ID","host":"baremetal0004d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.6094184,"logger":"baremetalhost_ironic","msg":"current provision state","host":"baremetal0004d1mdw1.sendgrid.net","lastError":"","current":"manageable","target":""} {"level":"info","ts":1581703727.60943,"logger":"baremetalhost_ironic","msg":"have manageable host","host":"baremetal0004d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.609438,"logger":"baremetalhost_ironic","msg":"updating hardware state","host":"baremetal0004d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.6298566,"logger":"baremetalhost_ironic","msg":"found existing node by ID","host":"baremetal0004d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.6298933,"logger":"baremetalhost","msg":"saving host status","Request.Namespace":"metal3","Request.Name":"baremetal0004d1mdw1.sendgrid.net","provisioningState":"ready","operational status":"OK","provisioning state":"ready"} {"level":"info","ts":1581703727.6373432,"logger":"baremetalhost","msg":"done","Request.Namespace":"metal3","Request.Name":"baremetal0004d1mdw1.sendgrid.net","provisioningState":"ready","requeue":true,"after":60} {"level":"info","ts":1581703727.637434,"logger":"baremetalhost","msg":"Reconciling BareMetalHost","Request.Namespace":"metal3","Request.Name":"baremetal0015d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.6375093,"logger":"baremetalhost","msg":"provisioning","Request.Namespace":"metal3","Request.Name":"baremetal0015d1mdw1.sendgrid.net","provisioningState":"provisioning"} {"level":"info","ts":1581703727.6551394,"logger":"baremetalhost_ironic","msg":"found existing node by ID","host":"baremetal0015d1mdw1.sendgrid.net"} {"level":"info","ts":1581703727.6551712,"logger":"baremetalhost_ironic","msg":"provisioning image to host","host":"baremetal0015d1mdw1.sendgrid.net","state":"manageable"} E0214 18:08:47.655283 1 runtime.go:78] Observed a panic: "invalid memory address or nil pointer dereference" (runtime error: invalid memory address or nil pointer dereference) goroutine 270 [running]: github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/runtime.logPanic(0x137efe0, 0x20309c0) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:74 +0xaa github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/runtime.HandleCrash(0x0, 0x0, 0x0) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:48 +0x82 panic(0x137efe0, 0x20309c0) /usr/local/go/src/runtime/panic.go:502 +0x229 github.com/metal3-io/baremetal-operator/pkg/provisioner/ironic.(*ironicProvisioner).Provision(0xc420169030, 0x1649da0, 0xc4207c7a40, 0x0, 0x0, 0x0, 0xa858015299444fb4, 0x4122f8, 0x10) /go/src/github.com/metal3-io/baremetal-operator/pkg/provisioner/ironic/ironic.go:924 +0x156 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).actionProvisioning(0xc4208b9ac0, 0x1672860, 0xc420169030, 0xc4207f7400, 0x9e0000c420787260, 0xc420787240) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:563 +0x117 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).handleProvisioning(0xc4207c79e0, 0xc4207f7400, 0xc4207871f0, 0xc4200e1790) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:246 +0xb8 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).(github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.handleProvisioning)-fm(0xc4207f7400, 0xc4207c7a10, 0xc42047ac90) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:45 +0x34 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).ReconcileState(0xc4207c79e0, 0xc4207f7400, 0x0, 0x0) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:109 +0x327 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).Reconcile(0xc4208b9ac0, 0xc42047a8da, 0x6, 0xc4206b17e0, 0x20, 0x204ac00, 0x0, 0x0, 0x0) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:281 +0x987 github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler(0xc4205623c0, 0x13c9ea0, 0xc420916ee0, 0x13c9e00) /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:256 +0x100 github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem(0xc4205623c0, 0xc4206be300) /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:232 +0xb7 github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker(0xc4205623c0) /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:211 +0x2b github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).(github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.worker)-fm() /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193 +0x2a github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1(0xc42065a490) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152 +0x54 github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil(0xc42065a490, 0x3b9aca00, 0x0, 0x1, 0xc420023a40) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153 +0xbd github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.Until(0xc42065a490, 0x3b9aca00, 0xc420023a40) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88 +0x4d created by github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1 /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193 +0x63b panic: runtime error: invalid memory address or nil pointer dereference [recovered] panic: runtime error: invalid memory address or nil pointer dereference [signal SIGSEGV: segmentation violation code=0x1 addr=0x18 pc=0xb40e76] goroutine 270 [running]: github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/runtime.HandleCrash(0x0, 0x0, 0x0) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go:55 +0x107 panic(0x137efe0, 0x20309c0) /usr/local/go/src/runtime/panic.go:502 +0x229 github.com/metal3-io/baremetal-operator/pkg/provisioner/ironic.(*ironicProvisioner).Provision(0xc420169030, 0x1649da0, 0xc4207c7a40, 0x0, 0x0, 0x0, 0xa858015299444fb4, 0x4122f8, 0x10) /go/src/github.com/metal3-io/baremetal-operator/pkg/provisioner/ironic/ironic.go:924 +0x156 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).actionProvisioning(0xc4208b9ac0, 0x1672860, 0xc420169030, 0xc4207f7400, 0x9e0000c420787260, 0xc420787240) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:563 +0x117 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).handleProvisioning(0xc4207c79e0, 0xc4207f7400, 0xc4207871f0, 0xc4200e1790) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:246 +0xb8 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).(github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.handleProvisioning)-fm(0xc4207f7400, 0xc4207c7a10, 0xc42047ac90) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:45 +0x34 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*hostStateMachine).ReconcileState(0xc4207c79e0, 0xc4207f7400, 0x0, 0x0) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/host_state_machine.go:109 +0x327 github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost.(*ReconcileBareMetalHost).Reconcile(0xc4208b9ac0, 0xc42047a8da, 0x6, 0xc4206b17e0, 0x20, 0x204ac00, 0x0, 0x0, 0x0) /go/src/github.com/metal3-io/baremetal-operator/pkg/controller/baremetalhost/baremetalhost_controller.go:281 +0x987 github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler(0xc4205623c0, 0x13c9ea0, 0xc420916ee0, 0x13c9e00) /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:256 +0x100 github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem(0xc4205623c0, 0xc4206be300) /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:232 +0xb7 github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker(0xc4205623c0) /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:211 +0x2b github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).(github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.worker)-fm() /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193 +0x2a github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1(0xc42065a490) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152 +0x54 github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil(0xc42065a490, 0x3b9aca00, 0x0, 0x1, 0xc420023a40) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153 +0xbd github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait.Until(0xc42065a490, 0x3b9aca00, 0xc420023a40) /go/src/github.com/metal3-io/baremetal-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88 +0x4d created by github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1 /go/src/github.com/metal3-io/baremetal-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:193 +0x63b ``` Steps to recreate: Use bmo as normal and get some bmh into the ready state. Bounce the pod. After the operator comes back up attempt to provision a new cluster the above error should be thrown. We have also encountered a similar error, although haven't been able to reproduce it 100% of the time as the one above, where a bmh previous to the pod bounce was in the provisioned state, in the example provided we provisioned the bmh by adding the image to the spec it wasn't provisioned as part of a cluster, but we have seen the exact same issue below when the bmhs were provisioned as part of a cluster, after the pod is bounced do not get adopted properly, i.e.: ``` baremetal0012d1mdw1.sendgrid.net error registration error ipmi://10.16.5.237/ unknown true Host adoption failed: Error while attempting to adopt node 60895ea0-419a-4138-aac3-f4fd72f320a4: Node 60895ea0-419a-4138-aac3-f4fd72f320a4 does not have any port associated with it.. ``` And the logs for the event: ``` {"level":"info","ts":1582047700.842536,"logger":"baremetalhost","msg":"Reconciling BareMetalHost","Request.Namespace":"metal3","Request.Name":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047700.856129,"logger":"baremetalhost_ironic","msg":"looking for existing node by name","host":"baremetal0012d1mdw1.sendgrid.net","name":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047700.8713014,"logger":"baremetalhost_ironic","msg":"re-registering host","host":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047700.8713233,"logger":"baremetalhost_ironic","msg":"validating management access","host":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047700.886314,"logger":"baremetalhost_ironic","msg":"looking for existing node by name","host":"baremetal0012d1mdw1.sendgrid.net","name":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047700.8993719,"logger":"baremetalhost_ironic","msg":"registering host in ironic","host":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047700.9414284,"logger":"baremetalhost_ironic","msg":"setting provisioning id","host":"baremetal0012d1mdw1.sendgrid.net","ID":"60895ea0-419a-4138-aac3-f4fd72f320a4"} {"level":"info","ts":1582047700.941458,"logger":"baremetalhost_ironic","msg":"setting instance info","host":"baremetal0012d1mdw1.sendgrid.net","image_source":"https://filestore-staging.sendgrid.net/met al3-image-build/fedora-31-metal3-0.0.20200213231225.qcow2","checksum":"https://filestore-staging.sendgrid.net/metal3-image-build/fedora-31-metal3-0.0.20200213231225.qcow2.md5sum"} {"level":"info","ts":1582047701.0355387,"logger":"baremetalhost_ironic","msg":"current provision state","host":"baremetal0012d1mdw1.sendgrid.net","lastError":"","current":"enroll","target":""} {"level":"info","ts":1582047701.0355792,"logger":"baremetalhost_ironic","msg":"changing provisioning state","host":"baremetal0012d1mdw1.sendgrid.net","current":"enroll","existing target":"","new target ":"manage"} {"level":"info","ts":1582047701.1334321,"logger":"baremetalhost","msg":"saving host status","Request.Namespace":"metal3","Request.Name":"baremetal0012d1mdw1.sendgrid.net","provisioningState":"provision ed","operational status":"OK","provisioning state":"provisioned"} {"level":"info","ts":1582047701.1406105,"logger":"baremetalhost","msg":"publishing event","reason":"Registered","message":"Registered new host"} {"level":"info","ts":1582047701.144376,"logger":"baremetalhost","msg":"done","Request.Namespace":"metal3","Request.Name":"baremetal0012d1mdw1.sendgrid.net","provisioningState":"provisioned","requeue":t rue,"after":10} A little while later {"level":"info","ts":1582047704.92819,"logger":"baremetalhost","msg":"Reconciling BareMetalHost","Request.Namespace":"metal3","Request.Name":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047704.9464087,"logger":"baremetalhost_ironic","msg":"found existing node by ID","host":"baremetal0012d1mdw1.sendgrid.net"} {"level":"info","ts":1582047704.9464579,"logger":"baremetalhost","msg":"saving host status","Request.Namespace":"metal3","Request.Name":"baremetal0012d1mdw1.sendgrid.net","provisioningState":"provisioned","operational status":"error","provisioning state":"registration error"} {"level":"info","ts":1582047704.9539585,"logger":"baremetalhost","msg":"publishing event","reason":"RegistrationError","message":"Host adoption failed: Error while attempting to adopt node 60895ea0-419a-4138-aac3-f4fd72f320a4: Node 60895ea0-419a-4138-aac3-f4fd72f320a4 does not have any port associated with it.."} {"level":"info","ts":1582047704.9578178,"logger":"baremetalhost","msg":"stopping on host error","Request.Namespace":"metal3","Request.Name":"baremetal0012d1mdw1.sendgrid.net","provisioningState":"provisioned","message":"Host adoption failed: Error while attempting to adopt node 60895ea0-419a-4138-aac3-f4fd72f320a4: Node 60895ea0-419a-4138-aac3-f4fd72f320a4 does not have any port associated with it.."} ```
non_main
on bmo pod restart unable to provision ready state bmh s there appears to be a bug where after creating a number of bmh that go through the registering inspecting ready phases after which the baremetal operator pod is restarted you are no longer able to provision any kube machines on these bmhs and the following error is thrown host validation error node does not have any port associated with it node does not have any port associated with it here s the full output and and logs from bmo kubectl get bmh ξ‚² βœ” ξ‚² ξ‚² name status provisioning status consumer bmc hardware profile online error sendgrid net ok inspecting ipmi true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true lukasz lukasz ξ‚° go src github com sendgrid cluster api provider bmo armada ξ‚° ξ‚  lukaszbranch ● ξ‚° ξ‚² βœ” ξ‚² ξ‚² lukasz lukasz ξ‚° go src github com sendgrid cluster api provider bmo armada ξ‚° ξ‚  lukaszbranch ● ξ‚° kubectl apply f out ξ‚² ↡ ξ‚² ξ‚² cluster cluster x io lukasz unchanged baremetalcluster infrastructure cluster x io lukasz unchanged kubeadmconfig bootstrap cluster x io lukasz controlplane created kubeadmconfig bootstrap cluster x io lukasz controlplane created kubeadmconfig bootstrap cluster x io lukasz controlplane created machine cluster x io lukasz controlplane created machine cluster x io lukasz controlplane created machine cluster x io lukasz controlplane created baremetalmachine infrastructure cluster x io lukasz controlplane created baremetalmachine infrastructure cluster x io lukasz controlplane created baremetalmachine infrastructure cluster x io lukasz controlplane created kubeadmconfigtemplate bootstrap cluster x io lukasz md created kubeadmconfigtemplate bootstrap cluster x io lukasz old md created machinedeployment cluster x io lukasz md created machinedeployment cluster x io lukasz old md created baremetalmachinetemplate infrastructure cluster x io lukasz md created baremetalmachinetemplate infrastructure cluster x io lukasz old md created lukasz lukasz ξ‚° go src github com sendgrid cluster api provider bmo armada ξ‚° ξ‚  lukaszbranch ● ξ‚° ξ‚² βœ” ξ‚² ξ‚² lukasz lukasz ξ‚° go src github com sendgrid cluster api provider bmo armada ξ‚° ξ‚  lukaszbranch ● ξ‚° kubectl get machine ξ‚² βœ” ξ‚² ξ‚² name providerid phase lukasz controlplane provisioning lukasz controlplane pending lukasz controlplane pending lukasz md pending lukasz old md pending lukasz lukasz ξ‚° go src github com sendgrid cluster api provider bmo armada ξ‚° ξ‚  lukaszbranch ● ξ‚° kubectl get bmh ξ‚² βœ” ξ‚² ξ‚² name status provisioning status consumer bmc hardware profile online error sendgrid net ok inspecting ipmi true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net error provisioning lukasz controlplane ipmi unknown true host validation error node does not have any port associated with it node does not have any port associated with it sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true sendgrid net ok ready ipmi unknown true lukasz lukasz ξ‚° go src github com sendgrid cluster api provider bmo armada ξ‚° ξ‚  lukaszbranch ● ξ‚° ξ‚² βœ” ξ‚² ξ‚² lukasz lukasz ξ‚° go src github com sendgrid cluster api provider bmo armada ξ‚° ξ‚  lukaszbranch ● ξ‚° level info ts logger baremetalhost msg reconciling baremetalhost request namespace request name sendgrid net level info ts logger baremetalhost msg inspecting hardware request namespace request name sendgrid net provisioningstate inspecting level info ts logger baremetalhost ironic msg inspecting hardware host sendgrid net status ok level info ts logger baremetalhost ironic msg looking for existing node by name host sendgrid net name sendgrid net level error ts logger controller runtime controller msg reconciler error controller baremetalhost controller request sendgrid net error action inspecting failed hardware inspection failed no ironic node for host errorverbose no ironic node for host nhardware inspection failed ngithub com io baremetal operator pkg controller baremetalhost reconcilebaremetalhost actioninspecting n t go src github com io baremetal operator pkg controller baremetalhost baremetalhost controller go ngithub com io baremetal operator pkg controller baremetalhost hoststatemachine handleinspecting n t go src github com io baremetal operator pkg controller baremetalhost host state machine go ngithub com io baremetal operator pkg controller baremetalhost hoststatemachine github com io baremetal operator pkg controller baremetalhost handleinspecting fm n t go src github com io baremetal operator pkg controller baremetalhost host state machine go ngithub com io baremetal operator pkg controller baremetalhost hoststatemachine reconcilestate n t go src github com io baremetal operator pkg controller baremetalhost host state machine go ngithub com io baremetal operator pkg controller baremetalhost reconcilebaremetalhost reconcile n t go src github com io baremetal operator pkg controller baremetalhost baremetalhost controller go ngithub com io baremetal operator vendor sigs io controller runtime pkg internal controller controller reconcilehandler n t go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go ngithub com io baremetal operator vendor sigs io controller runtime pkg internal controller controller processnextworkitem n t go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go ngithub com io baremetal operator vendor sigs io controller runtime pkg internal controller controller worker n t go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go ngithub com io baremetal operator vendor sigs io controller runtime pkg internal controller controller github com io baremetal operator vendor sigs io controller runtime pkg internal controller worker fm n t go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go ngithub com io baremetal operator vendor io apimachinery pkg util wait jitteruntil n t go src github com io baremetal operator vendor io apimachinery pkg util wait wait go ngithub com io baremetal operator vendor io apimachinery pkg util wait jitteruntil n t go src github com io baremetal operator vendor io apimachinery pkg util wait wait go ngithub com io baremetal operator vendor io apimachinery pkg util wait until n t go src github com io baremetal operator vendor io apimachinery pkg util wait wait go nruntime goexit n t usr local go src runtime asm s naction inspecting failed ngithub com io baremetal operator pkg controller baremetalhost reconcilebaremetalhost reconcile n t go src github com io baremetal operator pkg controller baremetalhost baremetalhost controller go ngithub com io baremetal operator vendor sigs io controller runtime pkg internal controller controller reconcilehandler n t go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go ngithub com io baremetal operator vendor sigs io controller runtime pkg internal controller controller processnextworkitem n t go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go ngithub com io baremetal operator vendor sigs io controller runtime pkg internal controller controller worker n t go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go ngithub com io baremetal operator vendor sigs io controller runtime pkg internal controller controller github com io baremetal operator vendor sigs io controller runtime pkg internal controller worker fm n t go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go ngithub com io baremetal operator vendor io apimachinery pkg util wait jitteruntil n t go src github com io baremetal operator vendor io apimachinery pkg util wait wait go ngithub com io baremetal operator vendor io apimachinery pkg util wait jitteruntil n t go src github com io baremetal operator vendor io apimachinery pkg util wait wait go ngithub com io baremetal operator vendor io apimachinery pkg util wait until n t go src github com io baremetal operator vendor io apimachinery pkg util wait wait go nruntime goexit n t usr local go src runtime asm s stacktrace github com io baremetal operator vendor github com go logr zapr zaplogger error n t go src github com io baremetal operator vendor github com go logr zapr zapr go ngithub com io baremetal operator vendor sigs io controller runtime pkg internal controller controller reconcilehandler n t go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go ngithub com io baremetal operator vendor sigs io controller runtime pkg internal controller controller processnextworkitem n t go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go ngithub com io baremetal operator vendor sigs io controller runtime pkg internal controller controller worker n t go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go ngithub com io baremetal operator vendor sigs io controller runtime pkg internal controller controller github com io baremetal operator vendor sigs io controller runtime pkg internal controller worker fm n t go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go ngithub com io baremetal operator vendor io apimachinery pkg util wait jitteruntil n t go src github com io baremetal operator vendor io apimachinery pkg util wait wait go ngithub com io baremetal operator vendor io apimachinery pkg util wait jitteruntil n t go src github com io baremetal operator vendor io apimachinery pkg util wait wait go ngithub com io baremetal operator vendor io apimachinery pkg util wait until n t go src github com io baremetal operator vendor io apimachinery pkg util wait wait go level info ts logger baremetalhost msg reconciling baremetalhost request namespace request name sendgrid net level info ts logger baremetalhost ironic msg validating management access host sendgrid net level info ts logger baremetalhost ironic msg found existing node by id host sendgrid net level info ts logger baremetalhost ironic msg current provision state host sendgrid net lasterror current manageable target level info ts logger baremetalhost ironic msg have manageable host host sendgrid net level info ts logger baremetalhost ironic msg updating hardware state host sendgrid net level info ts logger baremetalhost ironic msg found existing node by id host sendgrid net level info ts logger baremetalhost msg saving host status request namespace request name sendgrid net provisioningstate ready operational status ok provisioning state ready level info ts logger baremetalhost msg done request namespace request name sendgrid net provisioningstate ready requeue true after level info ts logger baremetalhost msg reconciling baremetalhost request namespace request name sendgrid net level info ts logger baremetalhost msg provisioning request namespace request name sendgrid net provisioningstate provisioning level info ts logger baremetalhost ironic msg found existing node by id host sendgrid net level info ts logger baremetalhost ironic msg provisioning image to host host sendgrid net state manageable runtime go observed a panic invalid memory address or nil pointer dereference runtime error invalid memory address or nil pointer dereference goroutine github com io baremetal operator vendor io apimachinery pkg util runtime logpanic go src github com io baremetal operator vendor io apimachinery pkg util runtime runtime go github com io baremetal operator vendor io apimachinery pkg util runtime handlecrash go src github com io baremetal operator vendor io apimachinery pkg util runtime runtime go panic usr local go src runtime panic go github com io baremetal operator pkg provisioner ironic ironicprovisioner provision go src github com io baremetal operator pkg provisioner ironic ironic go github com io baremetal operator pkg controller baremetalhost reconcilebaremetalhost actionprovisioning go src github com io baremetal operator pkg controller baremetalhost baremetalhost controller go github com io baremetal operator pkg controller baremetalhost hoststatemachine handleprovisioning go src github com io baremetal operator pkg controller baremetalhost host state machine go github com io baremetal operator pkg controller baremetalhost hoststatemachine github com io baremetal operator pkg controller baremetalhost handleprovisioning fm go src github com io baremetal operator pkg controller baremetalhost host state machine go github com io baremetal operator pkg controller baremetalhost hoststatemachine reconcilestate go src github com io baremetal operator pkg controller baremetalhost host state machine go github com io baremetal operator pkg controller baremetalhost reconcilebaremetalhost reconcile go src github com io baremetal operator pkg controller baremetalhost baremetalhost controller go github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller reconcilehandler go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller processnextworkitem go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller worker go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller github com io baremetal operator vendor sigs io controller runtime pkg internal controller worker fm go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go github com io baremetal operator vendor io apimachinery pkg util wait jitteruntil go src github com io baremetal operator vendor io apimachinery pkg util wait wait go github com io baremetal operator vendor io apimachinery pkg util wait jitteruntil go src github com io baremetal operator vendor io apimachinery pkg util wait wait go github com io baremetal operator vendor io apimachinery pkg util wait until go src github com io baremetal operator vendor io apimachinery pkg util wait wait go created by github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller start go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go panic runtime error invalid memory address or nil pointer dereference panic runtime error invalid memory address or nil pointer dereference goroutine github com io baremetal operator vendor io apimachinery pkg util runtime handlecrash go src github com io baremetal operator vendor io apimachinery pkg util runtime runtime go panic usr local go src runtime panic go github com io baremetal operator pkg provisioner ironic ironicprovisioner provision go src github com io baremetal operator pkg provisioner ironic ironic go github com io baremetal operator pkg controller baremetalhost reconcilebaremetalhost actionprovisioning go src github com io baremetal operator pkg controller baremetalhost baremetalhost controller go github com io baremetal operator pkg controller baremetalhost hoststatemachine handleprovisioning go src github com io baremetal operator pkg controller baremetalhost host state machine go github com io baremetal operator pkg controller baremetalhost hoststatemachine github com io baremetal operator pkg controller baremetalhost handleprovisioning fm go src github com io baremetal operator pkg controller baremetalhost host state machine go github com io baremetal operator pkg controller baremetalhost hoststatemachine reconcilestate go src github com io baremetal operator pkg controller baremetalhost host state machine go github com io baremetal operator pkg controller baremetalhost reconcilebaremetalhost reconcile go src github com io baremetal operator pkg controller baremetalhost baremetalhost controller go github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller reconcilehandler go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller processnextworkitem go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller worker go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller github com io baremetal operator vendor sigs io controller runtime pkg internal controller worker fm go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go github com io baremetal operator vendor io apimachinery pkg util wait jitteruntil go src github com io baremetal operator vendor io apimachinery pkg util wait wait go github com io baremetal operator vendor io apimachinery pkg util wait jitteruntil go src github com io baremetal operator vendor io apimachinery pkg util wait wait go github com io baremetal operator vendor io apimachinery pkg util wait until go src github com io baremetal operator vendor io apimachinery pkg util wait wait go created by github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller start go src github com io baremetal operator vendor sigs io controller runtime pkg internal controller controller go steps to recreate use bmo as normal and get some bmh into the ready state bounce the pod after the operator comes back up attempt to provision a new cluster the above error should be thrown we have also encountered a similar error although haven t been able to reproduce it of the time as the one above where a bmh previous to the pod bounce was in the provisioned state in the example provided we provisioned the bmh by adding the image to the spec it wasn t provisioned as part of a cluster but we have seen the exact same issue below when the bmhs were provisioned as part of a cluster after the pod is bounced do not get adopted properly i e sendgrid net error registration error ipmi unknown true host adoption failed error while attempting to adopt node node does not have any port associated with it and the logs for the event level info ts logger baremetalhost msg reconciling baremetalhost request namespace request name sendgrid net level info ts logger baremetalhost ironic msg looking for existing node by name host sendgrid net name sendgrid net level info ts logger baremetalhost ironic msg re registering host host sendgrid net level info ts logger baremetalhost ironic msg validating management access host sendgrid net level info ts logger baremetalhost ironic msg looking for existing node by name host sendgrid net name sendgrid net level info ts logger baremetalhost ironic msg registering host in ironic host sendgrid net level info ts logger baremetalhost ironic msg setting provisioning id host sendgrid net id level info ts logger baremetalhost ironic msg setting instance info host sendgrid net image source image build fedora checksum level info ts logger baremetalhost ironic msg current provision state host sendgrid net lasterror current enroll target level info ts logger baremetalhost ironic msg changing provisioning state host sendgrid net current enroll existing target new target manage level info ts logger baremetalhost msg saving host status request namespace request name sendgrid net provisioningstate provision ed operational status ok provisioning state provisioned level info ts logger baremetalhost msg publishing event reason registered message registered new host level info ts logger baremetalhost msg done request namespace request name sendgrid net provisioningstate provisioned requeue t rue after a little while later level info ts logger baremetalhost msg reconciling baremetalhost request namespace request name sendgrid net level info ts logger baremetalhost ironic msg found existing node by id host sendgrid net level info ts logger baremetalhost msg saving host status request namespace request name sendgrid net provisioningstate provisioned operational status error provisioning state registration error level info ts logger baremetalhost msg publishing event reason registrationerror message host adoption failed error while attempting to adopt node node does not have any port associated with it level info ts logger baremetalhost msg stopping on host error request namespace request name sendgrid net provisioningstate provisioned message host adoption failed error while attempting to adopt node node does not have any port associated with it
0
5,748
30,450,460,999
IssuesEvent
2023-07-16 08:16:14
0ptim/JellyChat
https://api.github.com/repos/0ptim/JellyChat
opened
Embeddings from production to local
docs/maintainance area:general
To debug and develop locally, there needs to be a way we can load the embeddings from the production Supabase DB to the local running Supabase instance.
True
Embeddings from production to local - To debug and develop locally, there needs to be a way we can load the embeddings from the production Supabase DB to the local running Supabase instance.
main
embeddings from production to local to debug and develop locally there needs to be a way we can load the embeddings from the production supabase db to the local running supabase instance
1
128,218
17,464,494,228
IssuesEvent
2021-08-06 14:57:36
raft-tech/TANF-app
https://api.github.com/repos/raft-tech/TANF-app
closed
Demo user research + prototyping progress to regional offices
Research & Design triage
Description: Why: get regional data leads invested in the new TDRS process and introduce Raft to regional team "Regional staff wants to have a chance to understand the direction that we're moving in with the new system. They'd like to learn more about what we've learned so far. The data group has at least 1 person from all 10 regions, so may be a good place to present this. Some of those folks aren't the program managers so we may need to broaden."--Lauren, 5/11/2020 AC: - [ ] Sample acceptance criteria Tasks: - [x] Ask Lauren if RO staff may want to attend - [ ] Observe session where OHS team shares their research learning with regional folks - [ ] Schedule session time/day - [ ] Maybe include personas - [ ] Maybe doing Design Principles exercise Some other thoughts: - read out of research - talk about how we'll be coming back to them in the future - we could do a lightweight design principles workshop - participation in workshop may help identify ROs to talk to - This could also help demo how we want to involve other user groups at many levels to our vendor. Notes: - Schedule after round 3 research sysn theiss has been documented adn shred within OFA data team; Lauren has approved.
1.0
Demo user research + prototyping progress to regional offices - Description: Why: get regional data leads invested in the new TDRS process and introduce Raft to regional team "Regional staff wants to have a chance to understand the direction that we're moving in with the new system. They'd like to learn more about what we've learned so far. The data group has at least 1 person from all 10 regions, so may be a good place to present this. Some of those folks aren't the program managers so we may need to broaden."--Lauren, 5/11/2020 AC: - [ ] Sample acceptance criteria Tasks: - [x] Ask Lauren if RO staff may want to attend - [ ] Observe session where OHS team shares their research learning with regional folks - [ ] Schedule session time/day - [ ] Maybe include personas - [ ] Maybe doing Design Principles exercise Some other thoughts: - read out of research - talk about how we'll be coming back to them in the future - we could do a lightweight design principles workshop - participation in workshop may help identify ROs to talk to - This could also help demo how we want to involve other user groups at many levels to our vendor. Notes: - Schedule after round 3 research sysn theiss has been documented adn shred within OFA data team; Lauren has approved.
non_main
demo user research prototyping progress to regional offices description why get regional data leads invested in the new tdrs process and introduce raft to regional team regional staff wants to have a chance to understand the direction that we re moving in with the new system they d like to learn more about what we ve learned so far the data group has at least person from all regions so may be a good place to present this some of those folks aren t the program managers so we may need to broaden lauren ac sample acceptance criteria tasks ask lauren if ro staff may want to attend observe session where ohs team shares their research learning with regional folks schedule session time day maybe include personas maybe doing design principles exercise some other thoughts read out of research talk about how we ll be coming back to them in the future we could do a lightweight design principles workshop participation in workshop may help identify ros to talk to this could also help demo how we want to involve other user groups at many levels to our vendor notes schedule after round research sysn theiss has been documented adn shred within ofa data team lauren has approved
0
4,981
2,893,749,461
IssuesEvent
2015-06-15 19:37:00
ldionne/hana
https://api.github.com/repos/ldionne/hana
closed
[Design][Documentation]Β Concepts do not convert from one to the other
documentation enhancement
In http://ldionne.com/hana/structboost_1_1hana_1_1Foldable.html it is written that any Foldable can be converted to a Sequence Do you mean that any instance/model of Foldable can be seen as an instance+model of a Sequence? The function `to<Tuple>` e.g. is a function from concrete models to a concrete model. Do we really need to make this conversion to be able to see a Foldable as a Sequence? I suggest to reword this kind of descriptions.
1.0
[Design][Documentation]Β Concepts do not convert from one to the other - In http://ldionne.com/hana/structboost_1_1hana_1_1Foldable.html it is written that any Foldable can be converted to a Sequence Do you mean that any instance/model of Foldable can be seen as an instance+model of a Sequence? The function `to<Tuple>` e.g. is a function from concrete models to a concrete model. Do we really need to make this conversion to be able to see a Foldable as a Sequence? I suggest to reword this kind of descriptions.
non_main
Β concepts do not convert from one to the other in it is written that any foldable can be converted to a sequence do you mean that any instance model of foldable can be seen as an instance model of a sequence the function to e g is a function from concrete models to a concrete model do we really need to make this conversion to be able to see a foldable as a sequence i suggest to reword this kind of descriptions
0
180,582
14,788,039,240
IssuesEvent
2021-01-12 08:37:53
rdsaliba/notorious-eng
https://api.github.com/repos/rdsaliba/notorious-eng
opened
UI - smoothering and improving UI
UI documentation enhancement
The PO made a few comments about the changes he would like to see happen in the UI. This is one of his priorities (even though it is not the highest). PO comments: - In general, the UI should be cleared. This means using more colors and more contrasts between the different UI elements so it's easier for the user to differentiate the elements. - In general, add more icons, graphics, background colors and images, and less text - Add an asset image on each thumbnail (image representing that asset) - Add transitions between UI views Acceptance Criteria: The first part of this issue is to create/edit mock-ups. The mock-ups should then be approved by the team (and maybe the PO) before moving on to editing the UI. - [ ] Mock-ups for the updated UI - [ ] Images of each asset on each thumbnail - [ ] Image of the asset in the asset info view - [ ] More background colors and contrasts - [ ] UI transitions between UI views - [ ] More colors, images, and graphics - [ ] Icons and images for buttons instead of text only.
1.0
UI - smoothering and improving UI - The PO made a few comments about the changes he would like to see happen in the UI. This is one of his priorities (even though it is not the highest). PO comments: - In general, the UI should be cleared. This means using more colors and more contrasts between the different UI elements so it's easier for the user to differentiate the elements. - In general, add more icons, graphics, background colors and images, and less text - Add an asset image on each thumbnail (image representing that asset) - Add transitions between UI views Acceptance Criteria: The first part of this issue is to create/edit mock-ups. The mock-ups should then be approved by the team (and maybe the PO) before moving on to editing the UI. - [ ] Mock-ups for the updated UI - [ ] Images of each asset on each thumbnail - [ ] Image of the asset in the asset info view - [ ] More background colors and contrasts - [ ] UI transitions between UI views - [ ] More colors, images, and graphics - [ ] Icons and images for buttons instead of text only.
non_main
ui smoothering and improving ui the po made a few comments about the changes he would like to see happen in the ui this is one of his priorities even though it is not the highest po comments in general the ui should be cleared this means using more colors and more contrasts between the different ui elements so it s easier for the user to differentiate the elements in general add more icons graphics background colors and images and less text add an asset image on each thumbnail image representing that asset add transitions between ui views acceptance criteria the first part of this issue is to create edit mock ups the mock ups should then be approved by the team and maybe the po before moving on to editing the ui mock ups for the updated ui images of each asset on each thumbnail image of the asset in the asset info view more background colors and contrasts ui transitions between ui views more colors images and graphics icons and images for buttons instead of text only
0
3,506
13,686,575,142
IssuesEvent
2020-09-30 08:53:52
darekkay/dashboard
https://api.github.com/repos/darekkay/dashboard
opened
Replace withErrorHandling with react-error-boundary
Type: Maintainance good first issue
The existing `withErrorHandling` should be replaced with [react-error-boundary](https://github.com/bvaughn/react-error-boundary), because it decouples the error state from the `widget` component (and provides some developer experience features).
True
Replace withErrorHandling with react-error-boundary - The existing `withErrorHandling` should be replaced with [react-error-boundary](https://github.com/bvaughn/react-error-boundary), because it decouples the error state from the `widget` component (and provides some developer experience features).
main
replace witherrorhandling with react error boundary the existing witherrorhandling should be replaced with because it decouples the error state from the widget component and provides some developer experience features
1
5,581
27,964,679,412
IssuesEvent
2023-03-24 18:23:38
BioArchLinux/Packages
https://api.github.com/repos/BioArchLinux/Packages
closed
[MAINTAIN] `bioarchlinux-keyring` has been _downgraded_. What is going on?, and increment `$epoch`.
maintain
I see that the package `bioarchlinux-keyring` from the Arch Linux repository "[bioarchlinux](https://wiki.archlinux.org/title/Unofficial_user_repositories#bioarchlinux) has been _downgraded_ in your repository!: ``` Package (1) Old Version New Version bioarchlinux/bioarchlinux-keyring 20220422-2 2022.04.16-1 ``` Can you please elaborate on that, what went wrong, and if you really need to replace with an older version, increment `$epoch` to ensure updates? Regards! **Packages (please complete the following information):** - Package Name: bioarchlinux-keyring
True
[MAINTAIN] `bioarchlinux-keyring` has been _downgraded_. What is going on?, and increment `$epoch`. - I see that the package `bioarchlinux-keyring` from the Arch Linux repository "[bioarchlinux](https://wiki.archlinux.org/title/Unofficial_user_repositories#bioarchlinux) has been _downgraded_ in your repository!: ``` Package (1) Old Version New Version bioarchlinux/bioarchlinux-keyring 20220422-2 2022.04.16-1 ``` Can you please elaborate on that, what went wrong, and if you really need to replace with an older version, increment `$epoch` to ensure updates? Regards! **Packages (please complete the following information):** - Package Name: bioarchlinux-keyring
main
bioarchlinux keyring has been downgraded what is going on and increment epoch i see that the package bioarchlinux keyring from the arch linux repository has been downgraded in your repository package old version new version bioarchlinux bioarchlinux keyring can you please elaborate on that what went wrong and if you really need to replace with an older version increment epoch to ensure updates regards packages please complete the following information package name bioarchlinux keyring
1
4,143
19,686,770,446
IssuesEvent
2022-01-11 23:21:54
truecharts/apps
https://api.github.com/repos/truecharts/apps
reopened
Add ApacheGuacamole
New App Request No-Maintainer
Home: [https://guacamole.apache.org/]() Official docker [https://registry.hub.docker.com/r/guacamole/guacamole]() needs also Postgre/MySQL 3rd Party docker [https://registry.hub.docker.com/r/jasonbean/guacamole]() embedded mysql.
True
Add ApacheGuacamole - Home: [https://guacamole.apache.org/]() Official docker [https://registry.hub.docker.com/r/guacamole/guacamole]() needs also Postgre/MySQL 3rd Party docker [https://registry.hub.docker.com/r/jasonbean/guacamole]() embedded mysql.
main
add apacheguacamole home official docker needs also postgre mysql party docker embedded mysql
1
50,138
13,187,343,730
IssuesEvent
2020-08-13 03:06:44
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
i3Monitoring ROOT output file names needs an additional identifier (Trac #187)
Migrated from Trac defect jeb + pnf
In current system, if JEB system is stopped and restarted, existing ROOT files will be overwritten. To avoid this, and to generate multiple ROOT files per client in this case, need to add a unique ID (in this case the PID will likely work well) to each filename, so that: EvtMon_PhysicsData_PhysicsFiltering_PFClient.sps-fpslave01.client1_Run00109709_Subrun00000000.root is EvtMon_PhysicsData_PhysicsFiltering_PFClient.sps-fpslave01.client1.PID_Run00109709_Subrun00000000.root where PID is the actual Process ID, which should be easily available to the process. <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/187 , reported by blaufuss and owned by blaufuss_</summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "In current system, if JEB system is stopped and restarted, existing ROOT files will be overwritten. To avoid this, and to generate multiple ROOT files\nper client in this case, need to add a unique ID (in this case the PID will\nlikely work well) to each filename, so that:\n\nEvtMon_PhysicsData_PhysicsFiltering_PFClient.sps-fpslave01.client1_Run00109709_Subrun00000000.root\n\nis \n\nEvtMon_PhysicsData_PhysicsFiltering_PFClient.sps-fpslave01.client1.PID_Run00109709_Subrun00000000.root\n\nwhere PID is the actual Process ID, which should be easily available to the process.", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "_ts": "1416713877066511", "component": "jeb + pnf", "summary": "i3Monitoring ROOT output file names needs an additional identifier", "priority": "normal", "keywords": "", "time": "2009-12-07T22:38:21", "milestone": "", "owner": "blaufuss", "type": "defect" } ``` </p> </details>
1.0
i3Monitoring ROOT output file names needs an additional identifier (Trac #187) - In current system, if JEB system is stopped and restarted, existing ROOT files will be overwritten. To avoid this, and to generate multiple ROOT files per client in this case, need to add a unique ID (in this case the PID will likely work well) to each filename, so that: EvtMon_PhysicsData_PhysicsFiltering_PFClient.sps-fpslave01.client1_Run00109709_Subrun00000000.root is EvtMon_PhysicsData_PhysicsFiltering_PFClient.sps-fpslave01.client1.PID_Run00109709_Subrun00000000.root where PID is the actual Process ID, which should be easily available to the process. <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/187 , reported by blaufuss and owned by blaufuss_</summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "In current system, if JEB system is stopped and restarted, existing ROOT files will be overwritten. To avoid this, and to generate multiple ROOT files\nper client in this case, need to add a unique ID (in this case the PID will\nlikely work well) to each filename, so that:\n\nEvtMon_PhysicsData_PhysicsFiltering_PFClient.sps-fpslave01.client1_Run00109709_Subrun00000000.root\n\nis \n\nEvtMon_PhysicsData_PhysicsFiltering_PFClient.sps-fpslave01.client1.PID_Run00109709_Subrun00000000.root\n\nwhere PID is the actual Process ID, which should be easily available to the process.", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "_ts": "1416713877066511", "component": "jeb + pnf", "summary": "i3Monitoring ROOT output file names needs an additional identifier", "priority": "normal", "keywords": "", "time": "2009-12-07T22:38:21", "milestone": "", "owner": "blaufuss", "type": "defect" } ``` </p> </details>
non_main
root output file names needs an additional identifier trac in current system if jeb system is stopped and restarted existing root files will be overwritten to avoid this and to generate multiple root files per client in this case need to add a unique id in this case the pid will likely work well to each filename so that evtmon physicsdata physicsfiltering pfclient sps root is evtmon physicsdata physicsfiltering pfclient sps pid root where pid is the actual process id which should be easily available to the process migrated from reported by blaufuss and owned by blaufuss json status closed changetime description in current system if jeb system is stopped and restarted existing root files will be overwritten to avoid this and to generate multiple root files nper client in this case need to add a unique id in this case the pid will nlikely work well to each filename so that n nevtmon physicsdata physicsfiltering pfclient sps root n nis n nevtmon physicsdata physicsfiltering pfclient sps pid root n nwhere pid is the actual process id which should be easily available to the process reporter blaufuss cc resolution fixed ts component jeb pnf summary root output file names needs an additional identifier priority normal keywords time milestone owner blaufuss type defect
0
4,842
24,960,337,139
IssuesEvent
2022-11-01 15:03:45
BioArchLinux/Packages
https://api.github.com/repos/BioArchLinux/Packages
opened
[MAINTAIN] java related packages polishment
maintain
<!-- Please report the error of one package in one issue! Use multi issues to report multi bugs. Thanks! --> **List of Pacakges** <details> - [ ] bbmap - [ ] beast - [ ] beast2 - [ ] canu - [ ] cytoscape - [ ] emboss - [ ] fastqc - [ ] figtree - [ ] gatk - [ ] insight-toolkit - [ ] jane - [ ] libsbml - [ ] mcscanx - [ ] mesquite - [ ] morphoj - [ ] open-delta - [ ] phyde - [ ] phylip - [ ] pibuss - [ ] phylonet - [ ] picard - [ ] sequencematrix - [ ] spread3 - [ ] spread - [ ] tempest - [ ] tracer - [ ] treemap - [ ] trimmomatic </details> **Description** These packages need to be checked the following points - [ ] use `java-runtime` and `java-enviroment` instead of jdk, jre, etc. - [ ] confirm the java version and limit the enviroment in PKGBUILD - [ ] check if they can be compiled from the source code [may be difficult, considering some software use quite low version java-enviroment] - [ ] check if PKGBUILD style can be auto updated
True
[MAINTAIN] java related packages polishment - <!-- Please report the error of one package in one issue! Use multi issues to report multi bugs. Thanks! --> **List of Pacakges** <details> - [ ] bbmap - [ ] beast - [ ] beast2 - [ ] canu - [ ] cytoscape - [ ] emboss - [ ] fastqc - [ ] figtree - [ ] gatk - [ ] insight-toolkit - [ ] jane - [ ] libsbml - [ ] mcscanx - [ ] mesquite - [ ] morphoj - [ ] open-delta - [ ] phyde - [ ] phylip - [ ] pibuss - [ ] phylonet - [ ] picard - [ ] sequencematrix - [ ] spread3 - [ ] spread - [ ] tempest - [ ] tracer - [ ] treemap - [ ] trimmomatic </details> **Description** These packages need to be checked the following points - [ ] use `java-runtime` and `java-enviroment` instead of jdk, jre, etc. - [ ] confirm the java version and limit the enviroment in PKGBUILD - [ ] check if they can be compiled from the source code [may be difficult, considering some software use quite low version java-enviroment] - [ ] check if PKGBUILD style can be auto updated
main
java related packages polishment please report the error of one package in one issue use multi issues to report multi bugs thanks list of pacakges bbmap beast canu cytoscape emboss fastqc figtree gatk insight toolkit jane libsbml mcscanx mesquite morphoj open delta phyde phylip pibuss phylonet picard sequencematrix spread tempest tracer treemap trimmomatic description these packages need to be checked the following points use java runtime and java enviroment instead of jdk jre etc confirm the java version and limit the enviroment in pkgbuild check if they can be compiled from the source code check if pkgbuild style can be auto updated
1
3,817
16,602,944,794
IssuesEvent
2021-06-01 22:18:37
jesus2099/konami-command
https://api.github.com/repos/jesus2099/konami-command
opened
Use regex includes
maintainability mb_ALL-LINKS ninja
β†– #488 – y-van-z/yvanzo’s All shook up links --- Refactor match and exclude, use regex include, instead. Because the following is making the script fail on [other URL](https://musicbrainz.org/artist/88d8f38f-adb4-48a0-8c1f-ec34f2a675ff/recordings?filter.artist_credit_id=&filter.name=%E6%9C%88%E5%A7%AB) than intended. And remembering what was intended is not even easy with such loose patterns. ```js // @exclude *.org/*/*edit* ```
True
Use regex includes - β†– #488 – y-van-z/yvanzo’s All shook up links --- Refactor match and exclude, use regex include, instead. Because the following is making the script fail on [other URL](https://musicbrainz.org/artist/88d8f38f-adb4-48a0-8c1f-ec34f2a675ff/recordings?filter.artist_credit_id=&filter.name=%E6%9C%88%E5%A7%AB) than intended. And remembering what was intended is not even easy with such loose patterns. ```js // @exclude *.org/*/*edit* ```
main
use regex includes β†– – y van z yvanzo’s all shook up links refactor match and exclude use regex include instead because the following is making the script fail on than intended and remembering what was intended is not even easy with such loose patterns js exclude org edit
1
80,925
15,604,981,454
IssuesEvent
2021-03-19 05:10:29
nhcarrigan/we-love-hacktoberfest
https://api.github.com/repos/nhcarrigan/we-love-hacktoberfest
closed
[OTHER] - Refactor codebase, implement HTTP endpoint
✨ goal: improvement 🏁 status: ready for dev πŸ’» aspect: code πŸ”’ staff only πŸŸ₯ priority: critical
# Other Issue ## Describe the issue Codebase needs refactored and cleaned up Need to implement an HTTP endpoint as well for uptime monitoring <!--A clear and concise description of the issue/concern--> ## Additional information <!--Add any other context about the problem here.-->
1.0
[OTHER] - Refactor codebase, implement HTTP endpoint - # Other Issue ## Describe the issue Codebase needs refactored and cleaned up Need to implement an HTTP endpoint as well for uptime monitoring <!--A clear and concise description of the issue/concern--> ## Additional information <!--Add any other context about the problem here.-->
non_main
refactor codebase implement http endpoint other issue describe the issue codebase needs refactored and cleaned up need to implement an http endpoint as well for uptime monitoring additional information
0
4,708
24,270,830,013
IssuesEvent
2022-09-28 10:07:29
mozilla/foundation.mozilla.org
https://api.github.com/repos/mozilla/foundation.mozilla.org
closed
SEO | PagesΒ have slow load speedΒ 
engineering Maintain
This is a set of pages that have indicated aslow load speed. In this scenario page speed is measured by the time it takes the crawler to load a webpage’s HTML code - load times for images, JavaScript and CSS are not factored in. The page in this list may have a much heavier load from dynamic content listing - and may benefit from some form of database caching. https://docs.google.com/spreadsheets/d/15HwgpxSYc4Zl809kcebAhLfLYXFuIk8ZP-Qvk3yVV8Q/edit#gid=1209509913
True
SEO | PagesΒ have slow load speedΒ  - This is a set of pages that have indicated aslow load speed. In this scenario page speed is measured by the time it takes the crawler to load a webpage’s HTML code - load times for images, JavaScript and CSS are not factored in. The page in this list may have a much heavier load from dynamic content listing - and may benefit from some form of database caching. https://docs.google.com/spreadsheets/d/15HwgpxSYc4Zl809kcebAhLfLYXFuIk8ZP-Qvk3yVV8Q/edit#gid=1209509913
main
seo pagesΒ have slow load speedΒ  this is a set of pages that have indicated aslow load speed in this scenario page speed is measured by the time it takes the crawler to load a webpage’s html code load times for images javascript and css are not factored in the page in this list may have a much heavier load from dynamic content listing and may benefit from some form of database caching
1
440,637
12,702,145,375
IssuesEvent
2020-06-22 19:35:08
IngenioUN/back_end
https://api.github.com/repos/IngenioUN/back_end
closed
Historia #47 - Agregar nuevos autores
Cap: Back-End Priority: High Type: New
**Tipo de usuario:** Type = 2 -> Administrador **Ruta para ver lista de solicitudes:** /user/author-requests **Ruta para ver informaciΓ³n detallada de solicitud:** /user/accepted-author/{userId}
1.0
Historia #47 - Agregar nuevos autores - **Tipo de usuario:** Type = 2 -> Administrador **Ruta para ver lista de solicitudes:** /user/author-requests **Ruta para ver informaciΓ³n detallada de solicitud:** /user/accepted-author/{userId}
non_main
historia agregar nuevos autores tipo de usuario type administrador ruta para ver lista de solicitudes user author requests ruta para ver informaciΓ³n detallada de solicitud user accepted author userid
0
386
3,420,554,919
IssuesEvent
2015-12-08 15:19:20
dotnet/roslyn-analyzers
https://api.github.com/repos/dotnet/roslyn-analyzers
opened
Fix DoNotIgnoreMethodResultsAnalyzer (CA1806 port) to handle PreserveSig methods
Area-Microsoft.Maintainability.Analyzers Bug
`DoNotIgnoreMethodResultsAnalyzer.IsHResultOrErrorCodeReturningMethod` does not completely match the FXCop implementation [here](http://index/UsageRules/R/88335f66dce7e0ad.html) - especially for ComImport types with methods marked with PreserveSig. We need to tune this method to match the FxCop behavior once https://github.com/dotnet/roslyn/issues/7282 is addressed. I will also add a skip test `UnusedComImportPreserveSig` for this issue.
True
Fix DoNotIgnoreMethodResultsAnalyzer (CA1806 port) to handle PreserveSig methods - `DoNotIgnoreMethodResultsAnalyzer.IsHResultOrErrorCodeReturningMethod` does not completely match the FXCop implementation [here](http://index/UsageRules/R/88335f66dce7e0ad.html) - especially for ComImport types with methods marked with PreserveSig. We need to tune this method to match the FxCop behavior once https://github.com/dotnet/roslyn/issues/7282 is addressed. I will also add a skip test `UnusedComImportPreserveSig` for this issue.
main
fix donotignoremethodresultsanalyzer port to handle preservesig methods donotignoremethodresultsanalyzer ishresultorerrorcodereturningmethod does not completely match the fxcop implementation especially for comimport types with methods marked with preservesig we need to tune this method to match the fxcop behavior once is addressed i will also add a skip test unusedcomimportpreservesig for this issue
1
417,105
28,110,113,387
IssuesEvent
2023-03-31 06:20:55
Nafeij/ped
https://api.github.com/repos/Nafeij/ped
opened
User Guide - Lack of explanations in examples
severity.VeryLow type.DocumentationBug
A single sentence describing the effect of each `Add *` command would be a good addition. E.g: > `add student n/Bellman telegram/97482842 e/e1234567@u.nus.edu score/100` _adds a student named Bellman with telegram number 97482842, email e1234567@u.nus.edu and score 100 to TrAcker_ <!--session: 1680242621233-1c655e1d-6e2e-479d-b22c-f6f42091d9f5--> <!--Version: Web v3.4.7-->
1.0
User Guide - Lack of explanations in examples - A single sentence describing the effect of each `Add *` command would be a good addition. E.g: > `add student n/Bellman telegram/97482842 e/e1234567@u.nus.edu score/100` _adds a student named Bellman with telegram number 97482842, email e1234567@u.nus.edu and score 100 to TrAcker_ <!--session: 1680242621233-1c655e1d-6e2e-479d-b22c-f6f42091d9f5--> <!--Version: Web v3.4.7-->
non_main
user guide lack of explanations in examples a single sentence describing the effect of each add command would be a good addition e g add student n bellman telegram e u nus edu score adds a student named bellman with telegram number email u nus edu and score to tracker
0
71,845
30,921,674,390
IssuesEvent
2023-08-06 01:18:11
Zahlungsmittel/Zahlungsmittel
https://api.github.com/repos/Zahlungsmittel/Zahlungsmittel
opened
[CLOSED] πŸ’₯ [DevOps] Update Cypress
devops test service: e2e end-to-end imported
<a href="https://github.com/mahula"><img src="https://avatars.githubusercontent.com/u/3883288?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [mahula](https://github.com/mahula)** _Tuesday Jun 13, 2023 at 13:04 GMT_ _Originally opened as https://github.com/gradido/gradido/issues/3055_ ---- ## πŸ’₯ DevOps ticket Update cypress relevant topics and packages. - The experimentalSessionAndOrigin configuration option was removed in Cypress version 12 You can safely remove this option from your config. Kann entfernt werden (siehe https://docs.cypress.io/guides/references/migration-guide) - Packages updaten - "@badeball/cypress-cucumber-preprocessor": "^12.0.0", - "@cypress/browserify-preprocessor": "^3.0.2", - "cypress": "^12.7.0", - "eslint-plugin-cypress": "^2.12.1" - ### Todo - [x] remove experimentalSessionAndOrigin configuration option ([it is a general availabitiy since Cypress version 12](https://docs.cypress.io/guides/references/migration-guide)) - [x] update packages `cypress`, `@badeball/cypress-cucumber-preprocessor`, `@cypress/browserify-preprocessor`, `eslint-plugin-cypress`
1.0
[CLOSED] πŸ’₯ [DevOps] Update Cypress - <a href="https://github.com/mahula"><img src="https://avatars.githubusercontent.com/u/3883288?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [mahula](https://github.com/mahula)** _Tuesday Jun 13, 2023 at 13:04 GMT_ _Originally opened as https://github.com/gradido/gradido/issues/3055_ ---- ## πŸ’₯ DevOps ticket Update cypress relevant topics and packages. - The experimentalSessionAndOrigin configuration option was removed in Cypress version 12 You can safely remove this option from your config. Kann entfernt werden (siehe https://docs.cypress.io/guides/references/migration-guide) - Packages updaten - "@badeball/cypress-cucumber-preprocessor": "^12.0.0", - "@cypress/browserify-preprocessor": "^3.0.2", - "cypress": "^12.7.0", - "eslint-plugin-cypress": "^2.12.1" - ### Todo - [x] remove experimentalSessionAndOrigin configuration option ([it is a general availabitiy since Cypress version 12](https://docs.cypress.io/guides/references/migration-guide)) - [x] update packages `cypress`, `@badeball/cypress-cucumber-preprocessor`, `@cypress/browserify-preprocessor`, `eslint-plugin-cypress`
non_main
πŸ’₯ update cypress issue by tuesday jun at gmt originally opened as πŸ’₯ devops ticket update cypress relevant topics and packages the experimentalsessionandorigin configuration option was removed in cypress version you can safely remove this option from your config kann entfernt werden siehe packages updaten badeball cypress cucumber preprocessor cypress browserify preprocessor cypress eslint plugin cypress todo remove experimentalsessionandorigin configuration option update packages cypress badeball cypress cucumber preprocessor cypress browserify preprocessor eslint plugin cypress
0
2,967
10,661,023,780
IssuesEvent
2019-10-18 11:19:31
arcticicestudio/snowsaw
https://api.github.com/repos/arcticicestudio/snowsaw
closed
App version with pure Go Git and SemVer libraries
context-workflow scope-compatibility scope-dx scope-maintainability scope-stability type-improvement
<img src="https://raw.githubusercontent.com/src-d/design/master/go-git/files/go-git-logo.png" /> <!-- <img src="https://raw.githubusercontent.com/src-d/design/master/go-git/files/go-git-github-readme-header.png" /> --> Currently the version of the application is determined by calling `git` commands in a new shell process. This works in most cases, but might fail if Git is not installed on the running system. To prevent further enlargement of the required development environment setup dependencies by adding more checks for external dependencies, the [`go-git` library v4][go-git] (`github.com/src-d/go-git/v4`) will be added: > A highly extensible Git implementation in pure Go. It allows to interact with the project repository and extract required information like the latest tag and commit of the current branch to assemble the application version. To simplify the processing and parsing of the version, the [`semver` library v3][semver-go] (`github.com/Masterminds/semver/v3`) will also be added. A new function will assemble the version of the application from the metadata of the Git repository. It will search for the latest [SemVer][] compatible version tag in the current branch and will fall back to the default version from the application configuration if none is found. If at least one tag is found but it is not the latest commit of the current branch, the build metadata will be appended, consisting of the amount of commits ahead and the shortened reference hash (8 digits) of the latest commit from the current branch. The function will be an early implementation of the Git `describe` command because support in `go-git` has not been implemented yet. See the [full compatibility comparison documentation with Git][go-git-comp] as well as the [proposed Git `describe` command implementation][go-git#816] for more details. [go-git-comp]: https://github.com/src-d/go-git/blob/master/COMPATIBILITY.md [go-git]: https://github.com/src-d/go-git [go-git#816]: https://github.com/src-d/go-git/pull/816 [semver-go]: https://github.com/Masterminds/semver [semver]: https://semver.org
True
App version with pure Go Git and SemVer libraries - <img src="https://raw.githubusercontent.com/src-d/design/master/go-git/files/go-git-logo.png" /> <!-- <img src="https://raw.githubusercontent.com/src-d/design/master/go-git/files/go-git-github-readme-header.png" /> --> Currently the version of the application is determined by calling `git` commands in a new shell process. This works in most cases, but might fail if Git is not installed on the running system. To prevent further enlargement of the required development environment setup dependencies by adding more checks for external dependencies, the [`go-git` library v4][go-git] (`github.com/src-d/go-git/v4`) will be added: > A highly extensible Git implementation in pure Go. It allows to interact with the project repository and extract required information like the latest tag and commit of the current branch to assemble the application version. To simplify the processing and parsing of the version, the [`semver` library v3][semver-go] (`github.com/Masterminds/semver/v3`) will also be added. A new function will assemble the version of the application from the metadata of the Git repository. It will search for the latest [SemVer][] compatible version tag in the current branch and will fall back to the default version from the application configuration if none is found. If at least one tag is found but it is not the latest commit of the current branch, the build metadata will be appended, consisting of the amount of commits ahead and the shortened reference hash (8 digits) of the latest commit from the current branch. The function will be an early implementation of the Git `describe` command because support in `go-git` has not been implemented yet. See the [full compatibility comparison documentation with Git][go-git-comp] as well as the [proposed Git `describe` command implementation][go-git#816] for more details. [go-git-comp]: https://github.com/src-d/go-git/blob/master/COMPATIBILITY.md [go-git]: https://github.com/src-d/go-git [go-git#816]: https://github.com/src-d/go-git/pull/816 [semver-go]: https://github.com/Masterminds/semver [semver]: https://semver.org
main
app version with pure go git and semver libraries currently the version of the application is determined by calling git commands in a new shell process this works in most cases but might fail if git is not installed on the running system to prevent further enlargement of the required development environment setup dependencies by adding more checks for external dependencies the github com src d go git will be added a highly extensible git implementation in pure go it allows to interact with the project repository and extract required information like the latest tag and commit of the current branch to assemble the application version to simplify the processing and parsing of the version the github com masterminds semver will also be added a new function will assemble the version of the application from the metadata of the git repository it will search for the latest compatible version tag in the current branch and will fall back to the default version from the application configuration if none is found if at least one tag is found but it is not the latest commit of the current branch the build metadata will be appended consisting of the amount of commits ahead and the shortened reference hash digits of the latest commit from the current branch the function will be an early implementation of the git describe command because support in go git has not been implemented yet see the as well as the for more details
1
2,892
10,319,644,974
IssuesEvent
2019-08-30 18:06:44
backdrop-ops/contrib
https://api.github.com/repos/backdrop-ops/contrib
closed
Requesting access to contrib
Maintainer application
Hello! I'm Les Lim (https://www.drupal.org/u/les-lim), and I'm the creator of Object Log module (https://www.drupal.org/project/object_log) and a co-maintainer for Mailsystem module (https://www.drupal.org/project/mailsystem) in Drupal 7. I have a working port of Object Log ready to go for Backdrop, and I'd like access to backdrop-ops/contrib to create a project for it there. I'm also interesting in porting the 7.x-3.x branch of Mailsystem to Backdrop - there's an existing port for the 7.x-2.x branch, but the 2.x branch is a fragile version that works by creating a generated PHP file in the public files directory.
True
Requesting access to contrib - Hello! I'm Les Lim (https://www.drupal.org/u/les-lim), and I'm the creator of Object Log module (https://www.drupal.org/project/object_log) and a co-maintainer for Mailsystem module (https://www.drupal.org/project/mailsystem) in Drupal 7. I have a working port of Object Log ready to go for Backdrop, and I'd like access to backdrop-ops/contrib to create a project for it there. I'm also interesting in porting the 7.x-3.x branch of Mailsystem to Backdrop - there's an existing port for the 7.x-2.x branch, but the 2.x branch is a fragile version that works by creating a generated PHP file in the public files directory.
main
requesting access to contrib hello i m les lim and i m the creator of object log module and a co maintainer for mailsystem module in drupal i have a working port of object log ready to go for backdrop and i d like access to backdrop ops contrib to create a project for it there i m also interesting in porting the x x branch of mailsystem to backdrop there s an existing port for the x x branch but the x branch is a fragile version that works by creating a generated php file in the public files directory
1
466,128
13,397,351,954
IssuesEvent
2020-09-03 11:27:15
googleapis/nodejs-language
https://api.github.com/repos/googleapis/nodejs-language
closed
Synthesis failed for nodejs-language
api: language autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate nodejs-language. :broken_heart: Here's the output from running `synth.py`: ``` xternal/gapic_generator_python/requirements.txt (line 4)) Using cached https://files.pythonhosted.org/packages/30/9e/f663a2aa66a09d838042ae1a2c5659828bb9b41ea3a6efa20a20fd92b121/Jinja2-2.11.2-py2.py3-none-any.whl Saved ./Jinja2-2.11.2-py2.py3-none-any.whl Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5)) Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6)) Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7)) Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz Complete output from command python setup.py egg_info: Traceback (most recent call last): File "<string>", line 1, in <module> File "/tmpfs/tmp/tmp_qnzrk07/setuptools-tmp/setuptools/__init__.py", line 6, in <module> import distutils.core File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 82, in create_module return importlib.import_module('._distutils', 'setuptools') File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) ModuleNotFoundError: No module named 'setuptools._distutils' ---------------------------------------- (Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-ljj0zq13/pypandoc/ ) ERROR: no such package '@gapic_generator_python_pip_deps//': pip_import failed: Collecting click==7.1.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 1)) Using cached https://files.pythonhosted.org/packages/d2/3d/fa76db83bf75c4f8d338c2fd15c8d33fdd7ad23a9b5e57eb6c5de26b430e/click-7.1.2-py2.py3-none-any.whl Saved ./click-7.1.2-py2.py3-none-any.whl Collecting google-api-core==1.22.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 2)) Using cached https://files.pythonhosted.org/packages/e0/2d/7c6c75013105e1d2b6eaa1bf18a56995be1dbc673c38885aea31136e9918/google_api_core-1.22.1-py2.py3-none-any.whl Saved ./google_api_core-1.22.1-py2.py3-none-any.whl Collecting googleapis-common-protos==1.52.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 3)) Using cached https://files.pythonhosted.org/packages/03/74/3956721ea1eb4bcf7502a311fdaa60b85bd751de4e57d1943afe9b334141/googleapis_common_protos-1.52.0-py2.py3-none-any.whl Saved ./googleapis_common_protos-1.52.0-py2.py3-none-any.whl Collecting jinja2==2.11.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 4)) Using cached https://files.pythonhosted.org/packages/30/9e/f663a2aa66a09d838042ae1a2c5659828bb9b41ea3a6efa20a20fd92b121/Jinja2-2.11.2-py2.py3-none-any.whl Saved ./Jinja2-2.11.2-py2.py3-none-any.whl Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5)) Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6)) Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7)) Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz Complete output from command python setup.py egg_info: Traceback (most recent call last): File "<string>", line 1, in <module> File "/tmpfs/tmp/tmp_qnzrk07/setuptools-tmp/setuptools/__init__.py", line 6, in <module> import distutils.core File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 82, in create_module return importlib.import_module('._distutils', 'setuptools') File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) ModuleNotFoundError: No module named 'setuptools._distutils' ---------------------------------------- (Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-ljj0zq13/pypandoc/ ) INFO: Elapsed time: 2.361s INFO: 0 processes. FAILED: Build did NOT complete successfully (0 packages loaded) FAILED: Build did NOT complete successfully (0 packages loaded) Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/nodejs-language/synth.py", line 16, in <module> version, File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 52, in node_library return self._generate_code(service, version, "nodejs", **kwargs) File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 183, in _generate_code shell.run(bazel_run_args) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bazel', '--max_idle_secs=240', 'build', '//google/cloud/language/v1:language-v1-nodejs']' returned non-zero exit status 1. 2020-08-31 04:16:36,129 autosynth [ERROR] > Synthesis failed 2020-08-31 04:16:36,129 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at df948af fix: move system and samples test from Node 10 to Node 12 (#503) 2020-08-31 04:16:36,134 autosynth [DEBUG] > Running: git checkout autosynth Switched to branch 'autosynth' 2020-08-31 04:16:36,138 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 690, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 539, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 670, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 375, in synthesize_loop has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 273, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/ab3da4e0-3750-4544-bd59-05f57c3553ea/targets/github%2Fsynthtool;config=default/tests;query=nodejs-language;failed=false).
1.0
Synthesis failed for nodejs-language - Hello! Autosynth couldn't regenerate nodejs-language. :broken_heart: Here's the output from running `synth.py`: ``` xternal/gapic_generator_python/requirements.txt (line 4)) Using cached https://files.pythonhosted.org/packages/30/9e/f663a2aa66a09d838042ae1a2c5659828bb9b41ea3a6efa20a20fd92b121/Jinja2-2.11.2-py2.py3-none-any.whl Saved ./Jinja2-2.11.2-py2.py3-none-any.whl Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5)) Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6)) Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7)) Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz Complete output from command python setup.py egg_info: Traceback (most recent call last): File "<string>", line 1, in <module> File "/tmpfs/tmp/tmp_qnzrk07/setuptools-tmp/setuptools/__init__.py", line 6, in <module> import distutils.core File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 82, in create_module return importlib.import_module('._distutils', 'setuptools') File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) ModuleNotFoundError: No module named 'setuptools._distutils' ---------------------------------------- (Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-ljj0zq13/pypandoc/ ) ERROR: no such package '@gapic_generator_python_pip_deps//': pip_import failed: Collecting click==7.1.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 1)) Using cached https://files.pythonhosted.org/packages/d2/3d/fa76db83bf75c4f8d338c2fd15c8d33fdd7ad23a9b5e57eb6c5de26b430e/click-7.1.2-py2.py3-none-any.whl Saved ./click-7.1.2-py2.py3-none-any.whl Collecting google-api-core==1.22.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 2)) Using cached https://files.pythonhosted.org/packages/e0/2d/7c6c75013105e1d2b6eaa1bf18a56995be1dbc673c38885aea31136e9918/google_api_core-1.22.1-py2.py3-none-any.whl Saved ./google_api_core-1.22.1-py2.py3-none-any.whl Collecting googleapis-common-protos==1.52.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 3)) Using cached https://files.pythonhosted.org/packages/03/74/3956721ea1eb4bcf7502a311fdaa60b85bd751de4e57d1943afe9b334141/googleapis_common_protos-1.52.0-py2.py3-none-any.whl Saved ./googleapis_common_protos-1.52.0-py2.py3-none-any.whl Collecting jinja2==2.11.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 4)) Using cached https://files.pythonhosted.org/packages/30/9e/f663a2aa66a09d838042ae1a2c5659828bb9b41ea3a6efa20a20fd92b121/Jinja2-2.11.2-py2.py3-none-any.whl Saved ./Jinja2-2.11.2-py2.py3-none-any.whl Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5)) Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6)) Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7)) Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz Complete output from command python setup.py egg_info: Traceback (most recent call last): File "<string>", line 1, in <module> File "/tmpfs/tmp/tmp_qnzrk07/setuptools-tmp/setuptools/__init__.py", line 6, in <module> import distutils.core File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 82, in create_module return importlib.import_module('._distutils', 'setuptools') File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) ModuleNotFoundError: No module named 'setuptools._distutils' ---------------------------------------- (Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-ljj0zq13/pypandoc/ ) INFO: Elapsed time: 2.361s INFO: 0 processes. FAILED: Build did NOT complete successfully (0 packages loaded) FAILED: Build did NOT complete successfully (0 packages loaded) Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/nodejs-language/synth.py", line 16, in <module> version, File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 52, in node_library return self._generate_code(service, version, "nodejs", **kwargs) File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 183, in _generate_code shell.run(bazel_run_args) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bazel', '--max_idle_secs=240', 'build', '//google/cloud/language/v1:language-v1-nodejs']' returned non-zero exit status 1. 2020-08-31 04:16:36,129 autosynth [ERROR] > Synthesis failed 2020-08-31 04:16:36,129 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at df948af fix: move system and samples test from Node 10 to Node 12 (#503) 2020-08-31 04:16:36,134 autosynth [DEBUG] > Running: git checkout autosynth Switched to branch 'autosynth' 2020-08-31 04:16:36,138 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 690, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 539, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 670, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 375, in synthesize_loop has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 273, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/ab3da4e0-3750-4544-bd59-05f57c3553ea/targets/github%2Fsynthtool;config=default/tests;query=nodejs-language;failed=false).
non_main
synthesis failed for nodejs language hello autosynth couldn t regenerate nodejs language broken heart here s the output from running synth py xternal gapic generator python requirements txt line using cached saved none any whl collecting markupsafe from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved markupsafe whl collecting protobuf from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved protobuf whl collecting pypandoc from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached complete output from command python setup py egg info traceback most recent call last file line in file tmpfs tmp tmp setuptools tmp setuptools init py line in import distutils core file tmpfs src github synthtool env lib site packages distutils hack init py line in create module return importlib import module distutils setuptools file home kbuilder pyenv versions lib importlib init py line in import module return bootstrap gcd import name package level modulenotfounderror no module named setuptools distutils command python setup py egg info failed with error code in tmpfs tmp pip build pypandoc error no such package gapic generator python pip deps pip import failed collecting click from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved click none any whl collecting google api core from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved google api core none any whl collecting googleapis common protos from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved googleapis common protos none any whl collecting from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved none any whl collecting markupsafe from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved markupsafe whl collecting protobuf from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved protobuf whl collecting pypandoc from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached complete output from command python setup py egg info traceback most recent call last file line in file tmpfs tmp tmp setuptools tmp setuptools init py line in import distutils core file tmpfs src github synthtool env lib site packages distutils hack init py line in create module return importlib import module distutils setuptools file home kbuilder pyenv versions lib importlib init py line in import module return bootstrap gcd import name package level modulenotfounderror no module named setuptools distutils command python setup py egg info failed with error code in tmpfs tmp pip build pypandoc info elapsed time info processes failed build did not complete successfully packages loaded failed build did not complete successfully packages loaded traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool nodejs language synth py line in version file tmpfs src github synthtool synthtool gcp gapic bazel py line in node library return self generate code service version nodejs kwargs file tmpfs src github synthtool synthtool gcp gapic bazel py line in generate code shell run bazel run args file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git reset hard head head is now at fix move system and samples test from node to node autosynth running git checkout autosynth switched to branch autosynth autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop has changes toolbox synthesize version in new branch synthesizer youngest file tmpfs src github synthtool autosynth synth py line in synthesize version in new branch synthesizer synthesize synth log path self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
4,330
21,780,014,594
IssuesEvent
2022-05-13 17:45:15
spack/spack
https://api.github.com/repos/spack/spack
closed
Possible redundancy bug deriving spec hash?
maintainers
I just saw this: https://github.com/spack/spack/blob/83b91246b101a30f78f246fb40fa9df54a1e8403/lib/spack/spack/spec.py#L4900-L4903 Should this be indexing something else if build_spec is found as a key? Or just nix the if and always look for that hash? It seems to do the same thing either way.
True
Possible redundancy bug deriving spec hash? - I just saw this: https://github.com/spack/spack/blob/83b91246b101a30f78f246fb40fa9df54a1e8403/lib/spack/spack/spec.py#L4900-L4903 Should this be indexing something else if build_spec is found as a key? Or just nix the if and always look for that hash? It seems to do the same thing either way.
main
possible redundancy bug deriving spec hash i just saw this should this be indexing something else if build spec is found as a key or just nix the if and always look for that hash it seems to do the same thing either way
1
638,292
20,720,877,078
IssuesEvent
2022-03-13 11:21:40
kiudee/chess-tuning-tools
https://api.github.com/repos/kiudee/chess-tuning-tools
closed
Log version of chess-tuning-tools
enhancement Priority: Low
It can be useful to see at a glance which version of chess-tuning-tools was used for a given log file.
1.0
Log version of chess-tuning-tools - It can be useful to see at a glance which version of chess-tuning-tools was used for a given log file.
non_main
log version of chess tuning tools it can be useful to see at a glance which version of chess tuning tools was used for a given log file
0
615,372
19,254,362,694
IssuesEvent
2021-12-09 09:42:20
googleapis/nodejs-pubsub
https://api.github.com/repos/googleapis/nodejs-pubsub
opened
subscriptions: should create a subscription with dead letter policy. failed
priority: p1 type: bug flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 98840fcc38a05be07de6e7dc97d8a79616940df9 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/a0cb08d8-f564-458e-92c9-c0182d1f2367), [Sponge](http://sponge2/a0cb08d8-f564-458e-92c9-c0182d1f2367) status: failed <details><summary>Test output</summary><br><pre>expected '' to include 'Created subscription sub5-8c0f3bbd-4216-4298-a5a3-67b8ec98e1ba with dead letter topic topic3-8c0f3bbd-4216-4298-a5a3-67b8ec98e1ba.' AssertionError: expected '' to include 'Created subscription sub5-8c0f3bbd-4216-4298-a5a3-67b8ec98e1ba with dead letter topic topic3-8c0f3bbd-4216-4298-a5a3-67b8ec98e1ba.' at Context.<anonymous> (build/system-test/subscriptions.test.js:182:23) -> /workspace/samples/system-test/subscriptions.test.ts:258:12 at processImmediate (internal/timers.js:461:21)</pre></details>
1.0
subscriptions: should create a subscription with dead letter policy. failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 98840fcc38a05be07de6e7dc97d8a79616940df9 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/a0cb08d8-f564-458e-92c9-c0182d1f2367), [Sponge](http://sponge2/a0cb08d8-f564-458e-92c9-c0182d1f2367) status: failed <details><summary>Test output</summary><br><pre>expected '' to include 'Created subscription sub5-8c0f3bbd-4216-4298-a5a3-67b8ec98e1ba with dead letter topic topic3-8c0f3bbd-4216-4298-a5a3-67b8ec98e1ba.' AssertionError: expected '' to include 'Created subscription sub5-8c0f3bbd-4216-4298-a5a3-67b8ec98e1ba with dead letter topic topic3-8c0f3bbd-4216-4298-a5a3-67b8ec98e1ba.' at Context.<anonymous> (build/system-test/subscriptions.test.js:182:23) -> /workspace/samples/system-test/subscriptions.test.ts:258:12 at processImmediate (internal/timers.js:461:21)</pre></details>
non_main
subscriptions should create a subscription with dead letter policy failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output expected to include created subscription with dead letter topic assertionerror expected to include created subscription with dead letter topic at context build system test subscriptions test js workspace samples system test subscriptions test ts at processimmediate internal timers js
0
446,287
12,854,602,850
IssuesEvent
2020-07-09 02:28:38
kubesphere/kubesphere
https://api.github.com/repos/kubesphere/kubesphere
opened
Admin can't open any page under system-workspace
area/console kind/bug priority/high
The `admin` user almost can open any page of the host cluster under multi-cluster env. <img width="1229" alt="Screen Shot 2020-07-09 at 10 25 50 AM" src="https://user-images.githubusercontent.com/28859385/86990267-cd4caf00-c1ce-11ea-99ac-c215d7bfcd5c.png">
1.0
Admin can't open any page under system-workspace - The `admin` user almost can open any page of the host cluster under multi-cluster env. <img width="1229" alt="Screen Shot 2020-07-09 at 10 25 50 AM" src="https://user-images.githubusercontent.com/28859385/86990267-cd4caf00-c1ce-11ea-99ac-c215d7bfcd5c.png">
non_main
admin can t open any page under system workspace the admin user almost can open any page of the host cluster under multi cluster env img width alt screen shot at am src
0
1,825
6,577,335,360
IssuesEvent
2017-09-12 00:11:29
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
module junos_config with the parameter lines + delete: the delete details doesnt apply
affects_2.1 bug_report networking waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> junos_config https://github.com/ansible/ansible-modules-core/blob/devel/network/junos/junos_config.py http://docs.ansible.com/ansible/junos_config_module.html ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ksator@ubuntu:~/ansible-training-for-junos$ ansible --version ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ksator@ubuntu:~/ansible-training-for-junos$ ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> default ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say β€œN/A” for anything that is not platform-specific. --> $ uname -a Linux ubuntu 3.19.0-25-generic #26~14.04.1-Ubuntu SMP Fri Jul 24 21:16:20 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux ##### SUMMARY <!--- Explain the problem briefly --> can not delete junos configuration lines on junos devices using the module junos_config with the parameter lines + delete (the delete details doesnt apply). ##### STEPS TO REPRODUCE <!--- --> I used this playbook https://github.com/ksator/ansible-training-for-junos/blob/master/junos_config/playbook.yml (module junos_config, with the parameter lines, and some lines starting wuth delete). ansible-playbook junos_config/playbook.yml <!--- Paste example playbooks or commands between quotes below --> ``` --- - name: core module junos_config hosts: Chris-EX4200-test connection: local gather_facts: no vars_prompt: - name: DEVICE_PASSWORD prompt: Device password private: yes tasks: - name: test junos_config: host: "{{ inventory_hostname }}" username: pytraining password: "{{ DEVICE_PASSWORD }}" lines: - delete system name-server 172.30.179.2 - set system host-name newname ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> The expected result with module junos_config with the parameter lines with lines starting with delete is: this module should remove these junos lines on remote devices. with the playbook above, this "system name-server 172.30.179.2" should be removed from the junos configuration of the host. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> the playbook did not remove "system name-server 172.30.179.2" on the host. however the details of the set command (set system host-name newname) has been added succesfully. <!--- Paste verbatim command output between quotes below --> ``` I ran this playbook (https://github.com/ksator/ansible-training-for-junos/blob/master/junos_config/playbook.yml). with extra verbosity (-vvvv). and I enabled netconf log on device (http://www.juniper.net/documentation/en_US/junos14.2/topics/topic-map/netconf-traceoptions.html). here's the output log. ksator@ubuntu:~/ansible-training-for-junos$ ansible-playbook junos_config/playbook.yml -i hosts -vvvv Using /etc/ansible/ansible.cfg as config file Loaded callback default of type stdout, v2.0 PLAYBOOK: playbook.yml ********************************************************* 1 plays in junos_config/playbook.yml Device password: PLAY [core module junos_config] ************************************************ TASK [test] ******************************************************************** task path: /home/ksator/ansible-training-for-junos/junos_config/playbook.yml:13 <172.30.179.113> ESTABLISH LOCAL CONNECTION FOR USER: ksator <172.30.179.113> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1466153012.52-229272728614829 `" && echo ansible-tmp-1466153012.52-229272728614829="` echo $HOME/.ansible/tmp/ansible-tmp-1466153012.52-229272728614829 `" ) && sleep 0' <172.30.179.113> PUT /tmp/tmptlRDup TO /home/ksator/.ansible/tmp/ansible-tmp-1466153012.52-229272728614829/junos_config <172.30.179.113> EXEC /bin/sh -c 'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python /home/ksator/.ansible/tmp/ansible-tmp-1466153012.52-229272728614829/junos_config; rm -rf "/home/ksator/.ansible/tmp/ansible-tmp-1466153012.52-229272728614829/" > /dev/null 2>&1 && sleep 0' changed: [172.30.179.113] => {"changed": true, "diff": {"prepared": "\n[edit system]\n- host-name ex4200-13;\n+ host-name newname;\n"}, "invocation": {"module_args": {"comment": "update config", "confirm": 1, "host": "172.30.179.113", "lines": ["delete system name-server 172.30.179.2", "set system host-name newname"], "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "provider": null, "replace": false, "rollback": null, "ssh_keyfile": null, "timeout": 0, "transport": "netconf", "username": "pytraining", "zeroize": false}, "module_name": "junos_config"}} PLAY RECAP ********************************************************************* 172.30.179.113 : ok=1 changed=1 unreachable=0 failed=0 ksator@ubuntu:~/ansible-training-for-junos$ netconf logs on junos device pytraining@ex4200-13> show log netconf-ops.log Jun 17 10:43:35 Started tracing session: 48789 Jun 17 10:43:35 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:hello xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0"><nc:capabilities><nc:capability>urn:ietf:params:netconf:capability:writable-running:1.0</nc:capability><nc:capability>urn:ietf:params:netconf:capability:rollback-on-error:1.0</nc:capability><nc:capability>urn:liberouter:params:netconf:capability:power-control:1.0</nc:capability><nc:capability>urn:ietf:params:netconf:capability:validate:1.0</nc:capability><nc:capability>urn:i Jun 17 10:43:36 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94529d66-3467-11e6-bdb3-000c29432de4"><get-configuration format="set"/></nc:rpc>]]>]]> Jun 17 10:43:36 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94529d66-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:36 [48789] Outgoing: <configuration xmlns="http://xml.juniper.net/xnm/1.1/xnm" junos:changed-seconds="1466152917" junos:changed-localtime="2016-06-17 10:41:57 CEST"> Jun 17 10:43:36 [48789] Outgoing: <version>14.1X53-D25.2</version> Jun 17 10:43:36 [48789] Outgoing: <groups> Jun 17 10:43:36 [48789] Outgoing: <name>access</name> Jun 17 10:43:36 [48789] Outgoing: <interfaces> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>&lt;ge-*&gt;</name> Jun 17 10:43:36 [48789] Outgoing: <mtu>4484</mtu> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: <port-mode>access</port-mode> Jun 17 10:43:36 [48789] Outgoing: </ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>&lt;xe-*&gt;</name> Jun 17 10:43:36 [48789] Outgoing: <mtu>4484</mtu> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: <port-mode>access</port-mode> Jun 17 10:43:36 [48789] Outgoing: </ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: </interfaces> Jun 17 10:43:36 [48789] Outgoing: </groups> Jun 17 10:43:36 [48789] Outgoing: <groups> Jun 17 10:43:36 [48789] Outgoing: <name>trunk</name> Jun 17 10:43:36 [48789] Outgoing: <interfaces> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>&lt;ge-*&gt;</name> Jun 17 10:43:36 [48789] Outgoing: <mtu>4484</mtu> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: <port-mode>trunk</port-mode> Jun 17 10:43:36 [48789] Outgoing: <vlan> Jun 17 10:43:36 [48789] Outgoing: <members>all</members> Jun 17 10:43:36 [48789] Outgoing: </vlan> Jun 17 10:43:36 [48789] Outgoing: </ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>&lt;xe-*&gt;</name> Jun 17 10:43:36 [48789] Outgoing: <mtu>4484</mtu> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: <port-mode>trunk</port-mode> Jun 17 10:43:36 [48789] Outgoing: <vlan> Jun 17 10:43:36 [48789] Outgoing: <members>all</members> Jun 17 10:43:36 [48789] Outgoing: </vlan> Jun 17 10:43:36 [48789] Outgoing: </ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: </interfaces> Jun 17 10:43:36 [48789] Outgoing: </groups> Jun 17 10:43:36 [48789] Outgoing: <system> Jun 17 10:43:36 [48789] Outgoing: <host-name>ex4200-13</host-name> Jun 17 10:43:36 [48789] Outgoing: <domain-name>poc-nl.jnpr.net</domain-name> Jun 17 10:43:36 [48789] Outgoing: <time-zone>Europe/Amsterdam</time-zone> Jun 17 10:43:36 [48789] Outgoing: <authentication-order>radius</authentication-order> Jun 17 10:43:36 [48789] Outgoing: <authentication-order>password</authentication-order> Jun 17 10:43:36 [48789] Outgoing: <root-authentication> Jun 17 10:43:36 [48789] Outgoing: <encrypted-password>$1$/NHg28eO$pqaVlLlPQ2thlQQ0ZB.Vx/</encrypted-password> Jun 17 10:43:36 [48789] Outgoing: </root-authentication> Jun 17 10:43:36 [48789] Outgoing: <name-server> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.179.2</name> Jun 17 10:43:36 [48789] Outgoing: </name-server> Jun 17 10:43:36 [48789] Outgoing: <name-server> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.179.3</name> Jun 17 10:43:36 [48789] Outgoing: </name-server> Jun 17 10:43:36 [48789] Outgoing: <radius-server> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.176.9</name> Jun 17 10:43:36 [48789] Outgoing: <secret>$9$DMHPTz36CtOqmBEclLXik.mfT6/t1Eyn/</secret> Jun 17 10:43:36 [48789] Outgoing: <retry>3</retry> Jun 17 10:43:36 [48789] Outgoing: </radius-server> Jun 17 10:43:36 [48789] Outgoing: <radius-server> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.176.4</name> Jun 17 10:43:36 [48789] Outgoing: <secret>$9$CgY9p1EcylvWx0B7VwgUDtuOBIEleWNVYre</secret> Jun 17 10:43:36 [48789] Outgoing: <retry>3</retry> Jun 17 10:43:36 [48789] Outgoing: </radius-server> Jun 17 10:43:36 [48789] Outgoing: <login> Jun 17 10:43:36 [48789] Outgoing: <user> Jun 17 10:43:36 [48789] Outgoing: <name>pytraining-13</name> Jun 17 10:43:36 [48789] Outgoing: <uid>2001</uid> Jun 17 10:43:36 [48789] Outgoing: <class>super-user</class> Jun 17 10:43:36 [48789] Outgoing: <authentication> Jun 17 10:43:36 [48789] Outgoing: <encrypted-password>$1$uH0eoaXS$tTd6k7k.AVkEDhdJ8V75F.</encrypted-password> Jun 17 10:43:36 [48789] Outgoing: </authentication> Jun 17 10:43:36 [48789] Outgoing: </user> Jun 17 10:43:36 [48789] Outgoing: <user> Jun 17 10:43:36 [48789] Outgoing: <name>remote</name> Jun 17 10:43:36 [48789] Outgoing: <uid>2000</uid> Jun 17 10:43:36 [48789] Outgoing: <class>super-user</class> Jun 17 10:43:36 [48789] Outgoing: </user> Jun 17 10:43:36 [48789] Outgoing: </login> Jun 17 10:43:36 [48789] Outgoing: <services> Jun 17 10:43:36 [48789] Outgoing: <ftp> Jun 17 10:43:36 [48789] Outgoing: </ftp> Jun 17 10:43:36 [48789] Outgoing: <ssh> Jun 17 10:43:36 [48789] Outgoing: <client-alive-interval>120</client-alive-interval> Jun 17 10:43:36 [48789] Outgoing: </ssh> Jun 17 10:43:36 [48789] Outgoing: <telnet> Jun 17 10:43:36 [48789] Outgoing: </telnet> Jun 17 10:43:36 [48789] Outgoing: <xnm-clear-text> Jun 17 10:43:36 [48789] Outgoing: </xnm-clear-text> Jun 17 10:43:36 [48789] Outgoing: <netconf> Jun 17 10:43:36 [48789] Outgoing: <ssh> Jun 17 10:43:36 [48789] Outgoing: </ssh> Jun 17 10:43:36 [48789] Outgoing: <traceoptions> Jun 17 10:43:36 [48789] Outgoing: <file> Jun 17 10:43:36 [48789] Outgoing: <filename>netconf-ops.log</filename> Jun 17 10:43:36 [48789] Outgoing: <size>3m</size> Jun 17 10:43:36 [48789] Outgoing: <files>20</files> Jun 17 10:43:36 [48789] Outgoing: <world-readable/> Jun 17 10:43:36 [48789] Outgoing: </file> Jun 17 10:43:36 [48789] Outgoing: <flag> Jun 17 10:43:36 [48789] Outgoing: <name>all</name> Jun 17 10:43:36 [48789] Outgoing: </flag> Jun 17 10:43:36 [48789] Outgoing: </traceoptions> Jun 17 10:43:36 [48789] Outgoing: </netconf> Jun 17 10:43:36 [48789] Outgoing: <web-management> Jun 17 10:43:36 [48789] Outgoing: <http> Jun 17 10:43:36 [48789] Outgoing: </http> Jun 17 10:43:36 [48789] Outgoing: </web-management> Jun 17 10:43:36 [48789] Outgoing: </services> Jun 17 10:43:36 [48789] Outgoing: <syslog> Jun 17 10:43:36 [48789] Outgoing: <user> Jun 17 10:43:36 [48789] Outgoing: <name>*</name> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>any</name> Jun 17 10:43:36 [48789] Outgoing: <emergency/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: </user> Jun 17 10:43:36 [48789] Outgoing: <host> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.189.13</name> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>any</name> Jun 17 10:43:36 [48789] Outgoing: <notice/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>authorization</name> Jun 17 10:43:36 [48789] Outgoing: <info/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>interactive-commands</name> Jun 17 10:43:36 [48789] Outgoing: <info/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: </host> Jun 17 10:43:36 [48789] Outgoing: <host> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.189.14</name> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>any</name> Jun 17 10:43:36 [48789] Outgoing: <notice/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>authorization</name> Jun 17 10:43:36 [48789] Outgoing: <info/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>interactive-commands</name> Jun 17 10:43:36 [48789] Outgoing: <info/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: </host> Jun 17 10:43:36 [48789] Outgoing: <file> Jun 17 10:43:36 [48789] Outgoing: <name>messages</name> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>any</name> Jun 17 10:43:36 [48789] Outgoing: <notice/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>authorization</name> Jun 17 10:43:36 [48789] Outgoing: <info/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: </file> Jun 17 10:43:36 [48789] Outgoing: </syslog> Jun 17 10:43:36 [48789] Outgoing: <compress-configuration-files/> Jun 17 10:43:36 [48789] Outgoing: <commit> Jun 17 10:43:36 [48789] Outgoing: <synchronize/> Jun 17 10:43:36 [48789] Outgoing: </commit> Jun 17 10:43:36 [48789] Outgoing: <ntp> Jun 17 10:43:36 [48789] Outgoing: <boot-server>172.30.179.3</boot-server> Jun 17 10:43:36 [48789] Outgoing: <server> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.179.3</name> Jun 17 10:43:36 [48789] Outgoing: </server> Jun 17 10:43:36 [48789] Outgoing: <server> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.179.2</name> Jun 17 10:43:36 [48789] Outgoing: </server> Jun 17 10:43:36 [48789] Outgoing: </ntp> Jun 17 10:43:36 [48789] Outgoing: </system> Jun 17 10:43:36 [48789] Outgoing: <interfaces> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>ge-0/0/0</name> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <inet> Jun 17 10:43:36 [48789] Outgoing: <address> Jun 17 10:43:36 [48789] Outgoing: <name>10.0.0.2/24</name> Jun 17 10:43:36 [48789] Outgoing: </address> Jun 17 10:43:36 [48789] Outgoing: </inet> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>lo0</name> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <inet> Jun 17 10:43:36 [48789] Outgoing: <address> Jun 17 10:43:36 [48789] Outgoing: <name>192.179.0.113/32</name> Jun 17 10:43:36 [48789] Outgoing: <preferred/> Jun 17 10:43:36 [48789] Outgoing: </address> Jun 17 10:43:36 [48789] Outgoing: <address> Jun 17 10:43:36 [48789] Outgoing: <name>127.0.0.1/32</name> Jun 17 10:43:36 [48789] Outgoing: </address> Jun 17 10:43:36 [48789] Outgoing: </inet> Jun 17 10:43:36 [48789] Outgoing: <iso> Jun 17 10:43:36 [48789] Outgoing: <address> Jun 17 10:43:36 [48789] Outgoing: <name>49.0179.0000.0000.0113.00</name> Jun 17 10:43:36 [48789] Outgoing: </address> Jun 17 10:43:36 [48789] Outgoing: </iso> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>me0</name> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <inet> Jun 17 10:43:36 [48789] Outgoing: <address> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.179.113/24</name> Jun 17 10:43:36 [48789] Outgoing: </address> Jun 17 10:43:36 [48789] Outgoing: </inet> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: </interfaces> Jun 17 10:43:36 [48789] Outgoing: <snmp> Jun 17 10:43:36 [48789] Outgoing: <location>AMS, EPOC 3.18</location> Jun 17 10:43:36 [48789] Outgoing: <contact>emea-poc@juniper.net</contact> Jun 17 10:43:36 [48789] Outgoing: <community> Jun 17 10:43:36 [48789] Outgoing: <name>public</name> Jun 17 10:43:36 [48789] Outgoing: <authorization>read-only</authorization> Jun 17 10:43:36 [48789] Outgoing: <clients> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.0.0/16</name> Jun 17 10:43:36 [48789] Outgoing: </clients> Jun 17 10:43:36 [48789] Outgoing: <clients> Jun 17 10:43:36 [48789] Outgoing: <name>0.0.0.0/0</name> Jun 17 10:43:36 [48789] Outgoing: <restrict/> Jun 17 10:43:36 [48789] Outgoing: </clients> Jun 17 10:43:36 [48789] Outgoing: </community> Jun 17 10:43:36 [48789] Outgoing: <community> Jun 17 10:43:36 [48789] Outgoing: <name>private</name> Jun 17 10:43:36 [48789] Outgoing: <authorization>read-write</authorization> Jun 17 10:43:36 [48789] Outgoing: <clients> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.0.0/16</name> Jun 17 10:43:36 [48789] Outgoing: </clients> Jun 17 10:43:36 [48789] Outgoing: <clients> Jun 17 10:43:36 [48789] Outgoing: <name>0.0.0.0/0</name> Jun 17 10:43:36 [48789] Outgoing: <restrict/> Jun 17 10:43:36 [48789] Outgoing: </clients> Jun 17 10:43:36 [48789] Outgoing: </community> Jun 17 10:43:36 [48789] Outgoing: <trap-options> Jun 17 10:43:36 [48789] Outgoing: <source-address> Jun 17 10:43:36 [48789] Outgoing: <address>172.30.179.113</address> Jun 17 10:43:36 [48789] Outgoing: </source-address> Jun 17 10:43:36 [48789] Outgoing: </trap-options> Jun 17 10:43:36 [48789] Outgoing: </snmp> Jun 17 10:43:36 [48789] Outgoing: <routing-options> Jun 17 10:43:36 [48789] Outgoing: <static> Jun 17 10:43:36 [48789] Outgoing: <route> Jun 17 10:43:36 [48789] Outgoing: <name>0.0.0.0/0</name> Jun 17 10:43:36 [48789] Outgoing: <next-hop>172.30.179.1</next-hop> Jun 17 10:43:36 [48789] Outgoing: <no-readvertise/> Jun 17 10:43:36 [48789] Outgoing: </route> Jun 17 10:43:36 [48789] Outgoing: </static> Jun 17 10:43:36 [48789] Outgoing: <router-id>192.179.0.113</router-id> Jun 17 10:43:36 [48789] Outgoing: <autonomous-system> Jun 17 10:43:36 [48789] Outgoing: <as-number>100</as-number> Jun 17 10:43:36 [48789] Outgoing: </autonomous-system> Jun 17 10:43:36 [48789] Outgoing: </routing-options> Jun 17 10:43:36 [48789] Outgoing: <vlans> Jun 17 10:43:36 [48789] Outgoing: <vlan> Jun 17 10:43:36 [48789] Outgoing: <name>test1</name> Jun 17 10:43:36 [48789] Outgoing: <vlan-id>1111</vlan-id> Jun 17 10:43:36 [48789] Outgoing: </vlan> Jun 17 10:43:36 [48789] Outgoing: <vlan> Jun 17 10:43:36 [48789] Outgoing: <name>test2</name> Jun 17 10:43:36 [48789] Outgoing: <vlan-id>1112</vlan-id> Jun 17 10:43:36 [48789] Outgoing: </vlan> Jun 17 10:43:36 [48789] Outgoing: <vlan> Jun 17 10:43:36 [48789] Outgoing: <name>test3</name> Jun 17 10:43:36 [48789] Outgoing: <vlan-id>1113</vlan-id> Jun 17 10:43:36 [48789] Outgoing: </vlan> Jun 17 10:43:36 [48789] Outgoing: </vlans> Jun 17 10:43:36 [48789] Outgoing: </configuration> Jun 17 10:43:36 [48789] Outgoing: </rpc-reply> Jun 17 10:43:36 [48789] Outgoing: ]]>]]> Jun 17 10:43:36 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94c9e182-3467-11e6-bdb3-000c29432de4"><lock-configuration/></nc:rpc>]]>]]> Jun 17 10:43:36 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94c9e182-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:36 [48789] Outgoing: </rpc-reply> Jun 17 10:43:36 [48789] Outgoing: ]]>]]> Jun 17 10:43:36 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94dc9a8e-3467-11e6-bdb3-000c29432de4"><load-configuration action="set" format="text"><configuration-set>set system host-name newname</configuration-set></load-configuration></nc:rpc>]]>]]> Jun 17 10:43:36 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94dc9a8e-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:36 [48789] Outgoing: <load-configuration-results> Jun 17 10:43:36 [48789] Outgoing: <ok/> Jun 17 10:43:36 [48789] Outgoing: </load-configuration-results> Jun 17 10:43:36 [48789] Outgoing: </rpc-reply> Jun 17 10:43:36 [48789] Outgoing: ]]>]]> Jun 17 10:43:37 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94f6f528-3467-11e6-bdb3-000c29432de4"><get-configuration compare="rollback" rollback="0" format="text"/></nc:rpc>]]>]]> Jun 17 10:43:37 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94f6f528-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:37 [48789] Outgoing: <configuration-information> Jun 17 10:43:37 [48789] Outgoing: <configuration-output> Jun 17 10:43:37 [48789] Outgoing: [edit system] Jun 17 10:43:37 [48789] Outgoing: - host-name ex4200-13; Jun 17 10:43:37 [48789] Outgoing: + host-name newname; Jun 17 10:43:37 [48789] Outgoing: </configuration-output> Jun 17 10:43:37 [48789] Outgoing: </configuration-information> Jun 17 10:43:37 [48789] Outgoing: </rpc-reply> Jun 17 10:43:37 [48789] Outgoing: ]]>]]> Jun 17 10:43:37 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:9565a07c-3467-11e6-bdb3-000c29432de4"><commit-configuration><check/></commit-configuration></nc:rpc>]]>]]> Jun 17 10:43:37 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:9565a07c-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:37 [48789] Outgoing: <commit-results> Jun 17 10:43:39 [48789] Outgoing: <routing-engine junos:style="normal"> Jun 17 10:43:39 [48789] Outgoing: <name>fpc0</name> Jun 17 10:43:39 [48789] Outgoing: <commit-check-success/> Jun 17 10:43:39 [48789] Outgoing: </routing-engine> Jun 17 10:43:39 [48789] Outgoing: </commit-results> Jun 17 10:43:39 [48789] Outgoing: <ok/> Jun 17 10:43:39 [48789] Outgoing: </rpc-reply> Jun 17 10:43:39 [48789] Outgoing: ]]>]]> Jun 17 10:43:39 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:96662b2c-3467-11e6-bdb3-000c29432de4"><commit-configuration><confirm-timeout>1</confirm-timeout><confirmed/><log>update config</log></commit-configuration></nc:rpc>]]>]]> Jun 17 10:43:39 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:96662b2c-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:39 [48789] Outgoing: <commit-results> Jun 17 10:43:41 [48789] Outgoing: <routing-engine junos:style="normal"> Jun 17 10:43:41 [48789] Outgoing: <name>fpc0</name> Jun 17 10:43:41 [48789] Outgoing: <commit-check-success/> Jun 17 10:43:44 [48789] Outgoing: <commit-success/> Jun 17 10:43:44 [48789] Outgoing: </routing-engine> Jun 17 10:43:44 [48789] Outgoing: </commit-results> Jun 17 10:43:44 [48789] Outgoing: <ok/> Jun 17 10:43:44 [48789] Outgoing: </rpc-reply> Jun 17 10:43:44 [48789] Outgoing: ]]>]]> Jun 17 10:43:44 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:99986dc8-3467-11e6-bdb3-000c29432de4"><unlock-configuration/></nc:rpc>]]>]]> Jun 17 10:43:44 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:99986dc8-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:44 [48789] Outgoing: </rpc-reply> Jun 17 10:43:46 [48789] Outgoing: ]]>]]> Jun 17 10:43:46 [48789] Outgoing: <!-- session end at 2016-06-17 10:43:46 CEST --> ```
True
module junos_config with the parameter lines + delete: the delete details doesnt apply - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> junos_config https://github.com/ansible/ansible-modules-core/blob/devel/network/junos/junos_config.py http://docs.ansible.com/ansible/junos_config_module.html ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ksator@ubuntu:~/ansible-training-for-junos$ ansible --version ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ksator@ubuntu:~/ansible-training-for-junos$ ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> default ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say β€œN/A” for anything that is not platform-specific. --> $ uname -a Linux ubuntu 3.19.0-25-generic #26~14.04.1-Ubuntu SMP Fri Jul 24 21:16:20 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux ##### SUMMARY <!--- Explain the problem briefly --> can not delete junos configuration lines on junos devices using the module junos_config with the parameter lines + delete (the delete details doesnt apply). ##### STEPS TO REPRODUCE <!--- --> I used this playbook https://github.com/ksator/ansible-training-for-junos/blob/master/junos_config/playbook.yml (module junos_config, with the parameter lines, and some lines starting wuth delete). ansible-playbook junos_config/playbook.yml <!--- Paste example playbooks or commands between quotes below --> ``` --- - name: core module junos_config hosts: Chris-EX4200-test connection: local gather_facts: no vars_prompt: - name: DEVICE_PASSWORD prompt: Device password private: yes tasks: - name: test junos_config: host: "{{ inventory_hostname }}" username: pytraining password: "{{ DEVICE_PASSWORD }}" lines: - delete system name-server 172.30.179.2 - set system host-name newname ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> The expected result with module junos_config with the parameter lines with lines starting with delete is: this module should remove these junos lines on remote devices. with the playbook above, this "system name-server 172.30.179.2" should be removed from the junos configuration of the host. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> the playbook did not remove "system name-server 172.30.179.2" on the host. however the details of the set command (set system host-name newname) has been added succesfully. <!--- Paste verbatim command output between quotes below --> ``` I ran this playbook (https://github.com/ksator/ansible-training-for-junos/blob/master/junos_config/playbook.yml). with extra verbosity (-vvvv). and I enabled netconf log on device (http://www.juniper.net/documentation/en_US/junos14.2/topics/topic-map/netconf-traceoptions.html). here's the output log. ksator@ubuntu:~/ansible-training-for-junos$ ansible-playbook junos_config/playbook.yml -i hosts -vvvv Using /etc/ansible/ansible.cfg as config file Loaded callback default of type stdout, v2.0 PLAYBOOK: playbook.yml ********************************************************* 1 plays in junos_config/playbook.yml Device password: PLAY [core module junos_config] ************************************************ TASK [test] ******************************************************************** task path: /home/ksator/ansible-training-for-junos/junos_config/playbook.yml:13 <172.30.179.113> ESTABLISH LOCAL CONNECTION FOR USER: ksator <172.30.179.113> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1466153012.52-229272728614829 `" && echo ansible-tmp-1466153012.52-229272728614829="` echo $HOME/.ansible/tmp/ansible-tmp-1466153012.52-229272728614829 `" ) && sleep 0' <172.30.179.113> PUT /tmp/tmptlRDup TO /home/ksator/.ansible/tmp/ansible-tmp-1466153012.52-229272728614829/junos_config <172.30.179.113> EXEC /bin/sh -c 'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python /home/ksator/.ansible/tmp/ansible-tmp-1466153012.52-229272728614829/junos_config; rm -rf "/home/ksator/.ansible/tmp/ansible-tmp-1466153012.52-229272728614829/" > /dev/null 2>&1 && sleep 0' changed: [172.30.179.113] => {"changed": true, "diff": {"prepared": "\n[edit system]\n- host-name ex4200-13;\n+ host-name newname;\n"}, "invocation": {"module_args": {"comment": "update config", "confirm": 1, "host": "172.30.179.113", "lines": ["delete system name-server 172.30.179.2", "set system host-name newname"], "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "provider": null, "replace": false, "rollback": null, "ssh_keyfile": null, "timeout": 0, "transport": "netconf", "username": "pytraining", "zeroize": false}, "module_name": "junos_config"}} PLAY RECAP ********************************************************************* 172.30.179.113 : ok=1 changed=1 unreachable=0 failed=0 ksator@ubuntu:~/ansible-training-for-junos$ netconf logs on junos device pytraining@ex4200-13> show log netconf-ops.log Jun 17 10:43:35 Started tracing session: 48789 Jun 17 10:43:35 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:hello xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0"><nc:capabilities><nc:capability>urn:ietf:params:netconf:capability:writable-running:1.0</nc:capability><nc:capability>urn:ietf:params:netconf:capability:rollback-on-error:1.0</nc:capability><nc:capability>urn:liberouter:params:netconf:capability:power-control:1.0</nc:capability><nc:capability>urn:ietf:params:netconf:capability:validate:1.0</nc:capability><nc:capability>urn:i Jun 17 10:43:36 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94529d66-3467-11e6-bdb3-000c29432de4"><get-configuration format="set"/></nc:rpc>]]>]]> Jun 17 10:43:36 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94529d66-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:36 [48789] Outgoing: <configuration xmlns="http://xml.juniper.net/xnm/1.1/xnm" junos:changed-seconds="1466152917" junos:changed-localtime="2016-06-17 10:41:57 CEST"> Jun 17 10:43:36 [48789] Outgoing: <version>14.1X53-D25.2</version> Jun 17 10:43:36 [48789] Outgoing: <groups> Jun 17 10:43:36 [48789] Outgoing: <name>access</name> Jun 17 10:43:36 [48789] Outgoing: <interfaces> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>&lt;ge-*&gt;</name> Jun 17 10:43:36 [48789] Outgoing: <mtu>4484</mtu> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: <port-mode>access</port-mode> Jun 17 10:43:36 [48789] Outgoing: </ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>&lt;xe-*&gt;</name> Jun 17 10:43:36 [48789] Outgoing: <mtu>4484</mtu> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: <port-mode>access</port-mode> Jun 17 10:43:36 [48789] Outgoing: </ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: </interfaces> Jun 17 10:43:36 [48789] Outgoing: </groups> Jun 17 10:43:36 [48789] Outgoing: <groups> Jun 17 10:43:36 [48789] Outgoing: <name>trunk</name> Jun 17 10:43:36 [48789] Outgoing: <interfaces> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>&lt;ge-*&gt;</name> Jun 17 10:43:36 [48789] Outgoing: <mtu>4484</mtu> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: <port-mode>trunk</port-mode> Jun 17 10:43:36 [48789] Outgoing: <vlan> Jun 17 10:43:36 [48789] Outgoing: <members>all</members> Jun 17 10:43:36 [48789] Outgoing: </vlan> Jun 17 10:43:36 [48789] Outgoing: </ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>&lt;xe-*&gt;</name> Jun 17 10:43:36 [48789] Outgoing: <mtu>4484</mtu> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: <port-mode>trunk</port-mode> Jun 17 10:43:36 [48789] Outgoing: <vlan> Jun 17 10:43:36 [48789] Outgoing: <members>all</members> Jun 17 10:43:36 [48789] Outgoing: </vlan> Jun 17 10:43:36 [48789] Outgoing: </ethernet-switching> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: </interfaces> Jun 17 10:43:36 [48789] Outgoing: </groups> Jun 17 10:43:36 [48789] Outgoing: <system> Jun 17 10:43:36 [48789] Outgoing: <host-name>ex4200-13</host-name> Jun 17 10:43:36 [48789] Outgoing: <domain-name>poc-nl.jnpr.net</domain-name> Jun 17 10:43:36 [48789] Outgoing: <time-zone>Europe/Amsterdam</time-zone> Jun 17 10:43:36 [48789] Outgoing: <authentication-order>radius</authentication-order> Jun 17 10:43:36 [48789] Outgoing: <authentication-order>password</authentication-order> Jun 17 10:43:36 [48789] Outgoing: <root-authentication> Jun 17 10:43:36 [48789] Outgoing: <encrypted-password>$1$/NHg28eO$pqaVlLlPQ2thlQQ0ZB.Vx/</encrypted-password> Jun 17 10:43:36 [48789] Outgoing: </root-authentication> Jun 17 10:43:36 [48789] Outgoing: <name-server> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.179.2</name> Jun 17 10:43:36 [48789] Outgoing: </name-server> Jun 17 10:43:36 [48789] Outgoing: <name-server> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.179.3</name> Jun 17 10:43:36 [48789] Outgoing: </name-server> Jun 17 10:43:36 [48789] Outgoing: <radius-server> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.176.9</name> Jun 17 10:43:36 [48789] Outgoing: <secret>$9$DMHPTz36CtOqmBEclLXik.mfT6/t1Eyn/</secret> Jun 17 10:43:36 [48789] Outgoing: <retry>3</retry> Jun 17 10:43:36 [48789] Outgoing: </radius-server> Jun 17 10:43:36 [48789] Outgoing: <radius-server> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.176.4</name> Jun 17 10:43:36 [48789] Outgoing: <secret>$9$CgY9p1EcylvWx0B7VwgUDtuOBIEleWNVYre</secret> Jun 17 10:43:36 [48789] Outgoing: <retry>3</retry> Jun 17 10:43:36 [48789] Outgoing: </radius-server> Jun 17 10:43:36 [48789] Outgoing: <login> Jun 17 10:43:36 [48789] Outgoing: <user> Jun 17 10:43:36 [48789] Outgoing: <name>pytraining-13</name> Jun 17 10:43:36 [48789] Outgoing: <uid>2001</uid> Jun 17 10:43:36 [48789] Outgoing: <class>super-user</class> Jun 17 10:43:36 [48789] Outgoing: <authentication> Jun 17 10:43:36 [48789] Outgoing: <encrypted-password>$1$uH0eoaXS$tTd6k7k.AVkEDhdJ8V75F.</encrypted-password> Jun 17 10:43:36 [48789] Outgoing: </authentication> Jun 17 10:43:36 [48789] Outgoing: </user> Jun 17 10:43:36 [48789] Outgoing: <user> Jun 17 10:43:36 [48789] Outgoing: <name>remote</name> Jun 17 10:43:36 [48789] Outgoing: <uid>2000</uid> Jun 17 10:43:36 [48789] Outgoing: <class>super-user</class> Jun 17 10:43:36 [48789] Outgoing: </user> Jun 17 10:43:36 [48789] Outgoing: </login> Jun 17 10:43:36 [48789] Outgoing: <services> Jun 17 10:43:36 [48789] Outgoing: <ftp> Jun 17 10:43:36 [48789] Outgoing: </ftp> Jun 17 10:43:36 [48789] Outgoing: <ssh> Jun 17 10:43:36 [48789] Outgoing: <client-alive-interval>120</client-alive-interval> Jun 17 10:43:36 [48789] Outgoing: </ssh> Jun 17 10:43:36 [48789] Outgoing: <telnet> Jun 17 10:43:36 [48789] Outgoing: </telnet> Jun 17 10:43:36 [48789] Outgoing: <xnm-clear-text> Jun 17 10:43:36 [48789] Outgoing: </xnm-clear-text> Jun 17 10:43:36 [48789] Outgoing: <netconf> Jun 17 10:43:36 [48789] Outgoing: <ssh> Jun 17 10:43:36 [48789] Outgoing: </ssh> Jun 17 10:43:36 [48789] Outgoing: <traceoptions> Jun 17 10:43:36 [48789] Outgoing: <file> Jun 17 10:43:36 [48789] Outgoing: <filename>netconf-ops.log</filename> Jun 17 10:43:36 [48789] Outgoing: <size>3m</size> Jun 17 10:43:36 [48789] Outgoing: <files>20</files> Jun 17 10:43:36 [48789] Outgoing: <world-readable/> Jun 17 10:43:36 [48789] Outgoing: </file> Jun 17 10:43:36 [48789] Outgoing: <flag> Jun 17 10:43:36 [48789] Outgoing: <name>all</name> Jun 17 10:43:36 [48789] Outgoing: </flag> Jun 17 10:43:36 [48789] Outgoing: </traceoptions> Jun 17 10:43:36 [48789] Outgoing: </netconf> Jun 17 10:43:36 [48789] Outgoing: <web-management> Jun 17 10:43:36 [48789] Outgoing: <http> Jun 17 10:43:36 [48789] Outgoing: </http> Jun 17 10:43:36 [48789] Outgoing: </web-management> Jun 17 10:43:36 [48789] Outgoing: </services> Jun 17 10:43:36 [48789] Outgoing: <syslog> Jun 17 10:43:36 [48789] Outgoing: <user> Jun 17 10:43:36 [48789] Outgoing: <name>*</name> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>any</name> Jun 17 10:43:36 [48789] Outgoing: <emergency/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: </user> Jun 17 10:43:36 [48789] Outgoing: <host> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.189.13</name> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>any</name> Jun 17 10:43:36 [48789] Outgoing: <notice/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>authorization</name> Jun 17 10:43:36 [48789] Outgoing: <info/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>interactive-commands</name> Jun 17 10:43:36 [48789] Outgoing: <info/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: </host> Jun 17 10:43:36 [48789] Outgoing: <host> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.189.14</name> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>any</name> Jun 17 10:43:36 [48789] Outgoing: <notice/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>authorization</name> Jun 17 10:43:36 [48789] Outgoing: <info/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>interactive-commands</name> Jun 17 10:43:36 [48789] Outgoing: <info/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: </host> Jun 17 10:43:36 [48789] Outgoing: <file> Jun 17 10:43:36 [48789] Outgoing: <name>messages</name> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>any</name> Jun 17 10:43:36 [48789] Outgoing: <notice/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: <contents> Jun 17 10:43:36 [48789] Outgoing: <name>authorization</name> Jun 17 10:43:36 [48789] Outgoing: <info/> Jun 17 10:43:36 [48789] Outgoing: </contents> Jun 17 10:43:36 [48789] Outgoing: </file> Jun 17 10:43:36 [48789] Outgoing: </syslog> Jun 17 10:43:36 [48789] Outgoing: <compress-configuration-files/> Jun 17 10:43:36 [48789] Outgoing: <commit> Jun 17 10:43:36 [48789] Outgoing: <synchronize/> Jun 17 10:43:36 [48789] Outgoing: </commit> Jun 17 10:43:36 [48789] Outgoing: <ntp> Jun 17 10:43:36 [48789] Outgoing: <boot-server>172.30.179.3</boot-server> Jun 17 10:43:36 [48789] Outgoing: <server> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.179.3</name> Jun 17 10:43:36 [48789] Outgoing: </server> Jun 17 10:43:36 [48789] Outgoing: <server> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.179.2</name> Jun 17 10:43:36 [48789] Outgoing: </server> Jun 17 10:43:36 [48789] Outgoing: </ntp> Jun 17 10:43:36 [48789] Outgoing: </system> Jun 17 10:43:36 [48789] Outgoing: <interfaces> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>ge-0/0/0</name> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <inet> Jun 17 10:43:36 [48789] Outgoing: <address> Jun 17 10:43:36 [48789] Outgoing: <name>10.0.0.2/24</name> Jun 17 10:43:36 [48789] Outgoing: </address> Jun 17 10:43:36 [48789] Outgoing: </inet> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>lo0</name> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <inet> Jun 17 10:43:36 [48789] Outgoing: <address> Jun 17 10:43:36 [48789] Outgoing: <name>192.179.0.113/32</name> Jun 17 10:43:36 [48789] Outgoing: <preferred/> Jun 17 10:43:36 [48789] Outgoing: </address> Jun 17 10:43:36 [48789] Outgoing: <address> Jun 17 10:43:36 [48789] Outgoing: <name>127.0.0.1/32</name> Jun 17 10:43:36 [48789] Outgoing: </address> Jun 17 10:43:36 [48789] Outgoing: </inet> Jun 17 10:43:36 [48789] Outgoing: <iso> Jun 17 10:43:36 [48789] Outgoing: <address> Jun 17 10:43:36 [48789] Outgoing: <name>49.0179.0000.0000.0113.00</name> Jun 17 10:43:36 [48789] Outgoing: </address> Jun 17 10:43:36 [48789] Outgoing: </iso> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: <interface> Jun 17 10:43:36 [48789] Outgoing: <name>me0</name> Jun 17 10:43:36 [48789] Outgoing: <unit> Jun 17 10:43:36 [48789] Outgoing: <name>0</name> Jun 17 10:43:36 [48789] Outgoing: <family> Jun 17 10:43:36 [48789] Outgoing: <inet> Jun 17 10:43:36 [48789] Outgoing: <address> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.179.113/24</name> Jun 17 10:43:36 [48789] Outgoing: </address> Jun 17 10:43:36 [48789] Outgoing: </inet> Jun 17 10:43:36 [48789] Outgoing: </family> Jun 17 10:43:36 [48789] Outgoing: </unit> Jun 17 10:43:36 [48789] Outgoing: </interface> Jun 17 10:43:36 [48789] Outgoing: </interfaces> Jun 17 10:43:36 [48789] Outgoing: <snmp> Jun 17 10:43:36 [48789] Outgoing: <location>AMS, EPOC 3.18</location> Jun 17 10:43:36 [48789] Outgoing: <contact>emea-poc@juniper.net</contact> Jun 17 10:43:36 [48789] Outgoing: <community> Jun 17 10:43:36 [48789] Outgoing: <name>public</name> Jun 17 10:43:36 [48789] Outgoing: <authorization>read-only</authorization> Jun 17 10:43:36 [48789] Outgoing: <clients> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.0.0/16</name> Jun 17 10:43:36 [48789] Outgoing: </clients> Jun 17 10:43:36 [48789] Outgoing: <clients> Jun 17 10:43:36 [48789] Outgoing: <name>0.0.0.0/0</name> Jun 17 10:43:36 [48789] Outgoing: <restrict/> Jun 17 10:43:36 [48789] Outgoing: </clients> Jun 17 10:43:36 [48789] Outgoing: </community> Jun 17 10:43:36 [48789] Outgoing: <community> Jun 17 10:43:36 [48789] Outgoing: <name>private</name> Jun 17 10:43:36 [48789] Outgoing: <authorization>read-write</authorization> Jun 17 10:43:36 [48789] Outgoing: <clients> Jun 17 10:43:36 [48789] Outgoing: <name>172.30.0.0/16</name> Jun 17 10:43:36 [48789] Outgoing: </clients> Jun 17 10:43:36 [48789] Outgoing: <clients> Jun 17 10:43:36 [48789] Outgoing: <name>0.0.0.0/0</name> Jun 17 10:43:36 [48789] Outgoing: <restrict/> Jun 17 10:43:36 [48789] Outgoing: </clients> Jun 17 10:43:36 [48789] Outgoing: </community> Jun 17 10:43:36 [48789] Outgoing: <trap-options> Jun 17 10:43:36 [48789] Outgoing: <source-address> Jun 17 10:43:36 [48789] Outgoing: <address>172.30.179.113</address> Jun 17 10:43:36 [48789] Outgoing: </source-address> Jun 17 10:43:36 [48789] Outgoing: </trap-options> Jun 17 10:43:36 [48789] Outgoing: </snmp> Jun 17 10:43:36 [48789] Outgoing: <routing-options> Jun 17 10:43:36 [48789] Outgoing: <static> Jun 17 10:43:36 [48789] Outgoing: <route> Jun 17 10:43:36 [48789] Outgoing: <name>0.0.0.0/0</name> Jun 17 10:43:36 [48789] Outgoing: <next-hop>172.30.179.1</next-hop> Jun 17 10:43:36 [48789] Outgoing: <no-readvertise/> Jun 17 10:43:36 [48789] Outgoing: </route> Jun 17 10:43:36 [48789] Outgoing: </static> Jun 17 10:43:36 [48789] Outgoing: <router-id>192.179.0.113</router-id> Jun 17 10:43:36 [48789] Outgoing: <autonomous-system> Jun 17 10:43:36 [48789] Outgoing: <as-number>100</as-number> Jun 17 10:43:36 [48789] Outgoing: </autonomous-system> Jun 17 10:43:36 [48789] Outgoing: </routing-options> Jun 17 10:43:36 [48789] Outgoing: <vlans> Jun 17 10:43:36 [48789] Outgoing: <vlan> Jun 17 10:43:36 [48789] Outgoing: <name>test1</name> Jun 17 10:43:36 [48789] Outgoing: <vlan-id>1111</vlan-id> Jun 17 10:43:36 [48789] Outgoing: </vlan> Jun 17 10:43:36 [48789] Outgoing: <vlan> Jun 17 10:43:36 [48789] Outgoing: <name>test2</name> Jun 17 10:43:36 [48789] Outgoing: <vlan-id>1112</vlan-id> Jun 17 10:43:36 [48789] Outgoing: </vlan> Jun 17 10:43:36 [48789] Outgoing: <vlan> Jun 17 10:43:36 [48789] Outgoing: <name>test3</name> Jun 17 10:43:36 [48789] Outgoing: <vlan-id>1113</vlan-id> Jun 17 10:43:36 [48789] Outgoing: </vlan> Jun 17 10:43:36 [48789] Outgoing: </vlans> Jun 17 10:43:36 [48789] Outgoing: </configuration> Jun 17 10:43:36 [48789] Outgoing: </rpc-reply> Jun 17 10:43:36 [48789] Outgoing: ]]>]]> Jun 17 10:43:36 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94c9e182-3467-11e6-bdb3-000c29432de4"><lock-configuration/></nc:rpc>]]>]]> Jun 17 10:43:36 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94c9e182-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:36 [48789] Outgoing: </rpc-reply> Jun 17 10:43:36 [48789] Outgoing: ]]>]]> Jun 17 10:43:36 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94dc9a8e-3467-11e6-bdb3-000c29432de4"><load-configuration action="set" format="text"><configuration-set>set system host-name newname</configuration-set></load-configuration></nc:rpc>]]>]]> Jun 17 10:43:36 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94dc9a8e-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:36 [48789] Outgoing: <load-configuration-results> Jun 17 10:43:36 [48789] Outgoing: <ok/> Jun 17 10:43:36 [48789] Outgoing: </load-configuration-results> Jun 17 10:43:36 [48789] Outgoing: </rpc-reply> Jun 17 10:43:36 [48789] Outgoing: ]]>]]> Jun 17 10:43:37 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94f6f528-3467-11e6-bdb3-000c29432de4"><get-configuration compare="rollback" rollback="0" format="text"/></nc:rpc>]]>]]> Jun 17 10:43:37 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:94f6f528-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:37 [48789] Outgoing: <configuration-information> Jun 17 10:43:37 [48789] Outgoing: <configuration-output> Jun 17 10:43:37 [48789] Outgoing: [edit system] Jun 17 10:43:37 [48789] Outgoing: - host-name ex4200-13; Jun 17 10:43:37 [48789] Outgoing: + host-name newname; Jun 17 10:43:37 [48789] Outgoing: </configuration-output> Jun 17 10:43:37 [48789] Outgoing: </configuration-information> Jun 17 10:43:37 [48789] Outgoing: </rpc-reply> Jun 17 10:43:37 [48789] Outgoing: ]]>]]> Jun 17 10:43:37 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:9565a07c-3467-11e6-bdb3-000c29432de4"><commit-configuration><check/></commit-configuration></nc:rpc>]]>]]> Jun 17 10:43:37 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:9565a07c-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:37 [48789] Outgoing: <commit-results> Jun 17 10:43:39 [48789] Outgoing: <routing-engine junos:style="normal"> Jun 17 10:43:39 [48789] Outgoing: <name>fpc0</name> Jun 17 10:43:39 [48789] Outgoing: <commit-check-success/> Jun 17 10:43:39 [48789] Outgoing: </routing-engine> Jun 17 10:43:39 [48789] Outgoing: </commit-results> Jun 17 10:43:39 [48789] Outgoing: <ok/> Jun 17 10:43:39 [48789] Outgoing: </rpc-reply> Jun 17 10:43:39 [48789] Outgoing: ]]>]]> Jun 17 10:43:39 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:96662b2c-3467-11e6-bdb3-000c29432de4"><commit-configuration><confirm-timeout>1</confirm-timeout><confirmed/><log>update config</log></commit-configuration></nc:rpc>]]>]]> Jun 17 10:43:39 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:96662b2c-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:39 [48789] Outgoing: <commit-results> Jun 17 10:43:41 [48789] Outgoing: <routing-engine junos:style="normal"> Jun 17 10:43:41 [48789] Outgoing: <name>fpc0</name> Jun 17 10:43:41 [48789] Outgoing: <commit-check-success/> Jun 17 10:43:44 [48789] Outgoing: <commit-success/> Jun 17 10:43:44 [48789] Outgoing: </routing-engine> Jun 17 10:43:44 [48789] Outgoing: </commit-results> Jun 17 10:43:44 [48789] Outgoing: <ok/> Jun 17 10:43:44 [48789] Outgoing: </rpc-reply> Jun 17 10:43:44 [48789] Outgoing: ]]>]]> Jun 17 10:43:44 [48789] Incoming: <?xml version="1.0" encoding="UTF-8"?><nc:rpc xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:99986dc8-3467-11e6-bdb3-000c29432de4"><unlock-configuration/></nc:rpc>]]>]]> Jun 17 10:43:44 [48789] Outgoing: <rpc-reply xmlns="urn:ietf:params:xml:ns:netconf:base:1.0" xmlns:junos="http://xml.juniper.net/junos/14.1X53/junos" xmlns:nc="urn:ietf:params:xml:ns:netconf:base:1.0" message-id="urn:uuid:99986dc8-3467-11e6-bdb3-000c29432de4"> Jun 17 10:43:44 [48789] Outgoing: </rpc-reply> Jun 17 10:43:46 [48789] Outgoing: ]]>]]> Jun 17 10:43:46 [48789] Outgoing: <!-- session end at 2016-06-17 10:43:46 CEST --> ```
main
module junos config with the parameter lines delete the delete details doesnt apply issue type bug report component name junos config ansible version ksator ubuntu ansible training for junos ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides ksator ubuntu ansible training for junos configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables default os environment mention the os you are running ansible from and the os you are managing or say β€œn a” for anything that is not platform specific uname a linux ubuntu generic ubuntu smp fri jul utc gnu linux summary can not delete junos configuration lines on junos devices using the module junos config with the parameter lines delete the delete details doesnt apply steps to reproduce i used this playbook module junos config with the parameter lines and some lines starting wuth delete ansible playbook junos config playbook yml name core module junos config hosts chris test connection local gather facts no vars prompt name device password prompt device password private yes tasks name test junos config host inventory hostname username pytraining password device password lines delete system name server set system host name newname expected results the expected result with module junos config with the parameter lines with lines starting with delete is this module should remove these junos lines on remote devices with the playbook above this system name server should be removed from the junos configuration of the host actual results the playbook did not remove system name server on the host however the details of the set command set system host name newname has been added succesfully i ran this playbook with extra verbosity vvvv and i enabled netconf log on device here s the output log ksator ubuntu ansible training for junos ansible playbook junos config playbook yml i hosts vvvv using etc ansible ansible cfg as config file loaded callback default of type stdout playbook playbook yml plays in junos config playbook yml device password play task task path home ksator ansible training for junos junos config playbook yml establish local connection for user ksator exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmptlrdup to home ksator ansible tmp ansible tmp junos config exec bin sh c lang c lc all c lc messages c usr bin python home ksator ansible tmp ansible tmp junos config rm rf home ksator ansible tmp ansible tmp dev null sleep changed changed true diff prepared n n host name n host name newname n invocation module args comment update config confirm host lines password value specified in no log parameter port null provider null replace false rollback null ssh keyfile null timeout transport netconf username pytraining zeroize false module name junos config play recap ok changed unreachable failed ksator ubuntu ansible training for junos netconf logs on junos device pytraining show log netconf ops log jun started tracing session jun incoming urn ietf params netconf capability writable running urn ietf params netconf capability rollback on error urn liberouter params netconf capability power control urn ietf params netconf capability validate urn i jun incoming jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing access jun outgoing jun outgoing jun outgoing lt ge gt jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing access jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing lt xe gt jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing access jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing trunk jun outgoing jun outgoing jun outgoing lt ge gt jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing trunk jun outgoing jun outgoing all jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing lt xe gt jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing trunk jun outgoing jun outgoing all jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing poc nl jnpr net jun outgoing europe amsterdam jun outgoing radius jun outgoing password jun outgoing jun outgoing vx jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing pytraining jun outgoing jun outgoing super user jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing remote jun outgoing jun outgoing super user jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing netconf ops log jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing all jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing any jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing any jun outgoing jun outgoing jun outgoing jun outgoing authorization jun outgoing jun outgoing jun outgoing jun outgoing interactive commands jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing any jun outgoing jun outgoing jun outgoing jun outgoing authorization jun outgoing jun outgoing jun outgoing jun outgoing interactive commands jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing messages jun outgoing jun outgoing any jun outgoing jun outgoing jun outgoing jun outgoing authorization jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing ge jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing ams epoc jun outgoing emea poc juniper net jun outgoing jun outgoing public jun outgoing read only jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing private jun outgoing read write jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun incoming jun outgoing jun outgoing jun outgoing jun incoming set system host name newname jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun incoming jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing host name jun outgoing host name newname jun outgoing jun outgoing jun outgoing jun outgoing jun incoming jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun incoming update config jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun outgoing jun incoming jun outgoing jun outgoing jun outgoing jun outgoing
1
32,269
12,102,230,257
IssuesEvent
2020-04-20 16:21:58
spinnaker/spinnaker
https://api.github.com/repos/spinnaker/spinnaker
closed
Spinnaker 1.19.4 - service accounts created in Front50 not synced to Fiat
sig/security
### Issue Summary: Service accounts are not synced to Fiat ### Cloud Provider(s): Google Cloud Platform ### Environment: Spinnaker 1.19.4 running on GKE ### Feature Area: Service accounts, Fiat, Front50 ### Description: When creating a new service account, it is not synced over to Fiat. ### Steps to Reproduce: ```` /# curl -X POST \ > -H "Content-type: application/json" \ > -d '{ "name": "vml2@e-conomic.net", "memberOf": ["vml"] }' \ > http://spin-front50.spinnaker:8080/serviceAccounts {"name":"vml2@e-conomic.net","lastModified":1587138472000,"lastModifiedBy":"anonymous","memberOf":["vml"]} ``` After syncing the roles, this shows up as a 404 in fiat. We noticed this when trying to switch to managed service accounts. And this is a stable reproduce of the issue. ``` root@nginx-8586cf59-5tn8z:/# curl -XGET http://spin-fiat.spinnaker:7003/authorize/vml2@e-conomic.net { "timestamp" : 1587138814886, "status" : 404, "error" : "Not Found", "message" : "No message available" } ### Additional Details:
True
Spinnaker 1.19.4 - service accounts created in Front50 not synced to Fiat - ### Issue Summary: Service accounts are not synced to Fiat ### Cloud Provider(s): Google Cloud Platform ### Environment: Spinnaker 1.19.4 running on GKE ### Feature Area: Service accounts, Fiat, Front50 ### Description: When creating a new service account, it is not synced over to Fiat. ### Steps to Reproduce: ```` /# curl -X POST \ > -H "Content-type: application/json" \ > -d '{ "name": "vml2@e-conomic.net", "memberOf": ["vml"] }' \ > http://spin-front50.spinnaker:8080/serviceAccounts {"name":"vml2@e-conomic.net","lastModified":1587138472000,"lastModifiedBy":"anonymous","memberOf":["vml"]} ``` After syncing the roles, this shows up as a 404 in fiat. We noticed this when trying to switch to managed service accounts. And this is a stable reproduce of the issue. ``` root@nginx-8586cf59-5tn8z:/# curl -XGET http://spin-fiat.spinnaker:7003/authorize/vml2@e-conomic.net { "timestamp" : 1587138814886, "status" : 404, "error" : "Not Found", "message" : "No message available" } ### Additional Details:
non_main
spinnaker service accounts created in not synced to fiat issue summary service accounts are not synced to fiat cloud provider s google cloud platform environment spinnaker running on gke feature area service accounts fiat description when creating a new service account it is not synced over to fiat steps to reproduce curl x post h content type application json d name e conomic net memberof name e conomic net lastmodified lastmodifiedby anonymous memberof after syncing the roles this shows up as a in fiat we noticed this when trying to switch to managed service accounts and this is a stable reproduce of the issue root nginx curl xget timestamp status error not found message no message available additional details
0
138,206
20,372,764,474
IssuesEvent
2022-02-21 12:53:31
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
closed
Template part transforms
Needs Design Feedback [Type] Discussion [Block] Template Part
I think it may be worth discussing and exploring the transform options for template parts. Currently the transform menu includes options to transform template parts in to Columns or Group blocks: <img width="392" alt="Screenshot 2021-02-23 at 10 14 43" src="https://user-images.githubusercontent.com/846565/108829485-229fca00-75c0-11eb-9dd3-1be96c741242.png"> However, neither of this options work. Selecting one just kicks you back to the wp-admin Dashboard πŸ› Should it be possible to transform template parts into other, non-template-part blocks at all? This seems like a potentially dangerous operation that might be better left to the "Detach blocks from template part" option in the ellipsis menu: <img width="554" alt="Screenshot 2021-02-23 at 10 20 46" src="https://user-images.githubusercontent.com/846565/108830037-d0ab7400-75c0-11eb-8958-6c42456733b3.png"> --- One transformation that should be possible in one way or another is switching template part variations. In previous issues the following design has been posited as an option to do this: <img src="https://user-images.githubusercontent.com/846565/106589428-eb4f7780-6543-11eb-9ec6-f8e81d790a69.png"> --- Closely related – in #28737 we are exploring how patterns that are contextually relevant to the selected template might be exposed via the transform menu. --- Finally, in https://github.com/WordPress/gutenberg/pull/27397#issuecomment-783315207 @mtias questioned whether it should be possible to quickly/easily wrap a template part inside another block. As @jasmussen mentioned, this could be useful for things like Sidebar template parts. My personal feeling is that this could possibly be better handled by the aforementioned patterns flow, as it is not a transform in the traditional sense, but it is worth discussing. ### Tentative action plan - [ ] Remove Group/Columns transform options on Template Part block – #29296 - [x] Add Template Part Switching - [ ] Provide a way to view block patterns that are contextually relevant to the selected Template Part – #28737 - [ ] Potentially create an affordance for template parts to be wrapped in other blocks
1.0
Template part transforms - I think it may be worth discussing and exploring the transform options for template parts. Currently the transform menu includes options to transform template parts in to Columns or Group blocks: <img width="392" alt="Screenshot 2021-02-23 at 10 14 43" src="https://user-images.githubusercontent.com/846565/108829485-229fca00-75c0-11eb-9dd3-1be96c741242.png"> However, neither of this options work. Selecting one just kicks you back to the wp-admin Dashboard πŸ› Should it be possible to transform template parts into other, non-template-part blocks at all? This seems like a potentially dangerous operation that might be better left to the "Detach blocks from template part" option in the ellipsis menu: <img width="554" alt="Screenshot 2021-02-23 at 10 20 46" src="https://user-images.githubusercontent.com/846565/108830037-d0ab7400-75c0-11eb-8958-6c42456733b3.png"> --- One transformation that should be possible in one way or another is switching template part variations. In previous issues the following design has been posited as an option to do this: <img src="https://user-images.githubusercontent.com/846565/106589428-eb4f7780-6543-11eb-9ec6-f8e81d790a69.png"> --- Closely related – in #28737 we are exploring how patterns that are contextually relevant to the selected template might be exposed via the transform menu. --- Finally, in https://github.com/WordPress/gutenberg/pull/27397#issuecomment-783315207 @mtias questioned whether it should be possible to quickly/easily wrap a template part inside another block. As @jasmussen mentioned, this could be useful for things like Sidebar template parts. My personal feeling is that this could possibly be better handled by the aforementioned patterns flow, as it is not a transform in the traditional sense, but it is worth discussing. ### Tentative action plan - [ ] Remove Group/Columns transform options on Template Part block – #29296 - [x] Add Template Part Switching - [ ] Provide a way to view block patterns that are contextually relevant to the selected Template Part – #28737 - [ ] Potentially create an affordance for template parts to be wrapped in other blocks
non_main
template part transforms i think it may be worth discussing and exploring the transform options for template parts currently the transform menu includes options to transform template parts in to columns or group blocks img width alt screenshot at src however neither of this options work selecting one just kicks you back to the wp admin dashboard πŸ› should it be possible to transform template parts into other non template part blocks at all this seems like a potentially dangerous operation that might be better left to the detach blocks from template part option in the ellipsis menu img width alt screenshot at src one transformation that should be possible in one way or another is switching template part variations in previous issues the following design has been posited as an option to do this img src closely related – in we are exploring how patterns that are contextually relevant to the selected template might be exposed via the transform menu finally in mtias questioned whether it should be possible to quickly easily wrap a template part inside another block as jasmussen mentioned this could be useful for things like sidebar template parts my personal feeling is that this could possibly be better handled by the aforementioned patterns flow as it is not a transform in the traditional sense but it is worth discussing tentative action plan remove group columns transform options on template part block – add template part switching provide a way to view block patterns that are contextually relevant to the selected template part –  potentially create an affordance for template parts to be wrapped in other blocks
0
5,704
30,081,740,273
IssuesEvent
2023-06-29 04:30:16
Apple-Dev-Projects/RoamX
https://api.github.com/repos/Apple-Dev-Projects/RoamX
closed
Configure CI/CD build action
maintain
Setup gated check-ins and create hosted build action for repository
True
Configure CI/CD build action - Setup gated check-ins and create hosted build action for repository
main
configure ci cd build action setup gated check ins and create hosted build action for repository
1
5,808
30,760,382,295
IssuesEvent
2023-07-29 16:13:10
NIAEFEUP/website-niaefeup-frontend
https://api.github.com/repos/NIAEFEUP/website-niaefeup-frontend
closed
global: Implement comprehensive color system
maintainability
So far, the custom colors added lack consistent and clear names, and would benefit from a more comprehensive system prepared for future additions and large number of elements. We should agree on a convention to implement for this matter.
True
global: Implement comprehensive color system - So far, the custom colors added lack consistent and clear names, and would benefit from a more comprehensive system prepared for future additions and large number of elements. We should agree on a convention to implement for this matter.
main
global implement comprehensive color system so far the custom colors added lack consistent and clear names and would benefit from a more comprehensive system prepared for future additions and large number of elements we should agree on a convention to implement for this matter
1
1,798
6,575,912,625
IssuesEvent
2017-09-11 17:48:35
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Add ability to handle snapshots to vsphere_guest module
affects_2.1 cloud feature_idea vmware waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> Original issue was https://github.com/ansible/ansible-modules-core/issues/2671. I used a wrong name in that ticket, I meant vsphere_guest. I was suggested by ansibot to re-open it here. ansibot commented 7 minutes ago @ronator: Thanks for taking the time to open this issue, but it was filed under the wrong github repository. Please create a new issue at https://github.com/ansible/ansible-modules-core/issues ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> vsphere_guest ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ansible 2.1.1.0 ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say β€œN/A” for anything that is not platform-specific. --> "N/A" ##### SUMMARY <!--- Explain the problem briefly --> The vmware_guest module lacks the ability to handle (list/create/remove) **snapshots** of vmware guests. This would be a killer feature! It would also allow more automization of the 'snapshot_to_clone' option, because for now using this option, snapshot must already be taken in vCenter.
True
Add ability to handle snapshots to vsphere_guest module - <!--- Verify first that your issue/request is not already reported in GitHub --> Original issue was https://github.com/ansible/ansible-modules-core/issues/2671. I used a wrong name in that ticket, I meant vsphere_guest. I was suggested by ansibot to re-open it here. ansibot commented 7 minutes ago @ronator: Thanks for taking the time to open this issue, but it was filed under the wrong github repository. Please create a new issue at https://github.com/ansible/ansible-modules-core/issues ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> vsphere_guest ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ansible 2.1.1.0 ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say β€œN/A” for anything that is not platform-specific. --> "N/A" ##### SUMMARY <!--- Explain the problem briefly --> The vmware_guest module lacks the ability to handle (list/create/remove) **snapshots** of vmware guests. This would be a killer feature! It would also allow more automization of the 'snapshot_to_clone' option, because for now using this option, snapshot must already be taken in vCenter.
main
add ability to handle snapshots to vsphere guest module original issue was i used a wrong name in that ticket i meant vsphere guest i was suggested by ansibot to re open it here ansibot commented minutes ago ronator thanks for taking the time to open this issue but it was filed under the wrong github repository please create a new issue at issue type feature idea component name vsphere guest ansible version ansible configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say β€œn a” for anything that is not platform specific n a summary the vmware guest module lacks the ability to handle list create remove snapshots of vmware guests this would be a killer feature it would also allow more automization of the snapshot to clone option because for now using this option snapshot must already be taken in vcenter
1
2,717
9,548,733,665
IssuesEvent
2019-05-02 06:50:01
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
reopened
Test classes are in same namespace as Type under Test
Area: analyzer Area: maintainability feature
Its a good practice to place the tests for a type in the exact same namespace as the type's namespace.
True
Test classes are in same namespace as Type under Test - Its a good practice to place the tests for a type in the exact same namespace as the type's namespace.
main
test classes are in same namespace as type under test its a good practice to place the tests for a type in the exact same namespace as the type s namespace
1
35,018
9,530,867,918
IssuesEvent
2019-04-29 14:47:33
w3f/polkadot-wiki
https://api.github.com/repos/w3f/polkadot-wiki
closed
Guide/Script: Docker file for Polkadot local testnet
build
We need some way for people to be able to play around with Polkadot through deploying parachains / creating referendums without giving them a proportionally high level of DOTs and potentially letting them create arbitrary governance proposals. The best idea is to create a docker container which would start a local Polkadot testnet for development purposes and allow the Polkadot UI to hook into this. Polkadot/Substrate has a developer flag for its chain option but as of right now there is no script which is taking advantage of it. The docker container / script should be accompanied by a short guide which shows how to pull down the container and start the Polkadot testnet locally. In future guides we can build from this one by doing things such as creating referendums and deploying parachains. As parachain development becomes more real (with the maturation of cumulus etc) this will be an easy environment for devs to get their feet wet through the process of parachain deployment.
1.0
Guide/Script: Docker file for Polkadot local testnet - We need some way for people to be able to play around with Polkadot through deploying parachains / creating referendums without giving them a proportionally high level of DOTs and potentially letting them create arbitrary governance proposals. The best idea is to create a docker container which would start a local Polkadot testnet for development purposes and allow the Polkadot UI to hook into this. Polkadot/Substrate has a developer flag for its chain option but as of right now there is no script which is taking advantage of it. The docker container / script should be accompanied by a short guide which shows how to pull down the container and start the Polkadot testnet locally. In future guides we can build from this one by doing things such as creating referendums and deploying parachains. As parachain development becomes more real (with the maturation of cumulus etc) this will be an easy environment for devs to get their feet wet through the process of parachain deployment.
non_main
guide script docker file for polkadot local testnet we need some way for people to be able to play around with polkadot through deploying parachains creating referendums without giving them a proportionally high level of dots and potentially letting them create arbitrary governance proposals the best idea is to create a docker container which would start a local polkadot testnet for development purposes and allow the polkadot ui to hook into this polkadot substrate has a developer flag for its chain option but as of right now there is no script which is taking advantage of it the docker container script should be accompanied by a short guide which shows how to pull down the container and start the polkadot testnet locally in future guides we can build from this one by doing things such as creating referendums and deploying parachains as parachain development becomes more real with the maturation of cumulus etc this will be an easy environment for devs to get their feet wet through the process of parachain deployment
0
3,370
13,041,800,285
IssuesEvent
2020-07-28 21:05:05
short-d/short
https://api.github.com/repos/short-d/short
closed
[Refactor] Extract out Error Modal into a separate component
good first issue maintainability p1
Currently, error modal code is nested inside Home page component (https://github.com/short-d/short/blob/master/frontend/src/component/pages/Home.tsx#L290-L308) which increased the complexity of home page. Let's extract it out into a separate component so that it is easier to read and understand the core logic of home page.
True
[Refactor] Extract out Error Modal into a separate component - Currently, error modal code is nested inside Home page component (https://github.com/short-d/short/blob/master/frontend/src/component/pages/Home.tsx#L290-L308) which increased the complexity of home page. Let's extract it out into a separate component so that it is easier to read and understand the core logic of home page.
main
extract out error modal into a separate component currently error modal code is nested inside home page component which increased the complexity of home page let s extract it out into a separate component so that it is easier to read and understand the core logic of home page
1
780,951
27,415,897,593
IssuesEvent
2023-03-01 13:46:04
MAIF/otoroshi
https://api.github.com/repos/MAIF/otoroshi
closed
remove old directories
roadmap-2023 priority/P1 size/S
- [x] archives - [x] clients - [x] connectors - [x] test-resources - [x] move useful client to a `tools` directory - [x] move wasm-manager to a `tools` directory
1.0
remove old directories - - [x] archives - [x] clients - [x] connectors - [x] test-resources - [x] move useful client to a `tools` directory - [x] move wasm-manager to a `tools` directory
non_main
remove old directories archives clients connectors test resources move useful client to a tools directory move wasm manager to a tools directory
0
555,894
16,472,236,706
IssuesEvent
2021-05-23 16:44:14
magento/magento2
https://api.github.com/repos/magento/magento2
closed
Vertical "ice skating" in zoom function in Chrome on product page with large image
Component: Framework/Image Event: cdkyiv2019 Issue: Confirmed Priority: P4 Progress: ready for dev Reproduced on 2.3.x Severity: S4 stale issue
### Preconditions (*) * Magento 2.2.7, but code in question has not been changed in 2.3.x * Chrome 72 ### Steps to reproduce (*) 1. Open product page with a large image included. 2. Open full-screen zoom 3. Zoom in all the way so that the image is larger than the window (shrink browser window if need be) 4. Click and drag image vertically, and then without letting go of mouse, reverse direction ### Expected result (*) Image pans in proportion to mouse movement. Image pan directly reverses course at the same time the mouse reverses course. ### Actual result (*) Image pans faster than mouse movement. Image pan continues in initial direction before later reversing course. See video: https://www.useloom.com/share/4ec4573b853c44a58999d464f5c21edd ### Analysis This issue only affects vertical panning, not horizontal panning. This is because of a new buggy approach that was introduced in db8476afd44fa54094d6d3453d14600cb5fc01f5 only for vertical panning, leaving the original approach for horizontal panning. This commit seems it was aimed at fixing #12704 (commit message matches that description), but I'm not sure since the timing doesn't quite match up.
1.0
Vertical "ice skating" in zoom function in Chrome on product page with large image - ### Preconditions (*) * Magento 2.2.7, but code in question has not been changed in 2.3.x * Chrome 72 ### Steps to reproduce (*) 1. Open product page with a large image included. 2. Open full-screen zoom 3. Zoom in all the way so that the image is larger than the window (shrink browser window if need be) 4. Click and drag image vertically, and then without letting go of mouse, reverse direction ### Expected result (*) Image pans in proportion to mouse movement. Image pan directly reverses course at the same time the mouse reverses course. ### Actual result (*) Image pans faster than mouse movement. Image pan continues in initial direction before later reversing course. See video: https://www.useloom.com/share/4ec4573b853c44a58999d464f5c21edd ### Analysis This issue only affects vertical panning, not horizontal panning. This is because of a new buggy approach that was introduced in db8476afd44fa54094d6d3453d14600cb5fc01f5 only for vertical panning, leaving the original approach for horizontal panning. This commit seems it was aimed at fixing #12704 (commit message matches that description), but I'm not sure since the timing doesn't quite match up.
non_main
vertical ice skating in zoom function in chrome on product page with large image preconditions magento but code in question has not been changed in x chrome steps to reproduce open product page with a large image included open full screen zoom zoom in all the way so that the image is larger than the window shrink browser window if need be click and drag image vertically and then without letting go of mouse reverse direction expected result image pans in proportion to mouse movement image pan directly reverses course at the same time the mouse reverses course actual result image pans faster than mouse movement image pan continues in initial direction before later reversing course see video analysis this issue only affects vertical panning not horizontal panning this is because of a new buggy approach that was introduced in only for vertical panning leaving the original approach for horizontal panning this commit seems it was aimed at fixing commit message matches that description but i m not sure since the timing doesn t quite match up
0
353,848
25,138,443,757
IssuesEvent
2022-11-09 20:44:01
dagster-io/dagster
https://api.github.com/repos/dagster-io/dagster
closed
[docs] - Add docs for env vars and secrets in Dagster Cloud
documentation
This issue serves to track research / content work for adding documentation about environment variables and secrets to Dagster Cloud. --- ### Message from the maintainers Impacted by this issue? Give it a πŸ‘! We factor engagement into prioritization.
1.0
[docs] - Add docs for env vars and secrets in Dagster Cloud - This issue serves to track research / content work for adding documentation about environment variables and secrets to Dagster Cloud. --- ### Message from the maintainers Impacted by this issue? Give it a πŸ‘! We factor engagement into prioritization.
non_main
add docs for env vars and secrets in dagster cloud this issue serves to track research content work for adding documentation about environment variables and secrets to dagster cloud message from the maintainers impacted by this issue give it a πŸ‘ we factor engagement into prioritization
0
50,609
3,006,510,347
IssuesEvent
2015-07-27 10:53:57
Itseez/opencv
https://api.github.com/repos/Itseez/opencv
opened
Niblack's adaptive thresholding
auto-transferred category: imgproc category: video feature priority: low
Transferred from http://code.opencv.org/issues/4024 ``` || Sancho McCann on 2014-11-24 19:46 || Priority: Low || Affected: None || Category: imgproc, video || Tracker: Feature || Difficulty: Medium || PR: https://github.com/Itseez/opencv_contrib/pull/181 || Platform: None / None ``` Niblack's adaptive thresholding ----------- ``` I'm interested in including Niblack's adaptive thresholding in OpenCV. Currently, the adaptive threshold is based on the mean within a block plus an offset that is a constant parameter. Niblack's threshold is based on the mean within a block plus an offset that is a constant parameter in [0,1] times the standard deviation of the pixel intensities within the block. ``` History ------- ##### Ilya Lavrenov on 2014-12-01 11:34 ``` Hi Sancho, Could you implement this type of adaptive thresholding? And send us an appropriate pull request. Thanks. - Target version changed from 2.4.11 to 3.0 - Difficulty changed from Easy to Medium - Assignee changed from Vadim Pisarevsky to Sancho McCann - Status changed from New to Open ``` ##### Steven Puttemans on 2015-03-03 13:19 ``` - Pull request set to https://github.com/Itseez/opencv_contrib/pull/181 ```
1.0
Niblack's adaptive thresholding - Transferred from http://code.opencv.org/issues/4024 ``` || Sancho McCann on 2014-11-24 19:46 || Priority: Low || Affected: None || Category: imgproc, video || Tracker: Feature || Difficulty: Medium || PR: https://github.com/Itseez/opencv_contrib/pull/181 || Platform: None / None ``` Niblack's adaptive thresholding ----------- ``` I'm interested in including Niblack's adaptive thresholding in OpenCV. Currently, the adaptive threshold is based on the mean within a block plus an offset that is a constant parameter. Niblack's threshold is based on the mean within a block plus an offset that is a constant parameter in [0,1] times the standard deviation of the pixel intensities within the block. ``` History ------- ##### Ilya Lavrenov on 2014-12-01 11:34 ``` Hi Sancho, Could you implement this type of adaptive thresholding? And send us an appropriate pull request. Thanks. - Target version changed from 2.4.11 to 3.0 - Difficulty changed from Easy to Medium - Assignee changed from Vadim Pisarevsky to Sancho McCann - Status changed from New to Open ``` ##### Steven Puttemans on 2015-03-03 13:19 ``` - Pull request set to https://github.com/Itseez/opencv_contrib/pull/181 ```
non_main
niblack s adaptive thresholding transferred from sancho mccann on priority low affected none category imgproc video tracker feature difficulty medium pr platform none none niblack s adaptive thresholding i m interested in including niblack s adaptive thresholding in opencv currently the adaptive threshold is based on the mean within a block plus an offset that is a constant parameter niblack s threshold is based on the mean within a block plus an offset that is a constant parameter in times the standard deviation of the pixel intensities within the block history ilya lavrenov on hi sancho could you implement this type of adaptive thresholding and send us an appropriate pull request thanks target version changed from to difficulty changed from easy to medium assignee changed from vadim pisarevsky to sancho mccann status changed from new to open steven puttemans on pull request set to
0
143,460
5,517,645,256
IssuesEvent
2017-03-18 00:39:17
SOCR/SOCRAT-issues
https://api.github.com/repos/SOCR/SOCRAT-issues
closed
Make Service / Page for Trellis Charts
component: charts priority: high
* have it be the default chart shown * but also not block bar charts as an issue (it doesn't like graph properly rn but there is at least the frame and, as explained in the journal, I spent a large enough chunk of time on this / thought adding a new file deserved an issue as a part of the epic of trellis?) <img width="275" alt="screen shot 2017-03-09 at 12 10 57 am" src="https://cloud.githubusercontent.com/assets/17788969/23761723/be457c76-04c1-11e7-9517-6589ef5d3624.png">
1.0
Make Service / Page for Trellis Charts - * have it be the default chart shown * but also not block bar charts as an issue (it doesn't like graph properly rn but there is at least the frame and, as explained in the journal, I spent a large enough chunk of time on this / thought adding a new file deserved an issue as a part of the epic of trellis?) <img width="275" alt="screen shot 2017-03-09 at 12 10 57 am" src="https://cloud.githubusercontent.com/assets/17788969/23761723/be457c76-04c1-11e7-9517-6589ef5d3624.png">
non_main
make service page for trellis charts have it be the default chart shown but also not block bar charts as an issue it doesn t like graph properly rn but there is at least the frame and as explained in the journal i spent a large enough chunk of time on this thought adding a new file deserved an issue as a part of the epic of trellis img width alt screen shot at am src
0
302,586
22,832,402,398
IssuesEvent
2022-07-12 13:58:25
milvus-io/milvus-operator
https://api.github.com/repos/milvus-io/milvus-operator
opened
Add markdown link check to avoid broken links
documentation good first issue
for example use: github-action-markdown-link-check@v1
1.0
Add markdown link check to avoid broken links - for example use: github-action-markdown-link-check@v1
non_main
add markdown link check to avoid broken links for example use github action markdown link check
0
43,879
23,411,836,937
IssuesEvent
2022-08-12 18:27:07
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
RuntimeType.GetGenericTypeDefinition() performance issue
tenet-performance
`RuntimeType.GetGenericTypeDefinition()` is a bit expensive. As in a Dictionary<type,type> cache lookup is faster. I would naively expect that the generic definition of a type is stored in a field of the constructed type and GetGenericTypeDefinition to be a simple field access.
True
RuntimeType.GetGenericTypeDefinition() performance issue - `RuntimeType.GetGenericTypeDefinition()` is a bit expensive. As in a Dictionary<type,type> cache lookup is faster. I would naively expect that the generic definition of a type is stored in a field of the constructed type and GetGenericTypeDefinition to be a simple field access.
non_main
runtimetype getgenerictypedefinition performance issue runtimetype getgenerictypedefinition is a bit expensive as in a dictionary cache lookup is faster i would naively expect that the generic definition of a type is stored in a field of the constructed type and getgenerictypedefinition to be a simple field access
0
253,195
8,052,639,554
IssuesEvent
2018-08-01 19:59:24
ARCH-commons/i2p-transform
https://api.github.com/repos/ARCH-commons/i2p-transform
closed
orphan providerids aren't legal
Oracle bug high priority
Getting DCQ error: "Orphan PROVIDERID(S) in the ENCOUNTER, DIAGNOSIS,MED_ADMIN, OBS_CLIN, PRESCRIBING, or PROCEDURES table." Will need to write a script to fill in the provider table.
1.0
orphan providerids aren't legal - Getting DCQ error: "Orphan PROVIDERID(S) in the ENCOUNTER, DIAGNOSIS,MED_ADMIN, OBS_CLIN, PRESCRIBING, or PROCEDURES table." Will need to write a script to fill in the provider table.
non_main
orphan providerids aren t legal getting dcq error orphan providerid s in the encounter diagnosis med admin obs clin prescribing or procedures table will need to write a script to fill in the provider table
0