Unnamed: 0
int64 1
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 3
438
| labels
stringlengths 4
308
| body
stringlengths 7
254k
| index
stringclasses 7
values | text_combine
stringlengths 96
254k
| label
stringclasses 2
values | text
stringlengths 96
246k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2,160
| 7,519,425,501
|
IssuesEvent
|
2018-04-12 11:34:33
|
RalfKoban/MiKo-Analyzers
|
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
|
closed
|
Interface methods should not return IList, IDictionary, ICollection
|
Area: analyzer Area: maintainability feature in progress
|
Interface methods or public class methods or properties should not get or return
- IList
- IDictionary
- ICollection
- ...
Instead they should get/return the read-only/immutable pendants.
Normally, you do not want to add or remove items or clear the list. So the interface should not allow to do so
|
True
|
Interface methods should not return IList, IDictionary, ICollection - Interface methods or public class methods or properties should not get or return
- IList
- IDictionary
- ICollection
- ...
Instead they should get/return the read-only/immutable pendants.
Normally, you do not want to add or remove items or clear the list. So the interface should not allow to do so
|
main
|
interface methods should not return ilist idictionary icollection interface methods or public class methods or properties should not get or return ilist idictionary icollection instead they should get return the read only immutable pendants normally you do not want to add or remove items or clear the list so the interface should not allow to do so
| 1
|
4,743
| 24,480,401,021
|
IssuesEvent
|
2022-10-08 18:59:24
|
centerofci/mathesar
|
https://api.github.com/repos/centerofci/mathesar
|
opened
|
Date input should close date picker when losing focus via Tab or Shift+Tab
|
type: bug work: frontend status: ready restricted: maintainers
|
## Steps to reproduce
1. Navigate to the Record Page for one Checkouts record.
1. Click on the "Checkout Time" input field to focus it.
1. Observe the date picker to appear (good).
1. Press Tab to focus on the Due Date input.
1. Expect the Checkout Time date picker to close.
1. Instead, observe the Checkout Time date picker remains open, even when the Due Date date picker has also opened.
CC @pavish
|
True
|
Date input should close date picker when losing focus via Tab or Shift+Tab - ## Steps to reproduce
1. Navigate to the Record Page for one Checkouts record.
1. Click on the "Checkout Time" input field to focus it.
1. Observe the date picker to appear (good).
1. Press Tab to focus on the Due Date input.
1. Expect the Checkout Time date picker to close.
1. Instead, observe the Checkout Time date picker remains open, even when the Due Date date picker has also opened.
CC @pavish
|
main
|
date input should close date picker when losing focus via tab or shift tab steps to reproduce navigate to the record page for one checkouts record click on the checkout time input field to focus it observe the date picker to appear good press tab to focus on the due date input expect the checkout time date picker to close instead observe the checkout time date picker remains open even when the due date date picker has also opened cc pavish
| 1
|
21,127
| 6,980,965,289
|
IssuesEvent
|
2017-12-13 05:14:09
|
hashicorp/packer
|
https://api.github.com/repos/hashicorp/packer
|
closed
|
Breaking change in 1.1.3
|
bug builder/amazon docs
|
[This commit](https://github.com/hashicorp/packer/commit/a90c45d9bb3f2abd56ea77c8a456df19baaa60a7#diff-76f53be4e00c8508514464a9b9235c4e) which made it into the `1.1.3` release introduces a dependency in AWS for the `ec2:DescribeInstanceStatus` permission on the role that is building AMI's.
This broke our pipelines which were previously working on `1.1.2`. (We used the docker `light` image, which wasn't pinned so it automatically put us on the latest release).
Anyway, the new permission it should probably be documented at https://www.packer.io/docs/builders/amazon.html#using-an-iam-task-or-instance-role
|
1.0
|
Breaking change in 1.1.3 - [This commit](https://github.com/hashicorp/packer/commit/a90c45d9bb3f2abd56ea77c8a456df19baaa60a7#diff-76f53be4e00c8508514464a9b9235c4e) which made it into the `1.1.3` release introduces a dependency in AWS for the `ec2:DescribeInstanceStatus` permission on the role that is building AMI's.
This broke our pipelines which were previously working on `1.1.2`. (We used the docker `light` image, which wasn't pinned so it automatically put us on the latest release).
Anyway, the new permission it should probably be documented at https://www.packer.io/docs/builders/amazon.html#using-an-iam-task-or-instance-role
|
non_main
|
breaking change in which made it into the release introduces a dependency in aws for the describeinstancestatus permission on the role that is building ami s this broke our pipelines which were previously working on we used the docker light image which wasn t pinned so it automatically put us on the latest release anyway the new permission it should probably be documented at
| 0
|
5,839
| 31,023,140,908
|
IssuesEvent
|
2023-08-10 07:17:03
|
jupyter-naas/awesome-notebooks
|
https://api.github.com/repos/jupyter-naas/awesome-notebooks
|
opened
|
Xero - Get Bank Transactions
|
templates maintainer
|
This notebook retrieves one or many bank transactions from Xero. It is usefull for organizations to keep track of their financial data.
|
True
|
Xero - Get Bank Transactions - This notebook retrieves one or many bank transactions from Xero. It is usefull for organizations to keep track of their financial data.
|
main
|
xero get bank transactions this notebook retrieves one or many bank transactions from xero it is usefull for organizations to keep track of their financial data
| 1
|
3,406
| 13,181,837,636
|
IssuesEvent
|
2020-08-12 14:54:03
|
duo-labs/cloudmapper
|
https://api.github.com/repos/duo-labs/cloudmapper
|
closed
|
Add command to show what resources are supported
|
map unmaintained_functionality
|
There isn't an easy way for people to know what resources CloudMapper supports for the different commands. Let's start by showing what `prepare` supports.
|
True
|
Add command to show what resources are supported - There isn't an easy way for people to know what resources CloudMapper supports for the different commands. Let's start by showing what `prepare` supports.
|
main
|
add command to show what resources are supported there isn t an easy way for people to know what resources cloudmapper supports for the different commands let s start by showing what prepare supports
| 1
|
3,994
| 18,509,642,327
|
IssuesEvent
|
2021-10-20 00:05:24
|
aws/aws-sam-cli
|
https://api.github.com/repos/aws/aws-sam-cli
|
closed
|
Cannot build from source under Ubuntu 20.04
|
platform/linux area/installation maintainer/need-response
|
<!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed).
If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. -->
### Description:
<!-- Briefly describe the bug you are facing.-->
I'm following instructions in `./installer/pyinstaller/Instruction.md`
I do have to `chmod +x ./installer/pyinstaller/build-linux.sh`
When I build inside the CentOS 7 image, per the instructions, I see that `libpython3.7-dev` is not provided. I've installed this on my host machine, and also installed this inside the CentOS docker image, no luck `yum install libpython3.7-dev.x86_64`.
### Steps to reproduce:
<!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) -->
```shell
docker run --mount type=bind,src="{Absolute path to AWS SAM CLI source}",dst="/aws-sam-cli" -it quay.io/pypa/manylinux2014_x86_64
cd aws-sam-cli
chmod +x ./installer/pyinstaller/build-linux.sh
./installer/pyinstaller/build-linux.sh aws-sam-cli-linux-x86_64.zip
```
### Observed result:
<!-- Please provide command output with `--debug` flag set. -->
Initially, an error saying that `libpython3.7` couldn't be found.
After running `yum install libpython3.7-dev.x86_64` in the container, I see `ERROR: No matching distribution found for pyrsistent==0.17.3`
### Expected result:
<!-- Describe what you expected. -->
Looks like a ZIP file should come out of this?
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: docker container for CentOS 7, per build instructions.
2. `sam --version`: N/A
3. AWS region: N/A
`Add --debug flag to command you are running`
|
True
|
Cannot build from source under Ubuntu 20.04 - <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed).
If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. -->
### Description:
<!-- Briefly describe the bug you are facing.-->
I'm following instructions in `./installer/pyinstaller/Instruction.md`
I do have to `chmod +x ./installer/pyinstaller/build-linux.sh`
When I build inside the CentOS 7 image, per the instructions, I see that `libpython3.7-dev` is not provided. I've installed this on my host machine, and also installed this inside the CentOS docker image, no luck `yum install libpython3.7-dev.x86_64`.
### Steps to reproduce:
<!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) -->
```shell
docker run --mount type=bind,src="{Absolute path to AWS SAM CLI source}",dst="/aws-sam-cli" -it quay.io/pypa/manylinux2014_x86_64
cd aws-sam-cli
chmod +x ./installer/pyinstaller/build-linux.sh
./installer/pyinstaller/build-linux.sh aws-sam-cli-linux-x86_64.zip
```
### Observed result:
<!-- Please provide command output with `--debug` flag set. -->
Initially, an error saying that `libpython3.7` couldn't be found.
After running `yum install libpython3.7-dev.x86_64` in the container, I see `ERROR: No matching distribution found for pyrsistent==0.17.3`
### Expected result:
<!-- Describe what you expected. -->
Looks like a ZIP file should come out of this?
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: docker container for CentOS 7, per build instructions.
2. `sam --version`: N/A
3. AWS region: N/A
`Add --debug flag to command you are running`
|
main
|
cannot build from source under ubuntu make sure we don t have an existing issue that reports the bug you are seeing both open and closed if you do find an existing issue re open or add a comment to that issue instead of creating a new one description i m following instructions in installer pyinstaller instruction md i do have to chmod x installer pyinstaller build linux sh when i build inside the centos image per the instructions i see that dev is not provided i ve installed this on my host machine and also installed this inside the centos docker image no luck yum install dev steps to reproduce shell docker run mount type bind src absolute path to aws sam cli source dst aws sam cli it quay io pypa cd aws sam cli chmod x installer pyinstaller build linux sh installer pyinstaller build linux sh aws sam cli linux zip observed result initially an error saying that couldn t be found after running yum install dev in the container i see error no matching distribution found for pyrsistent expected result looks like a zip file should come out of this additional environment details ex windows mac amazon linux etc os docker container for centos per build instructions sam version n a aws region n a add debug flag to command you are running
| 1
|
416,728
| 28,097,429,553
|
IssuesEvent
|
2023-03-30 16:47:00
|
WorldEnterpriseGroup/silkcorp
|
https://api.github.com/repos/WorldEnterpriseGroup/silkcorp
|
closed
|
Change the images of team member and content mentioned with images
|
documentation help wanted content writing copywriting
|
Website Link: https://silkcorp.org/
**Please Follow these Instructions:**
- Change all 4 images of team members
- Make an image having a message from Team member
- Change the image beside the message.
Find the attached image for reference.

|
1.0
|
Change the images of team member and content mentioned with images - Website Link: https://silkcorp.org/
**Please Follow these Instructions:**
- Change all 4 images of team members
- Make an image having a message from Team member
- Change the image beside the message.
Find the attached image for reference.

|
non_main
|
change the images of team member and content mentioned with images website link please follow these instructions change all images of team members make an image having a message from team member change the image beside the message find the attached image for reference
| 0
|
844
| 4,494,318,112
|
IssuesEvent
|
2016-08-31 05:27:07
|
ansible/ansible-modules-extras
|
https://api.github.com/repos/ansible/ansible-modules-extras
|
closed
|
gluster_volume module should parse "other names" of 'gluster peer status' command
|
feature_idea waiting_on_maintainer
|
Hi.
I have the following `gluster peer status` output:
```
Number of Peers: 2
Hostname: emma.int.prhv.afunix.org
Uuid: c536dce8-dda8-4a66-a840-2eea709572da
State: Peer in Cluster (Connected)
Other names:
emma
Hostname: amanda.int.prhv.afunix.org
Uuid: c6654b8f-3ca8-424f-8d0f-7c0b7d8108cc
State: Peer in Cluster (Connected)
Other names:
amanda
```
My inventory has short names configured (emma, amanda, ...).
When I try to configure a volume with gluster_volume module it fails.
Module cannot identify that `gluster peer probe` completed successfully, because it cannot match FQDN with short hostnames.
Gluster itself knows that peers have the specific short names, so should be able to the gluster_volume module.
|
True
|
gluster_volume module should parse "other names" of 'gluster peer status' command - Hi.
I have the following `gluster peer status` output:
```
Number of Peers: 2
Hostname: emma.int.prhv.afunix.org
Uuid: c536dce8-dda8-4a66-a840-2eea709572da
State: Peer in Cluster (Connected)
Other names:
emma
Hostname: amanda.int.prhv.afunix.org
Uuid: c6654b8f-3ca8-424f-8d0f-7c0b7d8108cc
State: Peer in Cluster (Connected)
Other names:
amanda
```
My inventory has short names configured (emma, amanda, ...).
When I try to configure a volume with gluster_volume module it fails.
Module cannot identify that `gluster peer probe` completed successfully, because it cannot match FQDN with short hostnames.
Gluster itself knows that peers have the specific short names, so should be able to the gluster_volume module.
|
main
|
gluster volume module should parse other names of gluster peer status command hi i have the following gluster peer status output number of peers hostname emma int prhv afunix org uuid state peer in cluster connected other names emma hostname amanda int prhv afunix org uuid state peer in cluster connected other names amanda my inventory has short names configured emma amanda when i try to configure a volume with gluster volume module it fails module cannot identify that gluster peer probe completed successfully because it cannot match fqdn with short hostnames gluster itself knows that peers have the specific short names so should be able to the gluster volume module
| 1
|
18,992
| 10,312,364,811
|
IssuesEvent
|
2019-08-29 19:40:50
|
jamijam/WebGoat-Legacy
|
https://api.github.com/repos/jamijam/WebGoat-Legacy
|
opened
|
CVE-2017-7525 (High) detected in jackson-databind-2.0.4.jar
|
security vulnerability
|
## CVE-2017-7525 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.0.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: /WebGoat-Legacy/pom.xml</p>
<p>Path to vulnerable library: 2/repository/com/fasterxml/jackson/core/jackson-databind/2.0.4/jackson-databind-2.0.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.0.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jamijam/WebGoat-Legacy/commit/47e282c80c0e0057166440fc7140b77c4f06feb6">47e282c80c0e0057166440fc7140b77c4f06feb6</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A deserialization flaw was discovered in the jackson-databind, versions before 2.6.7.1, 2.7.9.1 and 2.8.9, which could allow an unauthenticated user to perform code execution by sending the maliciously crafted input to the readValue method of the ObjectMapper.
<p>Publish Date: 2018-02-06
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7525>CVE-2017-7525</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-7525">https://nvd.nist.gov/vuln/detail/CVE-2017-7525</a></p>
<p>Release Date: 2018-02-06</p>
<p>Fix Resolution: 2.6.7.1,2.7.9.1,2.8.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-7525 (High) detected in jackson-databind-2.0.4.jar - ## CVE-2017-7525 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.0.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: /WebGoat-Legacy/pom.xml</p>
<p>Path to vulnerable library: 2/repository/com/fasterxml/jackson/core/jackson-databind/2.0.4/jackson-databind-2.0.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.0.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jamijam/WebGoat-Legacy/commit/47e282c80c0e0057166440fc7140b77c4f06feb6">47e282c80c0e0057166440fc7140b77c4f06feb6</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A deserialization flaw was discovered in the jackson-databind, versions before 2.6.7.1, 2.7.9.1 and 2.8.9, which could allow an unauthenticated user to perform code execution by sending the maliciously crafted input to the readValue method of the ObjectMapper.
<p>Publish Date: 2018-02-06
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7525>CVE-2017-7525</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-7525">https://nvd.nist.gov/vuln/detail/CVE-2017-7525</a></p>
<p>Release Date: 2018-02-06</p>
<p>Fix Resolution: 2.6.7.1,2.7.9.1,2.8.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_main
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to dependency file webgoat legacy pom xml path to vulnerable library repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details a deserialization flaw was discovered in the jackson databind versions before and which could allow an unauthenticated user to perform code execution by sending the maliciously crafted input to the readvalue method of the objectmapper publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
645
| 4,158,831,034
|
IssuesEvent
|
2016-06-17 05:55:36
|
StefMa/TimeTracking
|
https://api.github.com/repos/StefMa/TimeTracking
|
opened
|
Automatic release/publish to testes and Google Play
|
MAINTAINING
|
Currently we have to build the APK and release it by our self.
We have to create a script which shares the APK to testers (with the dev flavor) and to the play store with the prod flavor.
|
True
|
Automatic release/publish to testes and Google Play - Currently we have to build the APK and release it by our self.
We have to create a script which shares the APK to testers (with the dev flavor) and to the play store with the prod flavor.
|
main
|
automatic release publish to testes and google play currently we have to build the apk and release it by our self we have to create a script which shares the apk to testers with the dev flavor and to the play store with the prod flavor
| 1
|
3,106
| 11,868,468,938
|
IssuesEvent
|
2020-03-26 09:15:01
|
chocolatey-community/chocolatey-package-requests
|
https://api.github.com/repos/chocolatey-community/chocolatey-package-requests
|
closed
|
RFM - freac
|
Status: Available For Maintainer(s)
|
## Current Maintainer
- [x] I am the maintainer of the package and wish to pass it to someone else;
## I DON'T Want To Become The Maintainer
- [x] I have followed the Package Triage Process and I do NOT want to become maintainer of the package;
- [x] There is no existing open maintainer request for this package;
## Checklist
- [x] Issue title starts with 'RFM - '
## Existing Package Details
Package URL: https://chocolatey.org/packages/freac
Package source URL: https://github.com/abejenaru/chocolatey-packages/tree/master/automatic/freac
|
True
|
RFM - freac - ## Current Maintainer
- [x] I am the maintainer of the package and wish to pass it to someone else;
## I DON'T Want To Become The Maintainer
- [x] I have followed the Package Triage Process and I do NOT want to become maintainer of the package;
- [x] There is no existing open maintainer request for this package;
## Checklist
- [x] Issue title starts with 'RFM - '
## Existing Package Details
Package URL: https://chocolatey.org/packages/freac
Package source URL: https://github.com/abejenaru/chocolatey-packages/tree/master/automatic/freac
|
main
|
rfm freac current maintainer i am the maintainer of the package and wish to pass it to someone else i don t want to become the maintainer i have followed the package triage process and i do not want to become maintainer of the package there is no existing open maintainer request for this package checklist issue title starts with rfm existing package details package url package source url
| 1
|
25,273
| 4,151,867,250
|
IssuesEvent
|
2016-06-15 22:04:21
|
Wirlie/AllBanks-2
|
https://api.github.com/repos/Wirlie/AllBanks-2
|
reopened
|
Add a command to restart a plot. Like /plot clear
|
AllBanksLand bug enhancement goal high priority testing required work in progress
|
This command can help to restart an abandoned plot (also if you want to start from 0).
- [x] Add command: /plot clear
- [x] Add a confirmation before restarting a plot.
|
1.0
|
Add a command to restart a plot. Like /plot clear - This command can help to restart an abandoned plot (also if you want to start from 0).
- [x] Add command: /plot clear
- [x] Add a confirmation before restarting a plot.
|
non_main
|
add a command to restart a plot like plot clear this command can help to restart an abandoned plot also if you want to start from add command plot clear add a confirmation before restarting a plot
| 0
|
166,316
| 14,047,311,892
|
IssuesEvent
|
2020-11-02 06:53:46
|
JuanOliveros/git_web_practice
|
https://api.github.com/repos/JuanOliveros/git_web_practice
|
opened
|
Un commit que no sigue la convención de código o FIX a realizar
|
documentation
|
La convención de código a seguir:
- Para los arreglos: `<Identificador de la corrección>: <Comentario>`
- Para los arreglos con conflictos: `<Identificador de la corrección>: <Comentario por defecto del merge>`
Igualmente, solo hay 3 fixes a realizar. Al realizar uno y completarlo se creará un issue con las instrucciones a realizar para el siguiente.
Para realizar la corrección del mensaje de commit `git commit --amend` y `git commit push -f`
El último commit tiene el siguiente mensaje:
`se cambia el link de la imagen y el texto del titulo p5`
Este issue es solo un recordatorio de la convención de comentarios en los commits y puede ser cerrado.
|
1.0
|
Un commit que no sigue la convención de código o FIX a realizar - La convención de código a seguir:
- Para los arreglos: `<Identificador de la corrección>: <Comentario>`
- Para los arreglos con conflictos: `<Identificador de la corrección>: <Comentario por defecto del merge>`
Igualmente, solo hay 3 fixes a realizar. Al realizar uno y completarlo se creará un issue con las instrucciones a realizar para el siguiente.
Para realizar la corrección del mensaje de commit `git commit --amend` y `git commit push -f`
El último commit tiene el siguiente mensaje:
`se cambia el link de la imagen y el texto del titulo p5`
Este issue es solo un recordatorio de la convención de comentarios en los commits y puede ser cerrado.
|
non_main
|
un commit que no sigue la convención de código o fix a realizar la convención de código a seguir para los arreglos para los arreglos con conflictos igualmente solo hay fixes a realizar al realizar uno y completarlo se creará un issue con las instrucciones a realizar para el siguiente para realizar la corrección del mensaje de commit git commit amend y git commit push f el último commit tiene el siguiente mensaje se cambia el link de la imagen y el texto del titulo este issue es solo un recordatorio de la convención de comentarios en los commits y puede ser cerrado
| 0
|
1,836
| 6,577,368,557
|
IssuesEvent
|
2017-09-12 00:25:27
|
ansible/ansible-modules-core
|
https://api.github.com/repos/ansible/ansible-modules-core
|
closed
|
ec2_vol delete volume by name
|
affects_2.1 aws cloud feature_idea waiting_on_maintainer
|
<!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
cloud/amazon/ec2_vol
##### ANSIBLE VERSION
```
ansible 2.1.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
Nothing extra
##### OS / ENVIRONMENT
N/A
##### SUMMARY
It would be useful if the ec2_vol task was able to either delete volumes by name instead of just by id, OR if the state=list actually was able to use the name to filter the set of returned results, rather than returning all volumes, that then need to get filtered.
##### STEPS TO REPRODUCE
```
- name: Terminate EBS volumes
ec2_vol:
region: "{{ ec2_region }}"
name: "some_name"
state: absent
```
##### EXPECTED RESULTS
EBS volume can be deleted by name.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
```
<127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: vagrant
<127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1464717304.14-70488874953014 `" && echo ansible-tmp-1464717304.14-70488874953014="` echo $HOME/.ansible/tmp/ansible-tmp-1464717304.14-70488874953014
`" ) && sleep 0'
<127.0.0.1> PUT /tmp/tmpRRCEio TO /home/vagrant/.ansible/tmp/ansible-tmp-1464717304.14-70488874953014/ec2_vol
<127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/vagrant/.ansible/tmp/ansible-tmp-1464717304.14-70488874953014/ec2_vol; rm -rf "/home/vagrant/.ansible/tmp/ansible-tmp-1464717304.14-70488874953014/" > /dev/null 2>&1 && sleep 0'
failed: [localhost] (item=1) => {"failed": true, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "delete_on_termination": false,
"device_name": null, "ec2_url": null, "encrypted": false, "id": null, "instance": null, "iops": null, "name": "ml-cluster-denali-vol-1", "profile": null, "region": "us-east-1", "security_token": null, "snapshot": null, "state": "absent", "validate_certs": true, "volume_size": null, "volume_type": "standard", "zone": null}, "module_name": "ec2_vol"}, "item": "1", "msg": "Value (None) for parameter
volume is invalid. Expected: 'vol-...'."}
```
|
True
|
ec2_vol delete volume by name - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
cloud/amazon/ec2_vol
##### ANSIBLE VERSION
```
ansible 2.1.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
Nothing extra
##### OS / ENVIRONMENT
N/A
##### SUMMARY
It would be useful if the ec2_vol task was able to either delete volumes by name instead of just by id, OR if the state=list actually was able to use the name to filter the set of returned results, rather than returning all volumes, that then need to get filtered.
##### STEPS TO REPRODUCE
```
- name: Terminate EBS volumes
ec2_vol:
region: "{{ ec2_region }}"
name: "some_name"
state: absent
```
##### EXPECTED RESULTS
EBS volume can be deleted by name.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
```
<127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: vagrant
<127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1464717304.14-70488874953014 `" && echo ansible-tmp-1464717304.14-70488874953014="` echo $HOME/.ansible/tmp/ansible-tmp-1464717304.14-70488874953014
`" ) && sleep 0'
<127.0.0.1> PUT /tmp/tmpRRCEio TO /home/vagrant/.ansible/tmp/ansible-tmp-1464717304.14-70488874953014/ec2_vol
<127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/vagrant/.ansible/tmp/ansible-tmp-1464717304.14-70488874953014/ec2_vol; rm -rf "/home/vagrant/.ansible/tmp/ansible-tmp-1464717304.14-70488874953014/" > /dev/null 2>&1 && sleep 0'
failed: [localhost] (item=1) => {"failed": true, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "delete_on_termination": false,
"device_name": null, "ec2_url": null, "encrypted": false, "id": null, "instance": null, "iops": null, "name": "ml-cluster-denali-vol-1", "profile": null, "region": "us-east-1", "security_token": null, "snapshot": null, "state": "absent", "validate_certs": true, "volume_size": null, "volume_type": "standard", "zone": null}, "module_name": "ec2_vol"}, "item": "1", "msg": "Value (None) for parameter
volume is invalid. Expected: 'vol-...'."}
```
|
main
|
vol delete volume by name issue type feature idea component name cloud amazon vol ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration nothing extra os environment n a summary it would be useful if the vol task was able to either delete volumes by name instead of just by id or if the state list actually was able to use the name to filter the set of returned results rather than returning all volumes that then need to get filtered steps to reproduce name terminate ebs volumes vol region region name some name state absent expected results ebs volume can be deleted by name actual results establish local connection for user vagrant exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmprrceio to home vagrant ansible tmp ansible tmp vol exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python home vagrant ansible tmp ansible tmp vol rm rf home vagrant ansible tmp ansible tmp dev null sleep failed item failed true invocation module args aws access key null aws secret key null delete on termination false device name null url null encrypted false id null instance null iops null name ml cluster denali vol profile null region us east security token null snapshot null state absent validate certs true volume size null volume type standard zone null module name vol item msg value none for parameter volume is invalid expected vol
| 1
|
212,344
| 7,236,081,569
|
IssuesEvent
|
2018-02-13 04:43:31
|
GeoTIFF/geoblaze
|
https://api.github.com/repos/GeoTIFF/geoblaze
|
opened
|
[USER REQUEST] Describe 3rd Param for Histogram
|
top-priority
|
Currently, documentation doesn't describe that third options parameter for histogram. A user spent a long time and had to go through our source code to figure out what params to pass in with options. We need to update documentation.
|
1.0
|
[USER REQUEST] Describe 3rd Param for Histogram - Currently, documentation doesn't describe that third options parameter for histogram. A user spent a long time and had to go through our source code to figure out what params to pass in with options. We need to update documentation.
|
non_main
|
describe param for histogram currently documentation doesn t describe that third options parameter for histogram a user spent a long time and had to go through our source code to figure out what params to pass in with options we need to update documentation
| 0
|
2,431
| 8,621,114,567
|
IssuesEvent
|
2018-11-20 16:33:45
|
simplesamlphp/simplesamlphp
|
https://api.github.com/repos/simplesamlphp/simplesamlphp
|
closed
|
Double concatenation in translation string
|
maintainability
|
`lib/SimpleSAML/Error/Error.php` has the only double concatenation:
```
$moduleCode = explode(':', $this->errorCode, 2);
if (count($moduleCode) === 2) {
$this->module = $moduleCode[0];
$this->dictTitle = '{'.$this->module.':errors:title_'.$moduleCode[1].'}';
$this->dictDescr = '{'.$this->module.':errors:descr_'.$moduleCode[1].'}';
} else {
...
}
```
This is triggered by an errorcode with two colons in it, and needs there to be an `errors.definition.json` in the module's `dictionaries`-directory.
There is no case of either in the current codebase of SimpleSAMLPHP proper, it looks like dead code, so I propose to deprecate this in 1.15.
It might be necessary to support module-specific error codes, but not like this.
Part of #455
|
True
|
Double concatenation in translation string - `lib/SimpleSAML/Error/Error.php` has the only double concatenation:
```
$moduleCode = explode(':', $this->errorCode, 2);
if (count($moduleCode) === 2) {
$this->module = $moduleCode[0];
$this->dictTitle = '{'.$this->module.':errors:title_'.$moduleCode[1].'}';
$this->dictDescr = '{'.$this->module.':errors:descr_'.$moduleCode[1].'}';
} else {
...
}
```
This is triggered by an errorcode with two colons in it, and needs there to be an `errors.definition.json` in the module's `dictionaries`-directory.
There is no case of either in the current codebase of SimpleSAMLPHP proper, it looks like dead code, so I propose to deprecate this in 1.15.
It might be necessary to support module-specific error codes, but not like this.
Part of #455
|
main
|
double concatenation in translation string lib simplesaml error error php has the only double concatenation modulecode explode this errorcode if count modulecode this module modulecode this dicttitle this module errors title modulecode this dictdescr this module errors descr modulecode else this is triggered by an errorcode with two colons in it and needs there to be an errors definition json in the module s dictionaries directory there is no case of either in the current codebase of simplesamlphp proper it looks like dead code so i propose to deprecate this in it might be necessary to support module specific error codes but not like this part of
| 1
|
2,845
| 3,211,738,577
|
IssuesEvent
|
2015-10-06 12:33:38
|
coreos/rkt
|
https://api.github.com/repos/coreos/rkt
|
closed
|
Rkt should support setting supplemental groups on containers
|
area/usability kind/enhancement
|
Currently it is possible to set the UID/GID on a container process, but not possible to control the supplemental groups. It would be very convenient to have supplemental group control to enable sharing volumes across containers running as different UID/GIDs.
|
True
|
Rkt should support setting supplemental groups on containers - Currently it is possible to set the UID/GID on a container process, but not possible to control the supplemental groups. It would be very convenient to have supplemental group control to enable sharing volumes across containers running as different UID/GIDs.
|
non_main
|
rkt should support setting supplemental groups on containers currently it is possible to set the uid gid on a container process but not possible to control the supplemental groups it would be very convenient to have supplemental group control to enable sharing volumes across containers running as different uid gids
| 0
|
137,210
| 5,299,907,306
|
IssuesEvent
|
2017-02-10 02:03:38
|
copperhead/bugtracker
|
https://api.github.com/repos/copperhead/bugtracker
|
closed
|
IDS
|
enhancement priority-low
|
It would be cool to have optional built-in IDS support. It's not possible to do this well without it being built into the OS due to lack of privileges, especially as the app sandbox is hardened. It's an area where CopperheadOS could provide a real edge. Android has SafetyNet, but that's meant to protect the ecosystem as a whole, not individuals.
|
1.0
|
IDS - It would be cool to have optional built-in IDS support. It's not possible to do this well without it being built into the OS due to lack of privileges, especially as the app sandbox is hardened. It's an area where CopperheadOS could provide a real edge. Android has SafetyNet, but that's meant to protect the ecosystem as a whole, not individuals.
|
non_main
|
ids it would be cool to have optional built in ids support it s not possible to do this well without it being built into the os due to lack of privileges especially as the app sandbox is hardened it s an area where copperheados could provide a real edge android has safetynet but that s meant to protect the ecosystem as a whole not individuals
| 0
|
821,973
| 30,845,819,215
|
IssuesEvent
|
2023-08-02 13:40:16
|
plan-be/iode
|
https://api.github.com/repos/plan-be/iode
|
closed
|
PYTHON: Change compile options according to the compiler and buid config in cythonize_iode.py
|
enhancement priority: high python difficulty: low
|
Currently, the compile options are fixed for the MSVC compiler and the Debug build config:
```
extra_compile_args=["-Zi", "/Od", "/DVC", "/DSCRPROTO", "/DREALD"],
```
|
1.0
|
PYTHON: Change compile options according to the compiler and buid config in cythonize_iode.py - Currently, the compile options are fixed for the MSVC compiler and the Debug build config:
```
extra_compile_args=["-Zi", "/Od", "/DVC", "/DSCRPROTO", "/DREALD"],
```
|
non_main
|
python change compile options according to the compiler and buid config in cythonize iode py currently the compile options are fixed for the msvc compiler and the debug build config extra compile args
| 0
|
110,807
| 9,477,936,354
|
IssuesEvent
|
2019-04-19 20:32:26
|
cerner/terra-core
|
https://api.github.com/repos/cerner/terra-core
|
closed
|
Improve icon visual regression test coverage
|
Orion Reviewed icon intermediate issue testing
|
# Feature Request
## Description
Currently, our icon visual regression coverage is very minimal. We should expand it to better capture the full icon set. We've had a couple bugs slip through related to how the SVGs have been formatted that we can catch if we set up visual regression tests for the entire icon set.
|
1.0
|
Improve icon visual regression test coverage - # Feature Request
## Description
Currently, our icon visual regression coverage is very minimal. We should expand it to better capture the full icon set. We've had a couple bugs slip through related to how the SVGs have been formatted that we can catch if we set up visual regression tests for the entire icon set.
|
non_main
|
improve icon visual regression test coverage feature request description currently our icon visual regression coverage is very minimal we should expand it to better capture the full icon set we ve had a couple bugs slip through related to how the svgs have been formatted that we can catch if we set up visual regression tests for the entire icon set
| 0
|
990
| 4,756,646,363
|
IssuesEvent
|
2016-10-24 14:33:53
|
ansible/ansible-modules-extras
|
https://api.github.com/repos/ansible/ansible-modules-extras
|
closed
|
[Regression] asa_command: silently allows invalid command
|
affects_2.2 bug_report networking waiting_on_maintainer
|
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
asa_command
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel eb33ed4219) last updated 2016/09/27 09:18:44 (GMT +100)
lib/ansible/modules/core: (devel c03697c81e) last updated 2016/09/27 09:18:49 (GMT +100)
lib/ansible/modules/extras: (devel 119bc466be) last updated 2016/09/27 09:18:51 (GMT +100)
```
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
Running `asa_command` with an invalid `commands:` no longer bombs out
##### STEPS TO REPRODUCE
```
- name: run commands that include invalid command
asa_command:
commands:
- show version
- show foo
provider: "{{ cli }}"
register: result
ignore_errors: yes
```
##### EXPECTED RESULTS
##### ACTUAL RESULTS
```
Task to report error (gracefully)
```
|
True
|
[Regression] asa_command: silently allows invalid command - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
asa_command
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel eb33ed4219) last updated 2016/09/27 09:18:44 (GMT +100)
lib/ansible/modules/core: (devel c03697c81e) last updated 2016/09/27 09:18:49 (GMT +100)
lib/ansible/modules/extras: (devel 119bc466be) last updated 2016/09/27 09:18:51 (GMT +100)
```
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
Running `asa_command` with an invalid `commands:` no longer bombs out
##### STEPS TO REPRODUCE
```
- name: run commands that include invalid command
asa_command:
commands:
- show version
- show foo
provider: "{{ cli }}"
register: result
ignore_errors: yes
```
##### EXPECTED RESULTS
##### ACTUAL RESULTS
```
Task to report error (gracefully)
```
|
main
|
asa command silently allows invalid command issue type bug report component name asa command ansible version ansible devel last updated gmt lib ansible modules core devel last updated gmt lib ansible modules extras devel last updated gmt configuration os environment summary running asa command with an invalid commands no longer bombs out steps to reproduce name run commands that include invalid command asa command commands show version show foo provider cli register result ignore errors yes expected results actual results task to report error gracefully
| 1
|
4,834
| 24,912,106,884
|
IssuesEvent
|
2022-10-30 00:51:37
|
chocolatey-community/chocolatey-package-requests
|
https://api.github.com/repos/chocolatey-community/chocolatey-package-requests
|
closed
|
RFM - ds4windows
|
Status: Available For Maintainer(s)
|
## Current Maintainer
<!-- If you are not confirmed as a known maintainer, you may be asked to take additional steps to confirm your user account -->
- [x] I am the maintainer of the package and wish to pass it to someone else;
## Checklist
- [x] Issue title starts with 'RFM - '
## Existing Package Details
Package URL: https://chocolatey.org/packages/ds4windows
Package source URL: https://github.com/jtcmedia/chocolatey-packages/tree/master/ds4windows
|
True
|
RFM - ds4windows - ## Current Maintainer
<!-- If you are not confirmed as a known maintainer, you may be asked to take additional steps to confirm your user account -->
- [x] I am the maintainer of the package and wish to pass it to someone else;
## Checklist
- [x] Issue title starts with 'RFM - '
## Existing Package Details
Package URL: https://chocolatey.org/packages/ds4windows
Package source URL: https://github.com/jtcmedia/chocolatey-packages/tree/master/ds4windows
|
main
|
rfm current maintainer i am the maintainer of the package and wish to pass it to someone else checklist issue title starts with rfm existing package details package url package source url
| 1
|
3,122
| 11,956,415,536
|
IssuesEvent
|
2020-04-04 10:18:11
|
custom-cards/flex-table-card
|
https://api.github.com/repos/custom-cards/flex-table-card
|
closed
|
here somehow avoid that a 'null' is string-converted
|
maintaining todo :spiral_notepad:
|
https://github.com/custom-cards/flex-table-card/blob/3ec416da586266272e06f958622158e3b5ddf9b6/flex-table-card.js#L164-L168
---
###### This issue was generated by [todo](https://todo.jasonet.co) based on a `todo` comment in 3ec416da586266272e06f958622158e3b5ddf9b6. It's been assigned to @daringer because they committed the code.
|
True
|
here somehow avoid that a 'null' is string-converted - https://github.com/custom-cards/flex-table-card/blob/3ec416da586266272e06f958622158e3b5ddf9b6/flex-table-card.js#L164-L168
---
###### This issue was generated by [todo](https://todo.jasonet.co) based on a `todo` comment in 3ec416da586266272e06f958622158e3b5ddf9b6. It's been assigned to @daringer because they committed the code.
|
main
|
here somehow avoid that a null is string converted this issue was generated by based on a todo comment in it s been assigned to daringer because they committed the code
| 1
|
244,631
| 18,764,165,284
|
IssuesEvent
|
2021-11-05 20:34:00
|
satijalab/seurat
|
https://api.github.com/repos/satijalab/seurat
|
closed
|
Dockerfile
|
documentation
|
<!-- A clear description of what content at https://satijalab.org/seurat or in the Seurat function man pages is an issue. -->
Good morning, i was trying to reach the dockerfile, in order to build the docker with seurat, but the only thing i can find is https://hub.docker.com/r/satijalab/seurat the link to pull the docker. Can you please provide the dockerfile?
Thank you
Best
Luca
|
1.0
|
Dockerfile - <!-- A clear description of what content at https://satijalab.org/seurat or in the Seurat function man pages is an issue. -->
Good morning, i was trying to reach the dockerfile, in order to build the docker with seurat, but the only thing i can find is https://hub.docker.com/r/satijalab/seurat the link to pull the docker. Can you please provide the dockerfile?
Thank you
Best
Luca
|
non_main
|
dockerfile good morning i was trying to reach the dockerfile in order to build the docker with seurat but the only thing i can find is the link to pull the docker can you please provide the dockerfile thank you best luca
| 0
|
165,517
| 26,183,988,518
|
IssuesEvent
|
2023-01-02 20:09:28
|
flutter/website
|
https://api.github.com/repos/flutter/website
|
opened
|
Migrate to Bootstrap 5
|
infrastructure design p3-low blocked e2-days e3-weeks
|
### Describe the problem
Bootstrap 5 is the current release Bootstrap, replacing Bootstrap 4. We use it heavily across the site and we want to make sure we stay up to date. This will also allow us to eventually drop Jquery since Bootstrap 5 no longer uses it. Beyond that, this will also make a dark mode slightly easier with its (not yet released) [color mode functionality](https://getbootstrap.com/docs/5.3/customize/color-modes/).
### Expected fix
We should migrate away from Bootstrap 4 and to 5.
### Additional context
There may be some incompatibilities due to Bootstrap now using the Dart sass compiler which supports some newer features. We may have to configure Jekyll to use [sass-embedded](https://rubygems.org/gems/sass-embedded) instead somehow, which is implemented with Dart sass :)
|
1.0
|
Migrate to Bootstrap 5 - ### Describe the problem
Bootstrap 5 is the current release Bootstrap, replacing Bootstrap 4. We use it heavily across the site and we want to make sure we stay up to date. This will also allow us to eventually drop Jquery since Bootstrap 5 no longer uses it. Beyond that, this will also make a dark mode slightly easier with its (not yet released) [color mode functionality](https://getbootstrap.com/docs/5.3/customize/color-modes/).
### Expected fix
We should migrate away from Bootstrap 4 and to 5.
### Additional context
There may be some incompatibilities due to Bootstrap now using the Dart sass compiler which supports some newer features. We may have to configure Jekyll to use [sass-embedded](https://rubygems.org/gems/sass-embedded) instead somehow, which is implemented with Dart sass :)
|
non_main
|
migrate to bootstrap describe the problem bootstrap is the current release bootstrap replacing bootstrap we use it heavily across the site and we want to make sure we stay up to date this will also allow us to eventually drop jquery since bootstrap no longer uses it beyond that this will also make a dark mode slightly easier with its not yet released expected fix we should migrate away from bootstrap and to additional context there may be some incompatibilities due to bootstrap now using the dart sass compiler which supports some newer features we may have to configure jekyll to use instead somehow which is implemented with dart sass
| 0
|
151,701
| 13,429,844,132
|
IssuesEvent
|
2020-09-07 03:03:41
|
rorepoid/twgroup
|
https://api.github.com/repos/rorepoid/twgroup
|
closed
|
Desafío 1
|
documentation
|
Al momento de iniciar un nuevo proyecto en Laravel debemos realizar una serie de pasos para configurar el proyecto dependiendo de sus requerimientos. Imagina que necesitamos una plataforma sobre Laravel que utilizará un motor de base de datos MySQL/MariaDB, un servidor de correos SMTP y un servidor Redis.
¿Cuáles son los pasos que consideras necesarios para dejar la aplicación funcionando en modo de desarrollo? (Describe los comandos necesarios que ejecutarías y que archivos modificarías en base a los requerimientos mencionados).
|
1.0
|
Desafío 1 - Al momento de iniciar un nuevo proyecto en Laravel debemos realizar una serie de pasos para configurar el proyecto dependiendo de sus requerimientos. Imagina que necesitamos una plataforma sobre Laravel que utilizará un motor de base de datos MySQL/MariaDB, un servidor de correos SMTP y un servidor Redis.
¿Cuáles son los pasos que consideras necesarios para dejar la aplicación funcionando en modo de desarrollo? (Describe los comandos necesarios que ejecutarías y que archivos modificarías en base a los requerimientos mencionados).
|
non_main
|
desafío al momento de iniciar un nuevo proyecto en laravel debemos realizar una serie de pasos para configurar el proyecto dependiendo de sus requerimientos imagina que necesitamos una plataforma sobre laravel que utilizará un motor de base de datos mysql mariadb un servidor de correos smtp y un servidor redis ¿cuáles son los pasos que consideras necesarios para dejar la aplicación funcionando en modo de desarrollo describe los comandos necesarios que ejecutarías y que archivos modificarías en base a los requerimientos mencionados
| 0
|
1,482
| 6,416,004,678
|
IssuesEvent
|
2017-08-08 14:00:09
|
ansible/ansible-modules-extras
|
https://api.github.com/repos/ansible/ansible-modules-extras
|
closed
|
vca_vapp missing network_ip option
|
affects_2.3 cloud feature_idea vmware waiting_on_maintainer
|
Unable to set IP address when creating vCloud instances due to missing 'network_ip' option.
See:
https://github.com/vmware/vca-codesamples/blob/master/ansibleworkshopVMworld2015/lessons/lesson2/library/vca_vapp.py#L100
[module: cloud/vmware/vca_vapp.py]
|
True
|
vca_vapp missing network_ip option - Unable to set IP address when creating vCloud instances due to missing 'network_ip' option.
See:
https://github.com/vmware/vca-codesamples/blob/master/ansibleworkshopVMworld2015/lessons/lesson2/library/vca_vapp.py#L100
[module: cloud/vmware/vca_vapp.py]
|
main
|
vca vapp missing network ip option unable to set ip address when creating vcloud instances due to missing network ip option see
| 1
|
2,504
| 8,655,459,620
|
IssuesEvent
|
2018-11-27 16:00:29
|
codestation/qcma
|
https://api.github.com/repos/codestation/qcma
|
closed
|
Music damaged file
|
unmaintained
|
Hi, I recently Updated from 3.52 to the last firmware (3.61). But when I try to send music from the PC to PSVita, the music app of the psvita says me is a damaged file. I tried with converting the music (mp3 to mp3) with ffmpeg, vlc but nothing.
I use the qcma client of Linux installed from aur.
With the official cma I can send the music without any problem.
|
True
|
Music damaged file - Hi, I recently Updated from 3.52 to the last firmware (3.61). But when I try to send music from the PC to PSVita, the music app of the psvita says me is a damaged file. I tried with converting the music (mp3 to mp3) with ffmpeg, vlc but nothing.
I use the qcma client of Linux installed from aur.
With the official cma I can send the music without any problem.
|
main
|
music damaged file hi i recently updated from to the last firmware but when i try to send music from the pc to psvita the music app of the psvita says me is a damaged file i tried with converting the music to with ffmpeg vlc but nothing i use the qcma client of linux installed from aur with the official cma i can send the music without any problem
| 1
|
313,875
| 26,959,577,223
|
IssuesEvent
|
2023-02-08 17:09:11
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed
|
C-test-failure O-robot O-roachtest branch-master release-blocker T-disaster-recovery
|
roachtest.restore/tpce/8TB/aws/nodes=10/cpus=8 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/8311515?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/8311515?buildTab=artifacts#/restore/tpce/8TB/aws/nodes=10/cpus=8) on master @ [8866ecadd99e5f6dc8db9fed5d09bc57cd90b03c](https://github.com/cockroachdb/cockroach/commits/8866ecadd99e5f6dc8db9fed5d09bc57cd90b03c):
```
test artifacts and logs in: /artifacts/restore/tpce/8TB/aws/nodes=10/cpus=8/run_1
(monitor.go:127).Wait: monitor failure: monitor command failure: unexpected node event: 3: dead (exit status 7)
```
<p>Parameters: <code>ROACHTEST_cloud=aws</code>
, <code>ROACHTEST_cpu=8</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/disaster-recovery
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*restore/tpce/8TB/aws/nodes=10/cpus=8.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-23426
|
2.0
|
roachtest: restore/tpce/8TB/aws/nodes=10/cpus=8 failed - roachtest.restore/tpce/8TB/aws/nodes=10/cpus=8 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/8311515?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/8311515?buildTab=artifacts#/restore/tpce/8TB/aws/nodes=10/cpus=8) on master @ [8866ecadd99e5f6dc8db9fed5d09bc57cd90b03c](https://github.com/cockroachdb/cockroach/commits/8866ecadd99e5f6dc8db9fed5d09bc57cd90b03c):
```
test artifacts and logs in: /artifacts/restore/tpce/8TB/aws/nodes=10/cpus=8/run_1
(monitor.go:127).Wait: monitor failure: monitor command failure: unexpected node event: 3: dead (exit status 7)
```
<p>Parameters: <code>ROACHTEST_cloud=aws</code>
, <code>ROACHTEST_cpu=8</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/disaster-recovery
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*restore/tpce/8TB/aws/nodes=10/cpus=8.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-23426
|
non_main
|
roachtest restore tpce aws nodes cpus failed roachtest restore tpce aws nodes cpus with on master test artifacts and logs in artifacts restore tpce aws nodes cpus run monitor go wait monitor failure monitor command failure unexpected node event dead exit status parameters roachtest cloud aws roachtest cpu roachtest encrypted false roachtest fs roachtest localssd false roachtest ssd help see see cc cockroachdb disaster recovery jira issue crdb
| 0
|
1,545
| 6,572,237,119
|
IssuesEvent
|
2017-09-11 00:26:30
|
ansible/ansible-modules-extras
|
https://api.github.com/repos/ansible/ansible-modules-extras
|
closed
|
New ecs_service_facts module has a return behavior that is inconsistent when compared to existing *_facts modules.
|
affects_2.1 aws bug_report cloud feature_idea waiting_on_maintainer
|
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
`/ansible/ansible-modules-extras/cloud/amazon/ecs_service_facts.py`
##### ANSIBLE VERSION
```
2.1.0
```
##### SUMMARY
New ecs_service_facts module has a return behavior that is inconsistent when compared to existing *_facts modules.
The new `ecs_service_facts` module works without issues, but it automatically creates a `services` variable (i.e. by returning with `ecs_facts_result = dict(changed=False, ansible_facts=ecs_facts)`). Since the task/module does support querying specify services, it seemed that it would be better if the module behaved similar to other /cloud/amazon/*_fact modules where they use `register` to capture the output.
While the module allows to be queried with a single task, this could cause some confusion if a user executes the task multiple times (i.e. `services` will only contain the result from the last task).
I am willing to create a pull request but there may be reasons for this difference so I want to check prior to spending time on it / creating a breaking change.
##### STEPS TO REPRODUCE
Execute:
```
- name: Get the ECS service 1's info
ecs_service_facts:
region: us-east-1
cluster: ecs-cluster-name
service: service-name-1
details: "true"
- debug: var=services
- name: Get the ECS service 2's info
ecs_service_facts:
region: us-east-1
cluster: ecs-cluster-name
service: service-name-2
details: "true"
- debug: var=services
```
As expected, `services` will only contain the info for service-name-2.
Notes:
- no `register` used.
- can also use `register`, but to access the information, the accessor would be `variable_name.ansible_facts.services`.
##### EXPECTED RESULTS
Execute:
```
- name: Get the ECS service 1's info
ecs_service_facts:
region: us-east-1
cluster: ecs-cluster-name
service: service-name-1
details: "true"
register: service_1
- debug: var=service_1
- name: Get the ECS service 2's info
ecs_service_facts:
region: us-east-1
cluster: ecs-cluster-name
service: service-name-2
details: "true"
register: var=service_2
- debug: var=service_1
- debug: var=service_2
```
`service_1` and `service_2` would contain the information for each service.
|
True
|
New ecs_service_facts module has a return behavior that is inconsistent when compared to existing *_facts modules. - ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
`/ansible/ansible-modules-extras/cloud/amazon/ecs_service_facts.py`
##### ANSIBLE VERSION
```
2.1.0
```
##### SUMMARY
New ecs_service_facts module has a return behavior that is inconsistent when compared to existing *_facts modules.
The new `ecs_service_facts` module works without issues, but it automatically creates a `services` variable (i.e. by returning with `ecs_facts_result = dict(changed=False, ansible_facts=ecs_facts)`). Since the task/module does support querying specify services, it seemed that it would be better if the module behaved similar to other /cloud/amazon/*_fact modules where they use `register` to capture the output.
While the module allows to be queried with a single task, this could cause some confusion if a user executes the task multiple times (i.e. `services` will only contain the result from the last task).
I am willing to create a pull request but there may be reasons for this difference so I want to check prior to spending time on it / creating a breaking change.
##### STEPS TO REPRODUCE
Execute:
```
- name: Get the ECS service 1's info
ecs_service_facts:
region: us-east-1
cluster: ecs-cluster-name
service: service-name-1
details: "true"
- debug: var=services
- name: Get the ECS service 2's info
ecs_service_facts:
region: us-east-1
cluster: ecs-cluster-name
service: service-name-2
details: "true"
- debug: var=services
```
As expected, `services` will only contain the info for service-name-2.
Notes:
- no `register` used.
- can also use `register`, but to access the information, the accessor would be `variable_name.ansible_facts.services`.
##### EXPECTED RESULTS
Execute:
```
- name: Get the ECS service 1's info
ecs_service_facts:
region: us-east-1
cluster: ecs-cluster-name
service: service-name-1
details: "true"
register: service_1
- debug: var=service_1
- name: Get the ECS service 2's info
ecs_service_facts:
region: us-east-1
cluster: ecs-cluster-name
service: service-name-2
details: "true"
register: var=service_2
- debug: var=service_1
- debug: var=service_2
```
`service_1` and `service_2` would contain the information for each service.
|
main
|
new ecs service facts module has a return behavior that is inconsistent when compared to existing facts modules issue type feature idea component name ansible ansible modules extras cloud amazon ecs service facts py ansible version summary new ecs service facts module has a return behavior that is inconsistent when compared to existing facts modules the new ecs service facts module works without issues but it automatically creates a services variable i e by returning with ecs facts result dict changed false ansible facts ecs facts since the task module does support querying specify services it seemed that it would be better if the module behaved similar to other cloud amazon fact modules where they use register to capture the output while the module allows to be queried with a single task this could cause some confusion if a user executes the task multiple times i e services will only contain the result from the last task i am willing to create a pull request but there may be reasons for this difference so i want to check prior to spending time on it creating a breaking change steps to reproduce execute name get the ecs service s info ecs service facts region us east cluster ecs cluster name service service name details true debug var services name get the ecs service s info ecs service facts region us east cluster ecs cluster name service service name details true debug var services as expected services will only contain the info for service name notes no register used can also use register but to access the information the accessor would be variable name ansible facts services expected results execute name get the ecs service s info ecs service facts region us east cluster ecs cluster name service service name details true register service debug var service name get the ecs service s info ecs service facts region us east cluster ecs cluster name service service name details true register var service debug var service debug var service service and service would contain the information for each service
| 1
|
152,890
| 13,487,072,920
|
IssuesEvent
|
2020-09-11 10:23:12
|
cksystemsgroup/monster
|
https://api.github.com/repos/cksystemsgroup/monster
|
closed
|
Write down a Concept
|
documentation
|
Create a "big-picture" description, how the symbolic execution engine should work. This should ideally be written in markdown an should be convertable to browsable HTML.
We want to write that in the simplest way. We should describe the engine based on a running example of C* code. This piece of code should have at most one if statement.
|
1.0
|
Write down a Concept - Create a "big-picture" description, how the symbolic execution engine should work. This should ideally be written in markdown an should be convertable to browsable HTML.
We want to write that in the simplest way. We should describe the engine based on a running example of C* code. This piece of code should have at most one if statement.
|
non_main
|
write down a concept create a big picture description how the symbolic execution engine should work this should ideally be written in markdown an should be convertable to browsable html we want to write that in the simplest way we should describe the engine based on a running example of c code this piece of code should have at most one if statement
| 0
|
3,627
| 14,672,547,338
|
IssuesEvent
|
2020-12-30 10:53:23
|
Homebrew/homebrew-core
|
https://api.github.com/repos/Homebrew/homebrew-core
|
opened
|
luajit probably needs to be deprecated
|
help wanted maintainer feedback
|
- The latest release (stable OR beta) is from 2017
- It's heavily patched
- Every new macOS version requires an additional patch
- Upstream's recommendation is to “build from git HEAD”, and they won't apparently ship new releases: https://github.com/LuaJIT/LuaJIT/issues/648#issuecomment-752404043
The reason I'm not doing a pull request directly is that a lot of things depend on luajit, so I want to open a discussion and figure out the best way to handle this. Can some of these be migrated to one of the lua formulas?
|
True
|
luajit probably needs to be deprecated - - The latest release (stable OR beta) is from 2017
- It's heavily patched
- Every new macOS version requires an additional patch
- Upstream's recommendation is to “build from git HEAD”, and they won't apparently ship new releases: https://github.com/LuaJIT/LuaJIT/issues/648#issuecomment-752404043
The reason I'm not doing a pull request directly is that a lot of things depend on luajit, so I want to open a discussion and figure out the best way to handle this. Can some of these be migrated to one of the lua formulas?
|
main
|
luajit probably needs to be deprecated the latest release stable or beta is from it s heavily patched every new macos version requires an additional patch upstream s recommendation is to “build from git head” and they won t apparently ship new releases the reason i m not doing a pull request directly is that a lot of things depend on luajit so i want to open a discussion and figure out the best way to handle this can some of these be migrated to one of the lua formulas
| 1
|
83,641
| 3,638,064,985
|
IssuesEvent
|
2016-02-12 14:07:21
|
molgenis/molgenis
|
https://api.github.com/repos/molgenis/molgenis
|
closed
|
Charts won't plot TypeTest ID column versus TypeTest ID column
|
bug molgenis-dataexplorer priority-later
|
## Reproduce
Select Dataexplorer, Select TypeTest, Select charts
Create scatter plot for ID versus ID column.
## Expected
I can see the chart
## Actual
I get a somewhat obscure error:
```
18:32:36.221 [ajp-bio-8009-exec-161] ERROR org.molgenis.charts.ChartController - null
org.elasticsearch.action.search.SearchPhaseExecutionException: Failed to execute phase [query_fetch], all shards failed; shardFailures {[3rAobd3GTcOEsqtR2YsSGQ][molgenis][0]: QueryPhaseExecutionException[[molgenis][0]: query[ConstantScore(cache(+_type:org_molgenis_test_TypeTest +org.elasticsearch.index.search.nested.NonNestedDocsFilter@7b0ccb35))],from[0],size[1000],sort[<custom:"id": org.elasticsearch.index.fielddata.fieldcomparator.LongValuesComparatorSource@64be0481>,<custom:"id": org.elasticsearch.index.fielddata.fieldcomparator.LongValuesComparatorSource@73957d5e>]: Query Failed [Failed to execute main query]]; nested: ElasticsearchException[java.lang.NumberFormatException: Invalid shift value in prefixCoded bytes (is encoded value really an INT?)]; nested: UncheckedExecutionException[java.lang.NumberFormatException: Invalid shift value in prefixCoded bytes (is encoded value really an INT?)]; nested: NumberFormatException[Invalid shift value in prefixCoded bytes (is encoded value really an INT?)]; }
at org.elasticsearch.action.search.type.TransportSearchTypeAction$BaseAsyncAction.onFirstPhaseResult(TransportSearchTypeAction.java:233) ~[elasticsearch-1.4.4.jar:na]
at org.elasticsearch.action.search.type.TransportSearchTypeAction$BaseAsyncAction$1.onFailure(TransportSearchTypeAction.java:179) ~[elasticsearch-1.4.4.jar:na]
at org.elasticsearch.search.action.SearchServiceTransportAction$23.run(SearchServiceTransportAction.java:565) ~[elasticsearch-1.4.4.jar:na]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) ~[na:1.8.0_31]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) ~[na:1.8.0_31]
at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_31]
```
|
1.0
|
Charts won't plot TypeTest ID column versus TypeTest ID column - ## Reproduce
Select Dataexplorer, Select TypeTest, Select charts
Create scatter plot for ID versus ID column.
## Expected
I can see the chart
## Actual
I get a somewhat obscure error:
```
18:32:36.221 [ajp-bio-8009-exec-161] ERROR org.molgenis.charts.ChartController - null
org.elasticsearch.action.search.SearchPhaseExecutionException: Failed to execute phase [query_fetch], all shards failed; shardFailures {[3rAobd3GTcOEsqtR2YsSGQ][molgenis][0]: QueryPhaseExecutionException[[molgenis][0]: query[ConstantScore(cache(+_type:org_molgenis_test_TypeTest +org.elasticsearch.index.search.nested.NonNestedDocsFilter@7b0ccb35))],from[0],size[1000],sort[<custom:"id": org.elasticsearch.index.fielddata.fieldcomparator.LongValuesComparatorSource@64be0481>,<custom:"id": org.elasticsearch.index.fielddata.fieldcomparator.LongValuesComparatorSource@73957d5e>]: Query Failed [Failed to execute main query]]; nested: ElasticsearchException[java.lang.NumberFormatException: Invalid shift value in prefixCoded bytes (is encoded value really an INT?)]; nested: UncheckedExecutionException[java.lang.NumberFormatException: Invalid shift value in prefixCoded bytes (is encoded value really an INT?)]; nested: NumberFormatException[Invalid shift value in prefixCoded bytes (is encoded value really an INT?)]; }
at org.elasticsearch.action.search.type.TransportSearchTypeAction$BaseAsyncAction.onFirstPhaseResult(TransportSearchTypeAction.java:233) ~[elasticsearch-1.4.4.jar:na]
at org.elasticsearch.action.search.type.TransportSearchTypeAction$BaseAsyncAction$1.onFailure(TransportSearchTypeAction.java:179) ~[elasticsearch-1.4.4.jar:na]
at org.elasticsearch.search.action.SearchServiceTransportAction$23.run(SearchServiceTransportAction.java:565) ~[elasticsearch-1.4.4.jar:na]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) ~[na:1.8.0_31]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) ~[na:1.8.0_31]
at java.lang.Thread.run(Thread.java:745) ~[na:1.8.0_31]
```
|
non_main
|
charts won t plot typetest id column versus typetest id column reproduce select dataexplorer select typetest select charts create scatter plot for id versus id column expected i can see the chart actual i get a somewhat obscure error error org molgenis charts chartcontroller null org elasticsearch action search searchphaseexecutionexception failed to execute phase all shards failed shardfailures queryphaseexecutionexception query from size sort query failed nested elasticsearchexception nested uncheckedexecutionexception nested numberformatexception at org elasticsearch action search type transportsearchtypeaction baseasyncaction onfirstphaseresult transportsearchtypeaction java at org elasticsearch action search type transportsearchtypeaction baseasyncaction onfailure transportsearchtypeaction java at org elasticsearch search action searchservicetransportaction run searchservicetransportaction java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java
| 0
|
1,150
| 5,008,198,798
|
IssuesEvent
|
2016-12-12 18:50:41
|
ansible/ansible-modules-core
|
https://api.github.com/repos/ansible/ansible-modules-core
|
closed
|
iam_policy not using role structure for policy_document path
|
affects_2.0 aws bug_report cloud waiting_on_maintainer
|
##### Issue Type:
- Bug Report
##### Component Name:
iam_policy module
##### Ansible Version:
ansible 2.0.1.0
##### Ansible Configuration:
no changes to ansible.cfg
##### Environment:
control server Redhat 6.7
target server Redhat 6.7
##### Summary:
iam_policy policy_document parameter does not use role file structure
##### Steps To Reproduce:
aws.yml
```
---
- hosts: localhost
connection: local
gather_facts: true
roles:
- aws
```
roles/aws/task/main.yml
```
---
- name: configure iam policy
iam_policy:
iam_name: test
iam_type: group
policy_name: test
policy_document: test.json
# policy_document: roles/aws/files/test.json
region: us-east-1
state: present
```
roles/aws/files/test.json
```
---
<any valid policy>
```
run play
ansible-playbook aws.yml, task fails, no such file or directory test.json
however, if you change the policy_document argument to: roles/aws/files/test.json and it works
##### Expected Results:
Policy is created
##### Actual Results:
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: IOError: [Errno 2] No such file or directory: 'test.json'
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "parsed": false}
|
True
|
iam_policy not using role structure for policy_document path - ##### Issue Type:
- Bug Report
##### Component Name:
iam_policy module
##### Ansible Version:
ansible 2.0.1.0
##### Ansible Configuration:
no changes to ansible.cfg
##### Environment:
control server Redhat 6.7
target server Redhat 6.7
##### Summary:
iam_policy policy_document parameter does not use role file structure
##### Steps To Reproduce:
aws.yml
```
---
- hosts: localhost
connection: local
gather_facts: true
roles:
- aws
```
roles/aws/task/main.yml
```
---
- name: configure iam policy
iam_policy:
iam_name: test
iam_type: group
policy_name: test
policy_document: test.json
# policy_document: roles/aws/files/test.json
region: us-east-1
state: present
```
roles/aws/files/test.json
```
---
<any valid policy>
```
run play
ansible-playbook aws.yml, task fails, no such file or directory test.json
however, if you change the policy_document argument to: roles/aws/files/test.json and it works
##### Expected Results:
Policy is created
##### Actual Results:
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: IOError: [Errno 2] No such file or directory: 'test.json'
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "parsed": false}
|
main
|
iam policy not using role structure for policy document path issue type bug report component name iam policy module ansible version ansible ansible configuration no changes to ansible cfg environment control server redhat target server redhat summary iam policy policy document parameter does not use role file structure steps to reproduce aws yml hosts localhost connection local gather facts true roles aws roles aws task main yml name configure iam policy iam policy iam name test iam type group policy name test policy document test json policy document roles aws files test json region us east state present roles aws files test json run play ansible playbook aws yml task fails no such file or directory test json however if you change the policy document argument to roles aws files test json and it works expected results policy is created actual results an exception occurred during task execution to see the full traceback use vvv the error was ioerror no such file or directory test json fatal failed changed false failed true parsed false
| 1
|
4,167
| 19,982,012,114
|
IssuesEvent
|
2022-01-30 02:58:53
|
thumbor/thumbor-bootcamp
|
https://api.github.com/repos/thumbor/thumbor-bootcamp
|
opened
|
[Bootcamp Task] Remove thumbor upload into its own project
|
task L3 python maintainability
|
## Areas of Expertise
thumbor, open-source, testing, python, maintainability
Creating your own open source project will teach a lot of the concepts required for participating in another's open source project. Great experience!
## Summary
Create a new project `thumbor-upload` in thumbor's org and move the upload functionality into it. Once it has been moved, create a L2 task here to remove upload from thumbor's codebase (or do it in two PRs yourself).
## Involved Modules
* [thumbor](https://github.com/thumbor/thumbor/) - [upload](https://github.com/thumbor/thumbor/blob/master/thumbor/handlers/upload.py)
Don't forget to also check the tests and docs on uploading.
## Task Relevance
Removing code from thumbor's codebase is as important as adding. The leaner the project is, the faster we can move and onboard new people.
## Further Details
When you pick up this task, tag @heynemann to create the project for you and add you to the project team of contributors.
Then:
* Create the structure to run tests
* Add tests
* Add code
* Add github action to run the code on commit
* Add linting
* Add linting to the github action
* Add a nice README explaining how to use this new extensions
* Tag a version 0.1.0 (`git tag 0.1.0 && git push --tags`)
* Ask @heynemann to submit to pypi and he'll add you to the list of maintainers.
## Task Workflow
To complete this task, follow this workflow:
1. [Fork the involved repositories](http://help.github.com/fork-a-repo/)
2. In each repository there's a documentation on how to install dependencies and initialize your environment
3. Hack, in no particular order:
- Write code & tests
- Write new tests
- Write docs
- Improve design
- Check that all tests pass
- Repeat until you're satisfied
4. [Submit a pull request](https://docs.github.com/en/github/collaborating-with-pull-requests/proposing-changes-to-your-work-with-pull-requests/creating-a-pull-request).
|
True
|
[Bootcamp Task] Remove thumbor upload into its own project - ## Areas of Expertise
thumbor, open-source, testing, python, maintainability
Creating your own open source project will teach a lot of the concepts required for participating in another's open source project. Great experience!
## Summary
Create a new project `thumbor-upload` in thumbor's org and move the upload functionality into it. Once it has been moved, create a L2 task here to remove upload from thumbor's codebase (or do it in two PRs yourself).
## Involved Modules
* [thumbor](https://github.com/thumbor/thumbor/) - [upload](https://github.com/thumbor/thumbor/blob/master/thumbor/handlers/upload.py)
Don't forget to also check the tests and docs on uploading.
## Task Relevance
Removing code from thumbor's codebase is as important as adding. The leaner the project is, the faster we can move and onboard new people.
## Further Details
When you pick up this task, tag @heynemann to create the project for you and add you to the project team of contributors.
Then:
* Create the structure to run tests
* Add tests
* Add code
* Add github action to run the code on commit
* Add linting
* Add linting to the github action
* Add a nice README explaining how to use this new extensions
* Tag a version 0.1.0 (`git tag 0.1.0 && git push --tags`)
* Ask @heynemann to submit to pypi and he'll add you to the list of maintainers.
## Task Workflow
To complete this task, follow this workflow:
1. [Fork the involved repositories](http://help.github.com/fork-a-repo/)
2. In each repository there's a documentation on how to install dependencies and initialize your environment
3. Hack, in no particular order:
- Write code & tests
- Write new tests
- Write docs
- Improve design
- Check that all tests pass
- Repeat until you're satisfied
4. [Submit a pull request](https://docs.github.com/en/github/collaborating-with-pull-requests/proposing-changes-to-your-work-with-pull-requests/creating-a-pull-request).
|
main
|
remove thumbor upload into its own project areas of expertise thumbor open source testing python maintainability creating your own open source project will teach a lot of the concepts required for participating in another s open source project great experience summary create a new project thumbor upload in thumbor s org and move the upload functionality into it once it has been moved create a task here to remove upload from thumbor s codebase or do it in two prs yourself involved modules don t forget to also check the tests and docs on uploading task relevance removing code from thumbor s codebase is as important as adding the leaner the project is the faster we can move and onboard new people further details when you pick up this task tag heynemann to create the project for you and add you to the project team of contributors then create the structure to run tests add tests add code add github action to run the code on commit add linting add linting to the github action add a nice readme explaining how to use this new extensions tag a version git tag git push tags ask heynemann to submit to pypi and he ll add you to the list of maintainers task workflow to complete this task follow this workflow in each repository there s a documentation on how to install dependencies and initialize your environment hack in no particular order write code tests write new tests write docs improve design check that all tests pass repeat until you re satisfied
| 1
|
3,903
| 17,376,851,919
|
IssuesEvent
|
2021-07-30 23:28:21
|
chorman0773/Clever-ISA
|
https://api.github.com/repos/chorman0773/Clever-ISA
|
closed
|
Long Immediate Operand references "Operand Control Structure" but the term is not defined anywhere
|
I-unclear S-blocked-on-maintainer X-main
|
Long Immediate Operand references "Operand Control Structure" but the term is not defined anywhere
|
True
|
Long Immediate Operand references "Operand Control Structure" but the term is not defined anywhere - Long Immediate Operand references "Operand Control Structure" but the term is not defined anywhere
|
main
|
long immediate operand references operand control structure but the term is not defined anywhere long immediate operand references operand control structure but the term is not defined anywhere
| 1
|
227,547
| 18,068,431,798
|
IssuesEvent
|
2021-09-20 22:09:09
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: decommission/mixed-versions failed [should stop after beta1]
|
C-test-failure O-robot O-roachtest branch-master GA-blocker
|
roachtest.decommission/mixed-versions [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3462983&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3462983&tab=artifacts#/decommission/mixed-versions) on master @ [78c6771c7e9f7ba6431f44b067f27e0857341374](https://github.com/cockroachdb/cockroach/commits/78c6771c7e9f7ba6431f44b067f27e0857341374):
```
The test failed on branch=master, cloud=gce:
test timed out (see artifacts for details)
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*decommission/mixed-versions.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
2.0
|
roachtest: decommission/mixed-versions failed [should stop after beta1] - roachtest.decommission/mixed-versions [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3462983&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3462983&tab=artifacts#/decommission/mixed-versions) on master @ [78c6771c7e9f7ba6431f44b067f27e0857341374](https://github.com/cockroachdb/cockroach/commits/78c6771c7e9f7ba6431f44b067f27e0857341374):
```
The test failed on branch=master, cloud=gce:
test timed out (see artifacts for details)
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*decommission/mixed-versions.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
non_main
|
roachtest decommission mixed versions failed roachtest decommission mixed versions with on master the test failed on branch master cloud gce test timed out see artifacts for details reproduce see cc cockroachdb kv triage
| 0
|
548,064
| 16,056,442,761
|
IssuesEvent
|
2021-04-23 06:10:49
|
ucfopen/UDOIT
|
https://api.github.com/repos/ucfopen/UDOIT
|
closed
|
imgAltIsDifferent triggered even if aria-label used
|
enhancement low priority
|
The imgAltIsDifferent rule is triggered even if `aria-label` is used to provide alternative text for an `<img>` tag that has an `alt` attribute set to the file name.

|
1.0
|
imgAltIsDifferent triggered even if aria-label used - The imgAltIsDifferent rule is triggered even if `aria-label` is used to provide alternative text for an `<img>` tag that has an `alt` attribute set to the file name.

|
non_main
|
imgaltisdifferent triggered even if aria label used the imgaltisdifferent rule is triggered even if aria label is used to provide alternative text for an tag that has an alt attribute set to the file name
| 0
|
2,463
| 8,639,901,175
|
IssuesEvent
|
2018-11-23 22:31:46
|
F5OEO/rpitx
|
https://api.github.com/repos/F5OEO/rpitx
|
closed
|
At RPi 3 rpitx randomly shifting frequency during TX
|
V1 related (not maintained)
|
I've experienced a strange shifting of 2.8 MHz upwards from the carrier. Randomly the nominal 14MHz transmission is just shifted up to 16.8 MHz... then back.
Using latest Raspbian Jessie (as of today 22/04/2016), all the update, upgrade, disti upgrade done before. Cloned and installed riptx just right now from github.


|
True
|
At RPi 3 rpitx randomly shifting frequency during TX - I've experienced a strange shifting of 2.8 MHz upwards from the carrier. Randomly the nominal 14MHz transmission is just shifted up to 16.8 MHz... then back.
Using latest Raspbian Jessie (as of today 22/04/2016), all the update, upgrade, disti upgrade done before. Cloned and installed riptx just right now from github.


|
main
|
at rpi rpitx randomly shifting frequency during tx i ve experienced a strange shifting of mhz upwards from the carrier randomly the nominal transmission is just shifted up to mhz then back using latest raspbian jessie as of today all the update upgrade disti upgrade done before cloned and installed riptx just right now from github
| 1
|
707,082
| 24,294,236,927
|
IssuesEvent
|
2022-09-29 08:47:57
|
FinalProject-AIPARK/JenaPark-BE
|
https://api.github.com/repos/FinalProject-AIPARK/JenaPark-BE
|
closed
|
회원 정보 받는 api
|
Priority: Medium Status: Done
|
## 설명
로그인 성공 후 받는 토큰을 통해서 회원정보를 불러오느 api
## 할 일
- [x] Controller 수정
- [x] Service 수정
- [x] ResponseDto 생성
## 기타
참조 및 링크를 첨부하시오.
|
1.0
|
회원 정보 받는 api - ## 설명
로그인 성공 후 받는 토큰을 통해서 회원정보를 불러오느 api
## 할 일
- [x] Controller 수정
- [x] Service 수정
- [x] ResponseDto 생성
## 기타
참조 및 링크를 첨부하시오.
|
non_main
|
회원 정보 받는 api 설명 로그인 성공 후 받는 토큰을 통해서 회원정보를 불러오느 api 할 일 controller 수정 service 수정 responsedto 생성 기타 참조 및 링크를 첨부하시오
| 0
|
369,639
| 10,915,735,893
|
IssuesEvent
|
2019-11-21 11:49:43
|
incognitochain/incognito-wallet
|
https://api.github.com/repos/incognitochain/incognito-wallet
|
closed
|
Node go from online to offline
|
Priority: Critical Type: Bug
|
Node from online to offline although it's online before and staked successfully
|
1.0
|
Node go from online to offline - Node from online to offline although it's online before and staked successfully
|
non_main
|
node go from online to offline node from online to offline although it s online before and staked successfully
| 0
|
3,058
| 11,454,973,684
|
IssuesEvent
|
2020-02-06 18:08:23
|
18F/cg-product
|
https://api.github.com/repos/18F/cg-product
|
closed
|
As a federalist operator, I want to be able to migrate S3 origins in Cloud Front while the cdn broker is unavailable.
|
contractor-3-maintainability
|
The Federalist team has a backlog of sites awaiting S3 origin migration.
## Acceptance Criteria
* [ ] GIVEN an existing federalist site \
AND the need to update the origin of the site \
WHEN the federalist team invokes a migration via `cf task` \
THEN the origin is updated \
AND certification regeneration is NOT triggered.
---
## Implementation sketch
The federalist team executes migrations in tasks in Cloud Foundry. The task is defined here https://github.com/18F/federalist/tree/apburnes/bucket-migration/scripts/migrate-site-bucket. The task handles the bucket creation and replication.
The `update-service` call is here: https://github.com/18F/federalist/blob/6feb3c48147e7ea35d7a157a03a7e841bf393f83/scripts/migrate-site-bucket/migrate-site-bucket.sh#L364. This call in unavailable when the broker is unavailable.
The existing CDN broker is still running as it is processing renewals. The plan is to invoke the update-service endpoint via a POST. However, we must be sure the update call does not trigger a certificate regeneration. Therefore we need to filter requests to ONLY s3 origin migrations.
* [ ] Verify calls to update-service to change the origin will not trigger a certificate regeneration
* [ ] Implement a proxy filter to the update-service call to ensure only s3 origin migration updates are permitted
* [ ] Inform the federalist team to replace the `update-service` call in their task with a curl to the proxy filter
## Security considerations
The proxy application will require basic authentication at a minimum. The federalist team will need these credentials.
The proxy application will ensure we do not need to disclose broker credentials to the federalist team. Only the proxy will need these credentials.
|
True
|
As a federalist operator, I want to be able to migrate S3 origins in Cloud Front while the cdn broker is unavailable. - The Federalist team has a backlog of sites awaiting S3 origin migration.
## Acceptance Criteria
* [ ] GIVEN an existing federalist site \
AND the need to update the origin of the site \
WHEN the federalist team invokes a migration via `cf task` \
THEN the origin is updated \
AND certification regeneration is NOT triggered.
---
## Implementation sketch
The federalist team executes migrations in tasks in Cloud Foundry. The task is defined here https://github.com/18F/federalist/tree/apburnes/bucket-migration/scripts/migrate-site-bucket. The task handles the bucket creation and replication.
The `update-service` call is here: https://github.com/18F/federalist/blob/6feb3c48147e7ea35d7a157a03a7e841bf393f83/scripts/migrate-site-bucket/migrate-site-bucket.sh#L364. This call in unavailable when the broker is unavailable.
The existing CDN broker is still running as it is processing renewals. The plan is to invoke the update-service endpoint via a POST. However, we must be sure the update call does not trigger a certificate regeneration. Therefore we need to filter requests to ONLY s3 origin migrations.
* [ ] Verify calls to update-service to change the origin will not trigger a certificate regeneration
* [ ] Implement a proxy filter to the update-service call to ensure only s3 origin migration updates are permitted
* [ ] Inform the federalist team to replace the `update-service` call in their task with a curl to the proxy filter
## Security considerations
The proxy application will require basic authentication at a minimum. The federalist team will need these credentials.
The proxy application will ensure we do not need to disclose broker credentials to the federalist team. Only the proxy will need these credentials.
|
main
|
as a federalist operator i want to be able to migrate origins in cloud front while the cdn broker is unavailable the federalist team has a backlog of sites awaiting origin migration acceptance criteria given an existing federalist site and the need to update the origin of the site when the federalist team invokes a migration via cf task then the origin is updated and certification regeneration is not triggered implementation sketch the federalist team executes migrations in tasks in cloud foundry the task is defined here the task handles the bucket creation and replication the update service call is here this call in unavailable when the broker is unavailable the existing cdn broker is still running as it is processing renewals the plan is to invoke the update service endpoint via a post however we must be sure the update call does not trigger a certificate regeneration therefore we need to filter requests to only origin migrations verify calls to update service to change the origin will not trigger a certificate regeneration implement a proxy filter to the update service call to ensure only origin migration updates are permitted inform the federalist team to replace the update service call in their task with a curl to the proxy filter security considerations the proxy application will require basic authentication at a minimum the federalist team will need these credentials the proxy application will ensure we do not need to disclose broker credentials to the federalist team only the proxy will need these credentials
| 1
|
90,898
| 15,856,317,119
|
IssuesEvent
|
2021-04-08 02:03:22
|
rvvergara/react-native-learning-starter
|
https://api.github.com/repos/rvvergara/react-native-learning-starter
|
opened
|
CVE-2021-23337 (High) detected in lodash-4.17.14.tgz, lodash-4.17.15.tgz
|
security vulnerability
|
## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-4.17.14.tgz</b>, <b>lodash-4.17.15.tgz</b></p></summary>
<p>
<details><summary><b>lodash-4.17.14.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.14.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.14.tgz</a></p>
<p>Path to dependency file: react-native-learning-starter/package.json</p>
<p>Path to vulnerable library: react-native-learning-starter/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- eslint-6.8.0.tgz (Root Library)
- :x: **lodash-4.17.14.tgz** (Vulnerable Library)
</details>
<details><summary><b>lodash-4.17.15.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: react-native-learning-starter/package.json</p>
<p>Path to vulnerable library: react-native-learning-starter/node_modules/eslint/node_modules/inquirer/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- eslint-6.8.0.tgz (Root Library)
- inquirer-7.0.4.tgz
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23337 (High) detected in lodash-4.17.14.tgz, lodash-4.17.15.tgz - ## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-4.17.14.tgz</b>, <b>lodash-4.17.15.tgz</b></p></summary>
<p>
<details><summary><b>lodash-4.17.14.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.14.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.14.tgz</a></p>
<p>Path to dependency file: react-native-learning-starter/package.json</p>
<p>Path to vulnerable library: react-native-learning-starter/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- eslint-6.8.0.tgz (Root Library)
- :x: **lodash-4.17.14.tgz** (Vulnerable Library)
</details>
<details><summary><b>lodash-4.17.15.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: react-native-learning-starter/package.json</p>
<p>Path to vulnerable library: react-native-learning-starter/node_modules/eslint/node_modules/inquirer/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- eslint-6.8.0.tgz (Root Library)
- inquirer-7.0.4.tgz
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_main
|
cve high detected in lodash tgz lodash tgz cve high severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz lodash modular utilities library home page a href path to dependency file react native learning starter package json path to vulnerable library react native learning starter node modules lodash package json dependency hierarchy eslint tgz root library x lodash tgz vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file react native learning starter package json path to vulnerable library react native learning starter node modules eslint node modules inquirer node modules lodash package json dependency hierarchy eslint tgz root library inquirer tgz x lodash tgz vulnerable library vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource
| 0
|
185,362
| 21,788,706,747
|
IssuesEvent
|
2022-05-14 15:02:05
|
GNS3/gns3-web-ui
|
https://api.github.com/repos/GNS3/gns3-web-ui
|
closed
|
CVE-2022-24773 (Medium) detected in node-forge-0.10.0.tgz
|
security vulnerability
|
## CVE-2022-24773 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-12.2.13.tgz (Root Library)
- webpack-dev-server-3.11.2.tgz
- selfsigned-1.10.11.tgz
- :x: **node-forge-0.10.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GNS3/gns3-web-ui/commit/81ca3e2af27af6f8f6569c51ebf6bc06e70f8030">81ca3e2af27af6f8f6569c51ebf6bc06e70f8030</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not properly check `DigestInfo` for a proper ASN.1 structure. This can lead to successful verification with signatures that contain invalid structures but a valid digest. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds.
<p>Publish Date: 2022-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24773>CVE-2022-24773</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24773">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24773</a></p>
<p>Release Date: 2022-03-18</p>
<p>Fix Resolution (node-forge): 1.3.0</p>
<p>Direct dependency fix Resolution (@angular-devkit/build-angular): 13.3.3</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"@angular-devkit/build-angular","packageVersion":"12.2.13","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"@angular-devkit/build-angular:12.2.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"13.3.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2022-24773","vulnerabilityDetails":"Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not properly check `DigestInfo` for a proper ASN.1 structure. This can lead to successful verification with signatures that contain invalid structures but a valid digest. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24773","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2022-24773 (Medium) detected in node-forge-0.10.0.tgz - ## CVE-2022-24773 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-12.2.13.tgz (Root Library)
- webpack-dev-server-3.11.2.tgz
- selfsigned-1.10.11.tgz
- :x: **node-forge-0.10.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GNS3/gns3-web-ui/commit/81ca3e2af27af6f8f6569c51ebf6bc06e70f8030">81ca3e2af27af6f8f6569c51ebf6bc06e70f8030</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not properly check `DigestInfo` for a proper ASN.1 structure. This can lead to successful verification with signatures that contain invalid structures but a valid digest. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds.
<p>Publish Date: 2022-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24773>CVE-2022-24773</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24773">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24773</a></p>
<p>Release Date: 2022-03-18</p>
<p>Fix Resolution (node-forge): 1.3.0</p>
<p>Direct dependency fix Resolution (@angular-devkit/build-angular): 13.3.3</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"@angular-devkit/build-angular","packageVersion":"12.2.13","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"@angular-devkit/build-angular:12.2.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"13.3.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2022-24773","vulnerabilityDetails":"Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not properly check `DigestInfo` for a proper ASN.1 structure. This can lead to successful verification with signatures that contain invalid structures but a valid digest. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24773","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_main
|
cve medium detected in node forge tgz cve medium severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file package json path to vulnerable library node modules node forge package json dependency hierarchy build angular tgz root library webpack dev server tgz selfsigned tgz x node forge tgz vulnerable library found in head commit a href found in base branch master vulnerability details forge also called node forge is a native implementation of transport layer security in javascript prior to version rsa pkcs signature verification code does not properly check digestinfo for a proper asn structure this can lead to successful verification with signatures that contain invalid structures but a valid digest the issue has been addressed in node forge version there are currently no known workarounds publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge direct dependency fix resolution angular devkit build angular check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree angular devkit build angular isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails forge also called node forge is a native implementation of transport layer security in javascript prior to version rsa pkcs signature verification code does not properly check digestinfo for a proper asn structure this can lead to successful verification with signatures that contain invalid structures but a valid digest the issue has been addressed in node forge version there are currently no known workarounds vulnerabilityurl
| 0
|
170,620
| 14,265,909,897
|
IssuesEvent
|
2020-11-20 17:51:00
|
postmanlabs/postman-app-support
|
https://api.github.com/repos/postmanlabs/postman-app-support
|
closed
|
Reorder Collection Versions
|
feature product/documentation
|
While publishing my collection with multiple versions, right now I'm not able to reorder it. Is there any way to reorder my versions at the time of publishing my collection?
|
1.0
|
Reorder Collection Versions - While publishing my collection with multiple versions, right now I'm not able to reorder it. Is there any way to reorder my versions at the time of publishing my collection?
|
non_main
|
reorder collection versions while publishing my collection with multiple versions right now i m not able to reorder it is there any way to reorder my versions at the time of publishing my collection
| 0
|
3,795
| 16,218,791,910
|
IssuesEvent
|
2021-05-06 01:13:19
|
truecharts/apps
|
https://api.github.com/repos/truecharts/apps
|
closed
|
Add Podgrab
|
New App Request No-Maintainer
|
An extremely useful application for podcasts.
> Podgrab is a is a self-hosted podcast manager which automatically downloads latest podcast episodes.
Application Repo: https://github.com/akhilrex/podgrab
Thank you!
|
True
|
Add Podgrab - An extremely useful application for podcasts.
> Podgrab is a is a self-hosted podcast manager which automatically downloads latest podcast episodes.
Application Repo: https://github.com/akhilrex/podgrab
Thank you!
|
main
|
add podgrab an extremely useful application for podcasts podgrab is a is a self hosted podcast manager which automatically downloads latest podcast episodes application repo thank you
| 1
|
568,329
| 16,964,908,257
|
IssuesEvent
|
2021-06-29 09:41:42
|
status-im/StatusQ
|
https://api.github.com/repos/status-im/StatusQ
|
opened
|
Implement `StatusWindowsToolBar` component
|
priority 3: important type: feature
|
A component to render Microsoft Windows specific window toolbars.
Here's what they look like as per current design:
<img width="683" alt="Screenshot 2021-06-29 at 11 41 06" src="https://user-images.githubusercontent.com/445106/123775738-f201d100-d8ce-11eb-8529-0e71a953f652.png">
Figma: https://www.figma.com/file/MtAO3a7HnEH5xjCDVNilS7/%F0%9F%8E%A8-Design-System-%E2%8E%9C-Desktop?node-id=1%3A2
|
1.0
|
Implement `StatusWindowsToolBar` component - A component to render Microsoft Windows specific window toolbars.
Here's what they look like as per current design:
<img width="683" alt="Screenshot 2021-06-29 at 11 41 06" src="https://user-images.githubusercontent.com/445106/123775738-f201d100-d8ce-11eb-8529-0e71a953f652.png">
Figma: https://www.figma.com/file/MtAO3a7HnEH5xjCDVNilS7/%F0%9F%8E%A8-Design-System-%E2%8E%9C-Desktop?node-id=1%3A2
|
non_main
|
implement statuswindowstoolbar component a component to render microsoft windows specific window toolbars here s what they look like as per current design img width alt screenshot at src figma
| 0
|
635,210
| 20,381,981,307
|
IssuesEvent
|
2022-02-21 23:32:50
|
NerdyNomads/Text-Savvy
|
https://api.github.com/repos/NerdyNomads/Text-Savvy
|
opened
|
Create mock up of extension
|
high priority front-end
|
Create a UI mock up of the chrome extension dropdown panel with Figma.
Here is an example of UI design from an extension (Bubbles).

|
1.0
|
Create mock up of extension - Create a UI mock up of the chrome extension dropdown panel with Figma.
Here is an example of UI design from an extension (Bubbles).

|
non_main
|
create mock up of extension create a ui mock up of the chrome extension dropdown panel with figma here is an example of ui design from an extension bubbles
| 0
|
5,539
| 27,735,433,247
|
IssuesEvent
|
2023-03-15 10:53:30
|
precice/precice
|
https://api.github.com/repos/precice/precice
|
closed
|
Nightly build of dockerimage precice/precice:develop
|
maintainability compatibility
|
**Please describe the problem you are trying to solve.**
The python bindings use the docker image `precice/precice` provided via (https://github.com/precice/precice/blob/v2.3.0/.github/workflows/release-docker.yml) in their CI pipeline to create and push a docker image with the python bindings `precice/python-bindings` (https://github.com/precice/python-bindings/blob/develop/.github/workflows/build-docker.yml).
Currently `precice/precice` is only updated, when there is a release. See https://github.com/precice/precice/blob/v2.3.0/.github/workflows/release-docker.yml. But for being able to run tests on the latest version of preCICE, we actually need `precice/precice:develop`. This will also help us downstream to detect early.
**Describe the solution you propose.**
Change the action https://github.com/precice/precice/blob/v2.3.0/.github/workflows/release-docker.yml to create and push `precice/precice:develop` nightly.
**Describe alternatives you've considered**
We will use `precice/precice:latest` in the python bindings until this issue is solved. However, this is not optimal, because this does not allow us to detect compatibility issues with the develop version of preCICE early. We will only see compatibility issues as soon as preCICE is released and will, therefore, only be able to fix them after a release of preCICE and before a release of the bindings.
**Additional context**
As far as I understand we want to support the path marked as "nigthly" from the preCICE paper (https://arxiv.org/pdf/2109.14470.pdf). Solving this issue helps to support this path.

|
True
|
Nightly build of dockerimage precice/precice:develop - **Please describe the problem you are trying to solve.**
The python bindings use the docker image `precice/precice` provided via (https://github.com/precice/precice/blob/v2.3.0/.github/workflows/release-docker.yml) in their CI pipeline to create and push a docker image with the python bindings `precice/python-bindings` (https://github.com/precice/python-bindings/blob/develop/.github/workflows/build-docker.yml).
Currently `precice/precice` is only updated, when there is a release. See https://github.com/precice/precice/blob/v2.3.0/.github/workflows/release-docker.yml. But for being able to run tests on the latest version of preCICE, we actually need `precice/precice:develop`. This will also help us downstream to detect early.
**Describe the solution you propose.**
Change the action https://github.com/precice/precice/blob/v2.3.0/.github/workflows/release-docker.yml to create and push `precice/precice:develop` nightly.
**Describe alternatives you've considered**
We will use `precice/precice:latest` in the python bindings until this issue is solved. However, this is not optimal, because this does not allow us to detect compatibility issues with the develop version of preCICE early. We will only see compatibility issues as soon as preCICE is released and will, therefore, only be able to fix them after a release of preCICE and before a release of the bindings.
**Additional context**
As far as I understand we want to support the path marked as "nigthly" from the preCICE paper (https://arxiv.org/pdf/2109.14470.pdf). Solving this issue helps to support this path.

|
main
|
nightly build of dockerimage precice precice develop please describe the problem you are trying to solve the python bindings use the docker image precice precice provided via in their ci pipeline to create and push a docker image with the python bindings precice python bindings currently precice precice is only updated when there is a release see but for being able to run tests on the latest version of precice we actually need precice precice develop this will also help us downstream to detect early describe the solution you propose change the action to create and push precice precice develop nightly describe alternatives you ve considered we will use precice precice latest in the python bindings until this issue is solved however this is not optimal because this does not allow us to detect compatibility issues with the develop version of precice early we will only see compatibility issues as soon as precice is released and will therefore only be able to fix them after a release of precice and before a release of the bindings additional context as far as i understand we want to support the path marked as nigthly from the precice paper solving this issue helps to support this path
| 1
|
757,551
| 26,517,531,186
|
IssuesEvent
|
2023-01-18 22:12:31
|
SlimeVR/SlimeVR-Tracker-ESP
|
https://api.github.com/repos/SlimeVR/SlimeVR-Tracker-ESP
|
closed
|
[Magneto] Handle infinite samples with constant memory usage
|
Type: Feature Request Priority: Low Status: Unlabeled
|
the first thing magneto does with the input sample is first expand each sample into a 10 item row/column, treat the input as a 10xN matrix, and multiply it by it's own transpose, resulting in a 10x10 matrix.
expanding out the math, it turns out that taking each sample as a 10x1 matrix, multiplying by its own transpose, then summing each resulting matrix together results in the same matrix, except for floating point imprecision. I'm not sure whether the result is more or less precise, but according to my testing it's fairly close.
https://blog.demofox.org/2017/01/02/incremental-least-squares-surface-and-hyper-volume-fitting/ suggests we can do it with even less space than a full 10x10 matrix due to the structure of the input, 19 elements instead of 100. nevertheless, even if we don't go that far, we can still save quite a bit of memory. (currently, for mpu9250 at least, we save 3\*300 elements, which gets expanded to 10\*300 elements, before we end up with the 10x10 matrix of 100 elements.)
However, since we'd be doing the calculations on the fly, we lose the current method of outlier rejection. If we wanted to keep outlier rejection, we'd need to find a way to do it 'online', as samples are streaming in.
|
1.0
|
[Magneto] Handle infinite samples with constant memory usage - the first thing magneto does with the input sample is first expand each sample into a 10 item row/column, treat the input as a 10xN matrix, and multiply it by it's own transpose, resulting in a 10x10 matrix.
expanding out the math, it turns out that taking each sample as a 10x1 matrix, multiplying by its own transpose, then summing each resulting matrix together results in the same matrix, except for floating point imprecision. I'm not sure whether the result is more or less precise, but according to my testing it's fairly close.
https://blog.demofox.org/2017/01/02/incremental-least-squares-surface-and-hyper-volume-fitting/ suggests we can do it with even less space than a full 10x10 matrix due to the structure of the input, 19 elements instead of 100. nevertheless, even if we don't go that far, we can still save quite a bit of memory. (currently, for mpu9250 at least, we save 3\*300 elements, which gets expanded to 10\*300 elements, before we end up with the 10x10 matrix of 100 elements.)
However, since we'd be doing the calculations on the fly, we lose the current method of outlier rejection. If we wanted to keep outlier rejection, we'd need to find a way to do it 'online', as samples are streaming in.
|
non_main
|
handle infinite samples with constant memory usage the first thing magneto does with the input sample is first expand each sample into a item row column treat the input as a matrix and multiply it by it s own transpose resulting in a matrix expanding out the math it turns out that taking each sample as a matrix multiplying by its own transpose then summing each resulting matrix together results in the same matrix except for floating point imprecision i m not sure whether the result is more or less precise but according to my testing it s fairly close suggests we can do it with even less space than a full matrix due to the structure of the input elements instead of nevertheless even if we don t go that far we can still save quite a bit of memory currently for at least we save elements which gets expanded to elements before we end up with the matrix of elements however since we d be doing the calculations on the fly we lose the current method of outlier rejection if we wanted to keep outlier rejection we d need to find a way to do it online as samples are streaming in
| 0
|
4,892
| 25,124,310,645
|
IssuesEvent
|
2022-11-09 10:31:21
|
goharbor/community
|
https://api.github.com/repos/goharbor/community
|
closed
|
Pierre PÉRONNET changed employment
|
area/maintainer-nomination
|
Hi @holyhope, I see you have left OVHcloud are you still planing to be part of the maintainers team in the new company?
Thanks!
|
True
|
Pierre PÉRONNET changed employment - Hi @holyhope, I see you have left OVHcloud are you still planing to be part of the maintainers team in the new company?
Thanks!
|
main
|
pierre péronnet changed employment hi holyhope i see you have left ovhcloud are you still planing to be part of the maintainers team in the new company thanks
| 1
|
19,032
| 6,664,492,524
|
IssuesEvent
|
2017-10-02 20:21:14
|
dart-lang/build
|
https://api.github.com/repos/dart-lang/build
|
closed
|
Checking for existing outputs fails if an intermediate output is deleted
|
package:build_runner
|
Situation:
- There is a source file `source.dart` and two phases of builders `.dart` -> `.phase1` and `.phase1` -> `.phase2`
- The `.dart_tool` directory does not exist so there is no serialized asset graph
- `source.phase1` does *not* exist on disk, `source.phase2` *does* exist on disk.
Checking for existing outputs will not see a conflict with `source.phase2` because we only feed inputs that exist on disk [here](https://github.com/dart-lang/build/blob/465060fa2b9ab0f576c90c8385b49922e61c5fa2/build_runner/lib/src/generate/build_impl.dart#L318) - not outputs that we expect will exist.
If we run a build we will write `source.phase1` and then get an exception "Cannot overwrite inputs." when trying to write `sources.phase2`.
|
1.0
|
Checking for existing outputs fails if an intermediate output is deleted - Situation:
- There is a source file `source.dart` and two phases of builders `.dart` -> `.phase1` and `.phase1` -> `.phase2`
- The `.dart_tool` directory does not exist so there is no serialized asset graph
- `source.phase1` does *not* exist on disk, `source.phase2` *does* exist on disk.
Checking for existing outputs will not see a conflict with `source.phase2` because we only feed inputs that exist on disk [here](https://github.com/dart-lang/build/blob/465060fa2b9ab0f576c90c8385b49922e61c5fa2/build_runner/lib/src/generate/build_impl.dart#L318) - not outputs that we expect will exist.
If we run a build we will write `source.phase1` and then get an exception "Cannot overwrite inputs." when trying to write `sources.phase2`.
|
non_main
|
checking for existing outputs fails if an intermediate output is deleted situation there is a source file source dart and two phases of builders dart and the dart tool directory does not exist so there is no serialized asset graph source does not exist on disk source does exist on disk checking for existing outputs will not see a conflict with source because we only feed inputs that exist on disk not outputs that we expect will exist if we run a build we will write source and then get an exception cannot overwrite inputs when trying to write sources
| 0
|
4,082
| 19,285,632,005
|
IssuesEvent
|
2021-12-11 00:12:18
|
aws/aws-sam-cli
|
https://api.github.com/repos/aws/aws-sam-cli
|
closed
|
Debug dotnetcore 2.1 in Visual Studio 2017/2019
|
area/ide type/feature area/debugging stage/pm-review maintainer/need-response
|
### Describe your idea/feature/enhancement
We're currently migrating from a monolith ASP.NET Core app to a serverless architecture and SAM proved to be a valuable asset. However, the team is much more comfortable with VS (currently 2017 and we're moving to 2019) and not that much with VSCode.
Now that SAM supports debugging dotnetcore functions using VSCode (#568), we're wondering if this can ever translate to the full Visual Studio 2017/2019?
|
True
|
Debug dotnetcore 2.1 in Visual Studio 2017/2019 - ### Describe your idea/feature/enhancement
We're currently migrating from a monolith ASP.NET Core app to a serverless architecture and SAM proved to be a valuable asset. However, the team is much more comfortable with VS (currently 2017 and we're moving to 2019) and not that much with VSCode.
Now that SAM supports debugging dotnetcore functions using VSCode (#568), we're wondering if this can ever translate to the full Visual Studio 2017/2019?
|
main
|
debug dotnetcore in visual studio describe your idea feature enhancement we re currently migrating from a monolith asp net core app to a serverless architecture and sam proved to be a valuable asset however the team is much more comfortable with vs currently and we re moving to and not that much with vscode now that sam supports debugging dotnetcore functions using vscode we re wondering if this can ever translate to the full visual studio
| 1
|
516,145
| 14,975,963,202
|
IssuesEvent
|
2021-01-28 07:12:55
|
threefoldtech/home
|
https://api.github.com/repos/threefoldtech/home
|
opened
|
Deployed blog appears in the deployed solutions page but not in the deployed blogs overview.
|
priority_major type_bug
|
In VDC: jetserthing (Gold, testnet) a deployed blog using the example blog source from the manual results in successful deployment. But the deployed solution pages (generic and specific) display different results.



|
1.0
|
Deployed blog appears in the deployed solutions page but not in the deployed blogs overview. - In VDC: jetserthing (Gold, testnet) a deployed blog using the example blog source from the manual results in successful deployment. But the deployed solution pages (generic and specific) display different results.



|
non_main
|
deployed blog appears in the deployed solutions page but not in the deployed blogs overview in vdc jetserthing gold testnet a deployed blog using the example blog source from the manual results in successful deployment but the deployed solution pages generic and specific display different results
| 0
|
83,102
| 16,091,350,336
|
IssuesEvent
|
2021-04-26 17:05:31
|
microsoft/vscode-jupyter
|
https://api.github.com/repos/microsoft/vscode-jupyter
|
closed
|
No scrollbar generated for large outputs
|
bug upstream-vscode vscode-notebook
|
## Environment data
- VS Code version: 1.56.0-insider
- Jupyter Extension version (available under the Extensions sidebar):
- Python Extension version (available under the Extensions sidebar): v2021.6.780948196
- OS (Windows | Mac | Linux distro) and version: Ubuntu 20.04.2 LTS
- Python and/or Anaconda version: Python 3.8.3, conda 4.8.5
- Type of virtual environment used (N/A | venv | virtualenv | conda | ...): conda
- Jupyter server running: Local
## Expected behaviour
When running a code cell which generates large output (either text, or a series of plots), the ouput cells size should be limited and a scrollbar within the output cell should appear.
(Jupyter: Enable Scrolling For Cell Outputs
is active,
Jupyter: Max Output Size
is 50 (was 200 previously, but lowering it did not help)
## Actual behaviour
The entire output is shown as one large chunk without scrollbar.
## Steps to reproduce:
create a new file
save a .ipynb file
insert into the first code cell:
`for i in range(50):
print("test"`
insert into a second cell:
`from matplotlib import pyplot as plt
for i in range(5):
plt.figure()
plt.plot(range(10), range(10))
plt.show()`
run the notebook. (larger numbers lead to the same problem, so it is not just that the threshold is not reached
|
2.0
|
No scrollbar generated for large outputs - ## Environment data
- VS Code version: 1.56.0-insider
- Jupyter Extension version (available under the Extensions sidebar):
- Python Extension version (available under the Extensions sidebar): v2021.6.780948196
- OS (Windows | Mac | Linux distro) and version: Ubuntu 20.04.2 LTS
- Python and/or Anaconda version: Python 3.8.3, conda 4.8.5
- Type of virtual environment used (N/A | venv | virtualenv | conda | ...): conda
- Jupyter server running: Local
## Expected behaviour
When running a code cell which generates large output (either text, or a series of plots), the ouput cells size should be limited and a scrollbar within the output cell should appear.
(Jupyter: Enable Scrolling For Cell Outputs
is active,
Jupyter: Max Output Size
is 50 (was 200 previously, but lowering it did not help)
## Actual behaviour
The entire output is shown as one large chunk without scrollbar.
## Steps to reproduce:
create a new file
save a .ipynb file
insert into the first code cell:
`for i in range(50):
print("test"`
insert into a second cell:
`from matplotlib import pyplot as plt
for i in range(5):
plt.figure()
plt.plot(range(10), range(10))
plt.show()`
run the notebook. (larger numbers lead to the same problem, so it is not just that the threshold is not reached
|
non_main
|
no scrollbar generated for large outputs environment data vs code version insider jupyter extension version available under the extensions sidebar python extension version available under the extensions sidebar os windows mac linux distro and version ubuntu lts python and or anaconda version python conda type of virtual environment used n a venv virtualenv conda conda jupyter server running local expected behaviour when running a code cell which generates large output either text or a series of plots the ouput cells size should be limited and a scrollbar within the output cell should appear jupyter enable scrolling for cell outputs is active jupyter max output size is was previously but lowering it did not help actual behaviour the entire output is shown as one large chunk without scrollbar steps to reproduce create a new file save a ipynb file insert into the first code cell for i in range print test insert into a second cell from matplotlib import pyplot as plt for i in range plt figure plt plot range range plt show run the notebook larger numbers lead to the same problem so it is not just that the threshold is not reached
| 0
|
49,968
| 12,439,282,568
|
IssuesEvent
|
2020-05-26 09:50:33
|
docascod/DocsAsCode
|
https://api.github.com/repos/docascod/DocsAsCode
|
opened
|
default theme slides : add default images
|
enhancement fct_build
|
Add default (blank) images into slides default theme :
* title-page-background
* title-page-logo
* header-logo
* footer-logo
|
1.0
|
default theme slides : add default images - Add default (blank) images into slides default theme :
* title-page-background
* title-page-logo
* header-logo
* footer-logo
|
non_main
|
default theme slides add default images add default blank images into slides default theme title page background title page logo header logo footer logo
| 0
|
2,663
| 9,105,550,221
|
IssuesEvent
|
2019-02-20 21:06:35
|
lrozenblyum/chess
|
https://api.github.com/repos/lrozenblyum/chess
|
closed
|
IDE-specific derived resources
|
devenv maintainability
|
IDEA: Let's check whether we need the *.iml in version control.
When we change pom.xml, it's getting updated.
Find best practices.
https://intellij-support.jetbrains.com/hc/en-us/articles/206544839
- [x] IDEA: *.iml
- [x] IDEA: general storage
- [x] Eclipse: .settings
- [x] Eclipse; other
Caused by #245
|
True
|
IDE-specific derived resources - IDEA: Let's check whether we need the *.iml in version control.
When we change pom.xml, it's getting updated.
Find best practices.
https://intellij-support.jetbrains.com/hc/en-us/articles/206544839
- [x] IDEA: *.iml
- [x] IDEA: general storage
- [x] Eclipse: .settings
- [x] Eclipse; other
Caused by #245
|
main
|
ide specific derived resources idea let s check whether we need the iml in version control when we change pom xml it s getting updated find best practices idea iml idea general storage eclipse settings eclipse other caused by
| 1
|
14,018
| 24,208,421,355
|
IssuesEvent
|
2022-09-25 15:06:36
|
renovatebot/renovate
|
https://api.github.com/repos/renovatebot/renovate
|
opened
|
Error configuring GitLab CI_REGISTRY setting using registryAliases
|
type:bug status:requirements priority-5-triage
|
### How are you running Renovate?
Self-hosted
### If you're self-hosting Renovate, tell us what version of Renovate you run.
whitesource/renovate-on-prem v2.5.1 (renovate 32.185.3)
### If you're self-hosting Renovate, select which platform you are using.
GitLab self-hosted
### If you're self-hosting Renovate, tell us what version of the platform you run.
GitLab EE
### Was this something which used to work for you, and then stopped?
I never saw this working
### Describe the bug
As initially discussed [here](https://github.com/renovatebot/renovate/discussions/17777).
I'm having an issue configuring CI_REGISTRY following the documentation [here](https://docs.renovatebot.com/modules/manager/gitlabci/).
For a given GitLab repo I attempted to add a `registryAliases` entry by updating the `renovate.json` file as follows:
```
{
"$schema": "https://docs.renovatebot.com/renovate-schema.json",
"dependencyDashboard": true,
"registryAliases": {
"$CI_REGISTRY": "docker-registry.mycompany.net"
}
}
```
This approach does not work as I get the following `Action Required: Fix Renovate Configuration` generated under GitLab issues:
```
Location: renovate.json
Error type: The renovate configuration file contains some invalid settings
Message: Invalid 'registryAliases.registryAliases.$CI_REGISTRY' configuration: value is not a url
```
As mentioned in the initial discussion, it's expecting the URL to be in format `https://.....`. That validation might be too strict. E.g. in my example above, using `https://docker-registry.mycompany.net` will not generate the renovate error.
However, prepending the `https` results in incorrectly formatted docker images.
Perhaps there is an expectation that this setting belongs exclusively in `config.js`? In this case, a documentation fix seems in order.
Maybe the code needs to be modified to properly handle a `CI_REGISTRY` assignment?
### Relevant debug logs
<details><summary>Logs</summary>
```
```
</details>
### Have you created a minimal reproduction repository?
No reproduction repository
|
1.0
|
Error configuring GitLab CI_REGISTRY setting using registryAliases - ### How are you running Renovate?
Self-hosted
### If you're self-hosting Renovate, tell us what version of Renovate you run.
whitesource/renovate-on-prem v2.5.1 (renovate 32.185.3)
### If you're self-hosting Renovate, select which platform you are using.
GitLab self-hosted
### If you're self-hosting Renovate, tell us what version of the platform you run.
GitLab EE
### Was this something which used to work for you, and then stopped?
I never saw this working
### Describe the bug
As initially discussed [here](https://github.com/renovatebot/renovate/discussions/17777).
I'm having an issue configuring CI_REGISTRY following the documentation [here](https://docs.renovatebot.com/modules/manager/gitlabci/).
For a given GitLab repo I attempted to add a `registryAliases` entry by updating the `renovate.json` file as follows:
```
{
"$schema": "https://docs.renovatebot.com/renovate-schema.json",
"dependencyDashboard": true,
"registryAliases": {
"$CI_REGISTRY": "docker-registry.mycompany.net"
}
}
```
This approach does not work as I get the following `Action Required: Fix Renovate Configuration` generated under GitLab issues:
```
Location: renovate.json
Error type: The renovate configuration file contains some invalid settings
Message: Invalid 'registryAliases.registryAliases.$CI_REGISTRY' configuration: value is not a url
```
As mentioned in the initial discussion, it's expecting the URL to be in format `https://.....`. That validation might be too strict. E.g. in my example above, using `https://docker-registry.mycompany.net` will not generate the renovate error.
However, prepending the `https` results in incorrectly formatted docker images.
Perhaps there is an expectation that this setting belongs exclusively in `config.js`? In this case, a documentation fix seems in order.
Maybe the code needs to be modified to properly handle a `CI_REGISTRY` assignment?
### Relevant debug logs
<details><summary>Logs</summary>
```
```
</details>
### Have you created a minimal reproduction repository?
No reproduction repository
|
non_main
|
error configuring gitlab ci registry setting using registryaliases how are you running renovate self hosted if you re self hosting renovate tell us what version of renovate you run whitesource renovate on prem renovate if you re self hosting renovate select which platform you are using gitlab self hosted if you re self hosting renovate tell us what version of the platform you run gitlab ee was this something which used to work for you and then stopped i never saw this working describe the bug as initially discussed i m having an issue configuring ci registry following the documentation for a given gitlab repo i attempted to add a registryaliases entry by updating the renovate json file as follows schema dependencydashboard true registryaliases ci registry docker registry mycompany net this approach does not work as i get the following action required fix renovate configuration generated under gitlab issues location renovate json error type the renovate configuration file contains some invalid settings message invalid registryaliases registryaliases ci registry configuration value is not a url as mentioned in the initial discussion it s expecting the url to be in format that validation might be too strict e g in my example above using will not generate the renovate error however prepending the https results in incorrectly formatted docker images perhaps there is an expectation that this setting belongs exclusively in config js in this case a documentation fix seems in order maybe the code needs to be modified to properly handle a ci registry assignment relevant debug logs logs have you created a minimal reproduction repository no reproduction repository
| 0
|
169,630
| 26,834,761,518
|
IssuesEvent
|
2023-02-02 18:34:52
|
runtimeverification/haskell-backend
|
https://api.github.com/repos/runtimeverification/haskell-backend
|
closed
|
Validate predicate simplification rules
|
design cleanup
|
The backend should validate that predicate simplification rules (`\ceil(_) => ...`) have a predicate on the right-hand side. The related internal errors during execution may be removed.
|
1.0
|
Validate predicate simplification rules - The backend should validate that predicate simplification rules (`\ceil(_) => ...`) have a predicate on the right-hand side. The related internal errors during execution may be removed.
|
non_main
|
validate predicate simplification rules the backend should validate that predicate simplification rules ceil have a predicate on the right hand side the related internal errors during execution may be removed
| 0
|
4,999
| 25,722,559,400
|
IssuesEvent
|
2022-12-07 14:38:44
|
centerofci/mathesar
|
https://api.github.com/repos/centerofci/mathesar
|
opened
|
Data Explorer styling/frontend granular meta issue
|
work: frontend status: ready restricted: maintainers type: meta
|
This meta issue tracks pending items in Data Explorer prioritized by demo video readiness
1. **The following are essential for the demo video, since they will be visible during the demo**
- [ ] Show cell content on the Cell selection tab
- [ ] Automatically select inspector tabs based on column selection/cell selection
- [ ] Format content in cells based on display_options from underlying table
- [ ] Implement Hide transformation
- [ ] Show auto summarization modal when user clicks on a column with one-to-many link
- [ ] Style save exploration modal
1.1 *The following are essential for demo, but will take more effort & planning. We should try to de-prioritize some items here if we can safely exclude from the demo video*
- [ ] Format array cell content based on display_options
- [ ] Show count in the column selection left pane for added columns
- [ ] Show array icon in left pane for columns with one-to-many links
- Not sure if this is required, since we ask for explicit confirmation from user
- @ghislaineguerin agreed with not showing the array icon in left pane, but it was not finalized
2. **Following are part of styling & frontend issues, but doesn't necessarily block the demo video, based on the initial script. But they many be visible during the demo, so we need to decide if we want to prioritize any of them**
- [ ] Implement Sort transformation
- [ ] Show save status indicator when unsaved changes are present
- [ ] Show tooltips with help icon:
- [ ] Near base table name
- [ ] Near column count indicator in left pane
- [ ] Show error message when columns are missing in underlying table
- [ ] Show error message when filters are empty/invalid and user performs additional operations
- [ ] Show initial closeable help tutorial when user opens Data Explorer for the first time
|
True
|
Data Explorer styling/frontend granular meta issue - This meta issue tracks pending items in Data Explorer prioritized by demo video readiness
1. **The following are essential for the demo video, since they will be visible during the demo**
- [ ] Show cell content on the Cell selection tab
- [ ] Automatically select inspector tabs based on column selection/cell selection
- [ ] Format content in cells based on display_options from underlying table
- [ ] Implement Hide transformation
- [ ] Show auto summarization modal when user clicks on a column with one-to-many link
- [ ] Style save exploration modal
1.1 *The following are essential for demo, but will take more effort & planning. We should try to de-prioritize some items here if we can safely exclude from the demo video*
- [ ] Format array cell content based on display_options
- [ ] Show count in the column selection left pane for added columns
- [ ] Show array icon in left pane for columns with one-to-many links
- Not sure if this is required, since we ask for explicit confirmation from user
- @ghislaineguerin agreed with not showing the array icon in left pane, but it was not finalized
2. **Following are part of styling & frontend issues, but doesn't necessarily block the demo video, based on the initial script. But they many be visible during the demo, so we need to decide if we want to prioritize any of them**
- [ ] Implement Sort transformation
- [ ] Show save status indicator when unsaved changes are present
- [ ] Show tooltips with help icon:
- [ ] Near base table name
- [ ] Near column count indicator in left pane
- [ ] Show error message when columns are missing in underlying table
- [ ] Show error message when filters are empty/invalid and user performs additional operations
- [ ] Show initial closeable help tutorial when user opens Data Explorer for the first time
|
main
|
data explorer styling frontend granular meta issue this meta issue tracks pending items in data explorer prioritized by demo video readiness the following are essential for the demo video since they will be visible during the demo show cell content on the cell selection tab automatically select inspector tabs based on column selection cell selection format content in cells based on display options from underlying table implement hide transformation show auto summarization modal when user clicks on a column with one to many link style save exploration modal the following are essential for demo but will take more effort planning we should try to de prioritize some items here if we can safely exclude from the demo video format array cell content based on display options show count in the column selection left pane for added columns show array icon in left pane for columns with one to many links not sure if this is required since we ask for explicit confirmation from user ghislaineguerin agreed with not showing the array icon in left pane but it was not finalized following are part of styling frontend issues but doesn t necessarily block the demo video based on the initial script but they many be visible during the demo so we need to decide if we want to prioritize any of them implement sort transformation show save status indicator when unsaved changes are present show tooltips with help icon near base table name near column count indicator in left pane show error message when columns are missing in underlying table show error message when filters are empty invalid and user performs additional operations show initial closeable help tutorial when user opens data explorer for the first time
| 1
|
170,371
| 14,257,572,580
|
IssuesEvent
|
2020-11-20 04:01:51
|
microsoft/STL
|
https://api.github.com/repos/microsoft/STL
|
closed
|
README.md: The Block Diagram doesn't mention ConcRT
|
documentation resolved
|
https://github.com/microsoft/STL#block-diagram mentions VCStartup, VCRuntime, and the Universal CRT, but not ConcRT. As ConcRT's relationship with the STL is unusually circular, this should probably be mentioned.
|
1.0
|
README.md: The Block Diagram doesn't mention ConcRT - https://github.com/microsoft/STL#block-diagram mentions VCStartup, VCRuntime, and the Universal CRT, but not ConcRT. As ConcRT's relationship with the STL is unusually circular, this should probably be mentioned.
|
non_main
|
readme md the block diagram doesn t mention concrt mentions vcstartup vcruntime and the universal crt but not concrt as concrt s relationship with the stl is unusually circular this should probably be mentioned
| 0
|
232,615
| 17,788,777,328
|
IssuesEvent
|
2021-08-31 14:04:36
|
calliope-project/euro-calliope
|
https://api.github.com/repos/calliope-project/euro-calliope
|
closed
|
Explain background and purpose
|
documentation
|
Expand Sphinx documentation to describe the background and purpose of the repository
* Can build on Tim + Bryn's code sprint presentations
|
1.0
|
Explain background and purpose - Expand Sphinx documentation to describe the background and purpose of the repository
* Can build on Tim + Bryn's code sprint presentations
|
non_main
|
explain background and purpose expand sphinx documentation to describe the background and purpose of the repository can build on tim bryn s code sprint presentations
| 0
|
27,254
| 2,691,404,673
|
IssuesEvent
|
2015-03-31 21:21:51
|
NuGet/NuGetGallery
|
https://api.github.com/repos/NuGet/NuGetGallery
|
closed
|
Statistics navigation/drilldown suggestion
|
Priority - 2
|
I found the navigation/drilldown for the existing statistics pages didn't work the way I expected (after coming back to the packages with fresh expectations after a while).
**Problem:**
You cannot drill down into the by-version statistics for a package from the package's overall statistics.
**Repro:**
1. Navigate to https://www.nuget.org/stats
1. Click on the downloads number for Newtonsoft.Json in the left-hand column (Downloaded Packages)
1. Click on a Version number in the table
**Expected:**
To be taken to the statistics page for that specific version
**Actual:**
Taken to the package details page for that version
**Note:**
You can get to the version-specific statistics page for package versions that make the Downloaded Package Versions list, but there's no drill down that gets you there for a version that isn't in that report (even though the page can render).
|
1.0
|
Statistics navigation/drilldown suggestion - I found the navigation/drilldown for the existing statistics pages didn't work the way I expected (after coming back to the packages with fresh expectations after a while).
**Problem:**
You cannot drill down into the by-version statistics for a package from the package's overall statistics.
**Repro:**
1. Navigate to https://www.nuget.org/stats
1. Click on the downloads number for Newtonsoft.Json in the left-hand column (Downloaded Packages)
1. Click on a Version number in the table
**Expected:**
To be taken to the statistics page for that specific version
**Actual:**
Taken to the package details page for that version
**Note:**
You can get to the version-specific statistics page for package versions that make the Downloaded Package Versions list, but there's no drill down that gets you there for a version that isn't in that report (even though the page can render).
|
non_main
|
statistics navigation drilldown suggestion i found the navigation drilldown for the existing statistics pages didn t work the way i expected after coming back to the packages with fresh expectations after a while problem you cannot drill down into the by version statistics for a package from the package s overall statistics repro navigate to click on the downloads number for newtonsoft json in the left hand column downloaded packages click on a version number in the table expected to be taken to the statistics page for that specific version actual taken to the package details page for that version note you can get to the version specific statistics page for package versions that make the downloaded package versions list but there s no drill down that gets you there for a version that isn t in that report even though the page can render
| 0
|
865
| 4,534,587,162
|
IssuesEvent
|
2016-09-08 15:00:05
|
ansible/ansible-modules-core
|
https://api.github.com/repos/ansible/ansible-modules-core
|
closed
|
apache2_module fails for php7.0 on Ubuntu Xenial
|
bug_report waiting_on_maintainer
|
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
apache2_module
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0 (devel 982db58aff) last updated 2016/09/08 11:50:49 (GMT +100)
lib/ansible/modules/core: (detached HEAD db38f0c876) last updated 2016/09/08 13:03:40 (GMT +100)
lib/ansible/modules/extras: (detached HEAD 8bfdcfcab2) last updated 2016/09/08 11:51:00 (GMT +100)
config file = /home/rowan/.ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
N/A
##### SUMMARY
Ubuntu Xenial lists the php7.0 module as php7_module when running apache2ctl -M this breaks the regexp checking if the module is enabled.
I've made a work around here https://github.com/rwky/ansible-modules-core/commit/00ad6ef035a10dac7c84b7b68f04b00a739b104b but I didn't make a PR since I expect it may break other distros/versions.
Not entirely sure what the best solution to this is.
##### STEPS TO REPRODUCE
Run apache2_module with name=php7.0 state=present on a xenial server.
|
True
|
apache2_module fails for php7.0 on Ubuntu Xenial - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
apache2_module
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0 (devel 982db58aff) last updated 2016/09/08 11:50:49 (GMT +100)
lib/ansible/modules/core: (detached HEAD db38f0c876) last updated 2016/09/08 13:03:40 (GMT +100)
lib/ansible/modules/extras: (detached HEAD 8bfdcfcab2) last updated 2016/09/08 11:51:00 (GMT +100)
config file = /home/rowan/.ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
N/A
##### SUMMARY
Ubuntu Xenial lists the php7.0 module as php7_module when running apache2ctl -M this breaks the regexp checking if the module is enabled.
I've made a work around here https://github.com/rwky/ansible-modules-core/commit/00ad6ef035a10dac7c84b7b68f04b00a739b104b but I didn't make a PR since I expect it may break other distros/versions.
Not entirely sure what the best solution to this is.
##### STEPS TO REPRODUCE
Run apache2_module with name=php7.0 state=present on a xenial server.
|
main
|
module fails for on ubuntu xenial issue type bug report component name module ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file home rowan ansible cfg configured module search path default w o overrides configuration n a os environment n a summary ubuntu xenial lists the module as module when running m this breaks the regexp checking if the module is enabled i ve made a work around here but i didn t make a pr since i expect it may break other distros versions not entirely sure what the best solution to this is steps to reproduce run module with name state present on a xenial server
| 1
|
170,689
| 20,883,857,046
|
IssuesEvent
|
2022-03-23 01:20:49
|
turkdevops/sanity-nuxt-events
|
https://api.github.com/repos/turkdevops/sanity-nuxt-events
|
opened
|
CVE-2021-44906 (Medium) detected in minimist-0.0.8.tgz, minimist-1.2.0.tgz
|
security vulnerability
|
## CVE-2021-44906 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary>
<p>
<details><summary><b>minimist-0.0.8.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p>
<p>Path to dependency file: /web/package.json</p>
<p>Path to vulnerable library: /web/node_modules/minimist/package.json,/studio/node_modules/mkdirp/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- nuxt-2.14.12.tgz (Root Library)
- webpack-2.14.12.tgz
- cache-loader-4.1.0.tgz
- mkdirp-0.5.1.tgz
- :x: **minimist-0.0.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimist-1.2.0.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p>
<p>Path to dependency file: /web/package.json</p>
<p>Path to vulnerable library: /web/node_modules/loader-utils/node_modules/minimist/package.json,/studio/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- nuxt-2.14.12.tgz (Root Library)
- webpack-2.14.12.tgz
- cache-loader-4.1.0.tgz
- loader-utils-1.2.3.tgz
- json5-1.0.1.tgz
- :x: **minimist-1.2.0.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95).
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44906>CVE-2021-44906</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-44906">https://nvd.nist.gov/vuln/detail/CVE-2021-44906</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: BumperLane.Public.Service.Contracts - 0.23.35.214-prerelease;cloudscribe.templates - 5.2.0;Virteom.Tenant.Mobile.Bluetooth - 0.21.29.159-prerelease;ShowingVault.DotNet.Sdk - 0.13.41.190-prerelease;Envisia.DotNet.Templates - 3.0.1;Yarnpkg.Yarn - 0.26.1;Virteom.Tenant.Mobile.Framework.UWP - 0.20.41.103-prerelease;Virteom.Tenant.Mobile.Framework.iOS - 0.20.41.103-prerelease;BumperLane.Public.Api.V2.ClientModule - 0.23.35.214-prerelease;VueJS.NetCore - 1.1.1;Dianoga - 4.0.0,3.0.0-RC02;Virteom.Tenant.Mobile.Bluetooth.iOS - 0.20.41.103-prerelease;Virteom.Public.Utilities - 0.23.37.212-prerelease;Indianadavy.VueJsWebAPITemplate.CSharp - 1.0.1;NorDroN.AngularTemplate - 0.1.6;Virteom.Tenant.Mobile.Framework - 0.21.29.159-prerelease;Virteom.Tenant.Mobile.Bluetooth.Android - 0.20.41.103-prerelease;z4a-dotnet-scaffold - 1.0.0.2;Raml.Parser - 1.0.7;CoreVueWebTest - 3.0.101;dotnetng.template - 1.0.0.4;SitecoreMaster.TrueDynamicPlaceholders - 1.0.3;Virteom.Tenant.Mobile.Framework.Android - 0.20.41.103-prerelease;Fable.Template.Elmish.React - 0.1.6;BlazorPolyfill.Build - 6.0.100.2;Fable.Snowpack.Template - 2.1.0;BumperLane.Public.Api.Client - 0.23.35.214-prerelease;Yarn.MSBuild - 0.22.0,0.24.6;Blazor.TailwindCSS.BUnit - 1.0.2;Bridge.AWS - 0.3.30.36;tslint - 5.6.0;SAFE.Template - 3.0.1;GR.PageRender.Razor - 1.8.0;MIDIator.WebClient - 1.0.105</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-44906 (Medium) detected in minimist-0.0.8.tgz, minimist-1.2.0.tgz - ## CVE-2021-44906 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary>
<p>
<details><summary><b>minimist-0.0.8.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p>
<p>Path to dependency file: /web/package.json</p>
<p>Path to vulnerable library: /web/node_modules/minimist/package.json,/studio/node_modules/mkdirp/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- nuxt-2.14.12.tgz (Root Library)
- webpack-2.14.12.tgz
- cache-loader-4.1.0.tgz
- mkdirp-0.5.1.tgz
- :x: **minimist-0.0.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>minimist-1.2.0.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p>
<p>Path to dependency file: /web/package.json</p>
<p>Path to vulnerable library: /web/node_modules/loader-utils/node_modules/minimist/package.json,/studio/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- nuxt-2.14.12.tgz (Root Library)
- webpack-2.14.12.tgz
- cache-loader-4.1.0.tgz
- loader-utils-1.2.3.tgz
- json5-1.0.1.tgz
- :x: **minimist-1.2.0.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95).
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44906>CVE-2021-44906</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-44906">https://nvd.nist.gov/vuln/detail/CVE-2021-44906</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: BumperLane.Public.Service.Contracts - 0.23.35.214-prerelease;cloudscribe.templates - 5.2.0;Virteom.Tenant.Mobile.Bluetooth - 0.21.29.159-prerelease;ShowingVault.DotNet.Sdk - 0.13.41.190-prerelease;Envisia.DotNet.Templates - 3.0.1;Yarnpkg.Yarn - 0.26.1;Virteom.Tenant.Mobile.Framework.UWP - 0.20.41.103-prerelease;Virteom.Tenant.Mobile.Framework.iOS - 0.20.41.103-prerelease;BumperLane.Public.Api.V2.ClientModule - 0.23.35.214-prerelease;VueJS.NetCore - 1.1.1;Dianoga - 4.0.0,3.0.0-RC02;Virteom.Tenant.Mobile.Bluetooth.iOS - 0.20.41.103-prerelease;Virteom.Public.Utilities - 0.23.37.212-prerelease;Indianadavy.VueJsWebAPITemplate.CSharp - 1.0.1;NorDroN.AngularTemplate - 0.1.6;Virteom.Tenant.Mobile.Framework - 0.21.29.159-prerelease;Virteom.Tenant.Mobile.Bluetooth.Android - 0.20.41.103-prerelease;z4a-dotnet-scaffold - 1.0.0.2;Raml.Parser - 1.0.7;CoreVueWebTest - 3.0.101;dotnetng.template - 1.0.0.4;SitecoreMaster.TrueDynamicPlaceholders - 1.0.3;Virteom.Tenant.Mobile.Framework.Android - 0.20.41.103-prerelease;Fable.Template.Elmish.React - 0.1.6;BlazorPolyfill.Build - 6.0.100.2;Fable.Snowpack.Template - 2.1.0;BumperLane.Public.Api.Client - 0.23.35.214-prerelease;Yarn.MSBuild - 0.22.0,0.24.6;Blazor.TailwindCSS.BUnit - 1.0.2;Bridge.AWS - 0.3.30.36;tslint - 5.6.0;SAFE.Template - 3.0.1;GR.PageRender.Razor - 1.8.0;MIDIator.WebClient - 1.0.105</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_main
|
cve medium detected in minimist tgz minimist tgz cve medium severity vulnerability vulnerable libraries minimist tgz minimist tgz minimist tgz parse argument options library home page a href path to dependency file web package json path to vulnerable library web node modules minimist package json studio node modules mkdirp node modules minimist package json dependency hierarchy nuxt tgz root library webpack tgz cache loader tgz mkdirp tgz x minimist tgz vulnerable library minimist tgz parse argument options library home page a href path to dependency file web package json path to vulnerable library web node modules loader utils node modules minimist package json studio node modules minimist package json dependency hierarchy nuxt tgz root library webpack tgz cache loader tgz loader utils tgz tgz x minimist tgz vulnerable library found in base branch main vulnerability details minimist is vulnerable to prototype pollution via file index js function setkey lines publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bumperlane public service contracts prerelease cloudscribe templates virteom tenant mobile bluetooth prerelease showingvault dotnet sdk prerelease envisia dotnet templates yarnpkg yarn virteom tenant mobile framework uwp prerelease virteom tenant mobile framework ios prerelease bumperlane public api clientmodule prerelease vuejs netcore dianoga virteom tenant mobile bluetooth ios prerelease virteom public utilities prerelease indianadavy vuejswebapitemplate csharp nordron angulartemplate virteom tenant mobile framework prerelease virteom tenant mobile bluetooth android prerelease dotnet scaffold raml parser corevuewebtest dotnetng template sitecoremaster truedynamicplaceholders virteom tenant mobile framework android prerelease fable template elmish react blazorpolyfill build fable snowpack template bumperlane public api client prerelease yarn msbuild blazor tailwindcss bunit bridge aws tslint safe template gr pagerender razor midiator webclient step up your open source security game with whitesource
| 0
|
2,933
| 3,255,175,199
|
IssuesEvent
|
2015-10-20 06:59:37
|
piwik/piwik
|
https://api.github.com/repos/piwik/piwik
|
opened
|
Remove word "website" from website selector
|
c: Usability
|
As mentioned here https://github.com/piwik/piwik/issues/8712 I do now think as well that we should remove the word "Website" from the sites selector eg as suggested here: https://github.com/piwik/piwik/issues/8712#issuecomment-148390762
All other selectors use a proper icon instead but the selector doesn't yet.

Instead of the word "website" I'd rather see an icon, maybe one similar to the one used in Piwik Mobile eg http://fortawesome.github.io/Font-Awesome/icon/globe/ but could be also a different one since the globe could be also used for language etc.
Because of the word website one cannot see much of the website's name which can be especially tricky if one has many websites with similar names, maybe even many websites that start with the same name.
|
True
|
Remove word "website" from website selector - As mentioned here https://github.com/piwik/piwik/issues/8712 I do now think as well that we should remove the word "Website" from the sites selector eg as suggested here: https://github.com/piwik/piwik/issues/8712#issuecomment-148390762
All other selectors use a proper icon instead but the selector doesn't yet.

Instead of the word "website" I'd rather see an icon, maybe one similar to the one used in Piwik Mobile eg http://fortawesome.github.io/Font-Awesome/icon/globe/ but could be also a different one since the globe could be also used for language etc.
Because of the word website one cannot see much of the website's name which can be especially tricky if one has many websites with similar names, maybe even many websites that start with the same name.
|
non_main
|
remove word website from website selector as mentioned here i do now think as well that we should remove the word website from the sites selector eg as suggested here all other selectors use a proper icon instead but the selector doesn t yet instead of the word website i d rather see an icon maybe one similar to the one used in piwik mobile eg but could be also a different one since the globe could be also used for language etc because of the word website one cannot see much of the website s name which can be especially tricky if one has many websites with similar names maybe even many websites that start with the same name
| 0
|
1,501
| 2,514,821,676
|
IssuesEvent
|
2015-01-15 14:43:53
|
eclipsesource/tabris-js
|
https://api.github.com/repos/eclipsesource/tabris-js
|
closed
|
Support search actions
|
feature priority: high
|
Support actions that integrate a search field with proposals in the navigation bar when executed.
|
1.0
|
Support search actions - Support actions that integrate a search field with proposals in the navigation bar when executed.
|
non_main
|
support search actions support actions that integrate a search field with proposals in the navigation bar when executed
| 0
|
836
| 4,473,748,310
|
IssuesEvent
|
2016-08-26 06:17:59
|
Particular/NServiceBus.SqlServer
|
https://api.github.com/repos/Particular/NServiceBus.SqlServer
|
closed
|
Setup mulit-catalog deployment with Linked Servers with Sql Server
|
Tag: Maintainer Prio Type: Spike
|
## Goal
Validate if that is a valid option to provide multi instance deployments without need for sharing connection strings at the endpoint configuration level
## Results
### Proof-of-concept
The proof of concept consisted of a simple application that was using direct ADO.NET to send messages on my local machine between two linked SQL Server instances (2014 Express and 2014 Developer). These tests were single-threaded. Results were:
* The single-threaded send throughput is roughly 2-3 times lower when sending to the linked server
* When using local SQL transaction in the client and, inside it, sending the linked instance, the transaction is escalated to a distributed one (fails if DTC is disabled). Not sure about the actual mechanism of that escalation (because from client code it is a normal `SqlTransaction` object) but throwing an exception in the middle causes whole thing to roll back (in case there are multiple send operations)
* When sending to a linked server without using `SqlTransaction` does not cause the DTC escalation. This is probably because just creating the `SqlTransaction` object adds the local instance as a participant in the transaction.
### Modified SQL transport
After the initial promising results SQL Server transport has been modified to accepts 4-part addresses in form of `table@[instance].[catalog].[schema]`.
* It has been proved that, on a single instance, multi-catalog can be handled without the DTC escalation not the outbox. It works out-of-the-box
* A bug was found in the isolated send implementation (requiring new transaction instead of suppressing an existing one). This bug caused linked server sends to escalate to DTC immediately in outbox mode. **PR pending**
* A simple throughput test has been conducted where 20000 messages were queued up in the queue. The processor was running with concurrency limited to 20 and a handler that would send a message to a linked server for each incoming message. Simple measurements showed throughput of roughly **1000 msg/s** on **c4.2xlarge** AWS instance with EBS volume at **3000 IOPS**
#### Results
* Multi-instance is easy to implement on top of V3 SQL Server transport by extending the `QueueAddress` class.
* Linked server sends are slower than local sends but the additional cost does not seem to be prohibitive. If we decide to go with this approach we should stress that **always** the most performant approach is to have all the queues in a single instance.
* Linked server security concerns has not been investigated. It would probably require setting up a domain controller.
* There is a problem with `ReplyTo` address handling. Generating proper `ReplyTo` address requires, for the endpoint instance to know the catalog and instance it connects to. the catalog can be obtained from the connection string but the instance is more problematic because the connection string would accept *aliases* such as `.` or `(local)` or a host name (while other machines might need to use the IP).
#### Possible workarounds
* Use views and `instead of`triggers to send to linked servers instead of doing it directly. This is not an option because it requires DTC between the linked servers (probably because the trigger has access to the local `inserted` table)
* Explicitly enable "linked servers". In that mode the `Data Source` property in the connection string has to be the network-accessible name of the instance -- same one as is used to created links to this server from other servers. In this mode the instance name is going to be part of the transport address
* Allow to override the SQL command used to send a message providing i.e. calling a stored procedure instead of directly inserting. The SP can send to the linked server. Here we have the advantage that the naming scheme of linked servers can be administered centrally.
### Three-part addressing
Because of the problem outlined above, another experiment was about using three-part (instead of four-part) addressing where instance is not part of the address. This approach requires that catalog names are globally unique but offers some big advantages over four-part addressing. The biggest one is ability to overcome the reply address problem in an elegant way:
* The instance, on startup, reads the list of servers linked to its instance
* For each linked server it reads the list of catalogs that it has access to on that linked server
* Above information is used to determine, for each destination catalog, weather it is on local instance or a linked instance.
The only problem with the above approach is that in the outbox mode it unnecessarily escalates to DTC. Implementing [this small fix](https://github.com/Particular/NServiceBus/issues/3882) should prevent this.
|
True
|
Setup mulit-catalog deployment with Linked Servers with Sql Server - ## Goal
Validate if that is a valid option to provide multi instance deployments without need for sharing connection strings at the endpoint configuration level
## Results
### Proof-of-concept
The proof of concept consisted of a simple application that was using direct ADO.NET to send messages on my local machine between two linked SQL Server instances (2014 Express and 2014 Developer). These tests were single-threaded. Results were:
* The single-threaded send throughput is roughly 2-3 times lower when sending to the linked server
* When using local SQL transaction in the client and, inside it, sending the linked instance, the transaction is escalated to a distributed one (fails if DTC is disabled). Not sure about the actual mechanism of that escalation (because from client code it is a normal `SqlTransaction` object) but throwing an exception in the middle causes whole thing to roll back (in case there are multiple send operations)
* When sending to a linked server without using `SqlTransaction` does not cause the DTC escalation. This is probably because just creating the `SqlTransaction` object adds the local instance as a participant in the transaction.
### Modified SQL transport
After the initial promising results SQL Server transport has been modified to accepts 4-part addresses in form of `table@[instance].[catalog].[schema]`.
* It has been proved that, on a single instance, multi-catalog can be handled without the DTC escalation not the outbox. It works out-of-the-box
* A bug was found in the isolated send implementation (requiring new transaction instead of suppressing an existing one). This bug caused linked server sends to escalate to DTC immediately in outbox mode. **PR pending**
* A simple throughput test has been conducted where 20000 messages were queued up in the queue. The processor was running with concurrency limited to 20 and a handler that would send a message to a linked server for each incoming message. Simple measurements showed throughput of roughly **1000 msg/s** on **c4.2xlarge** AWS instance with EBS volume at **3000 IOPS**
#### Results
* Multi-instance is easy to implement on top of V3 SQL Server transport by extending the `QueueAddress` class.
* Linked server sends are slower than local sends but the additional cost does not seem to be prohibitive. If we decide to go with this approach we should stress that **always** the most performant approach is to have all the queues in a single instance.
* Linked server security concerns has not been investigated. It would probably require setting up a domain controller.
* There is a problem with `ReplyTo` address handling. Generating proper `ReplyTo` address requires, for the endpoint instance to know the catalog and instance it connects to. the catalog can be obtained from the connection string but the instance is more problematic because the connection string would accept *aliases* such as `.` or `(local)` or a host name (while other machines might need to use the IP).
#### Possible workarounds
* Use views and `instead of`triggers to send to linked servers instead of doing it directly. This is not an option because it requires DTC between the linked servers (probably because the trigger has access to the local `inserted` table)
* Explicitly enable "linked servers". In that mode the `Data Source` property in the connection string has to be the network-accessible name of the instance -- same one as is used to created links to this server from other servers. In this mode the instance name is going to be part of the transport address
* Allow to override the SQL command used to send a message providing i.e. calling a stored procedure instead of directly inserting. The SP can send to the linked server. Here we have the advantage that the naming scheme of linked servers can be administered centrally.
### Three-part addressing
Because of the problem outlined above, another experiment was about using three-part (instead of four-part) addressing where instance is not part of the address. This approach requires that catalog names are globally unique but offers some big advantages over four-part addressing. The biggest one is ability to overcome the reply address problem in an elegant way:
* The instance, on startup, reads the list of servers linked to its instance
* For each linked server it reads the list of catalogs that it has access to on that linked server
* Above information is used to determine, for each destination catalog, weather it is on local instance or a linked instance.
The only problem with the above approach is that in the outbox mode it unnecessarily escalates to DTC. Implementing [this small fix](https://github.com/Particular/NServiceBus/issues/3882) should prevent this.
|
main
|
setup mulit catalog deployment with linked servers with sql server goal validate if that is a valid option to provide multi instance deployments without need for sharing connection strings at the endpoint configuration level results proof of concept the proof of concept consisted of a simple application that was using direct ado net to send messages on my local machine between two linked sql server instances express and developer these tests were single threaded results were the single threaded send throughput is roughly times lower when sending to the linked server when using local sql transaction in the client and inside it sending the linked instance the transaction is escalated to a distributed one fails if dtc is disabled not sure about the actual mechanism of that escalation because from client code it is a normal sqltransaction object but throwing an exception in the middle causes whole thing to roll back in case there are multiple send operations when sending to a linked server without using sqltransaction does not cause the dtc escalation this is probably because just creating the sqltransaction object adds the local instance as a participant in the transaction modified sql transport after the initial promising results sql server transport has been modified to accepts part addresses in form of table it has been proved that on a single instance multi catalog can be handled without the dtc escalation not the outbox it works out of the box a bug was found in the isolated send implementation requiring new transaction instead of suppressing an existing one this bug caused linked server sends to escalate to dtc immediately in outbox mode pr pending a simple throughput test has been conducted where messages were queued up in the queue the processor was running with concurrency limited to and a handler that would send a message to a linked server for each incoming message simple measurements showed throughput of roughly msg s on aws instance with ebs volume at iops results multi instance is easy to implement on top of sql server transport by extending the queueaddress class linked server sends are slower than local sends but the additional cost does not seem to be prohibitive if we decide to go with this approach we should stress that always the most performant approach is to have all the queues in a single instance linked server security concerns has not been investigated it would probably require setting up a domain controller there is a problem with replyto address handling generating proper replyto address requires for the endpoint instance to know the catalog and instance it connects to the catalog can be obtained from the connection string but the instance is more problematic because the connection string would accept aliases such as or local or a host name while other machines might need to use the ip possible workarounds use views and instead of triggers to send to linked servers instead of doing it directly this is not an option because it requires dtc between the linked servers probably because the trigger has access to the local inserted table explicitly enable linked servers in that mode the data source property in the connection string has to be the network accessible name of the instance same one as is used to created links to this server from other servers in this mode the instance name is going to be part of the transport address allow to override the sql command used to send a message providing i e calling a stored procedure instead of directly inserting the sp can send to the linked server here we have the advantage that the naming scheme of linked servers can be administered centrally three part addressing because of the problem outlined above another experiment was about using three part instead of four part addressing where instance is not part of the address this approach requires that catalog names are globally unique but offers some big advantages over four part addressing the biggest one is ability to overcome the reply address problem in an elegant way the instance on startup reads the list of servers linked to its instance for each linked server it reads the list of catalogs that it has access to on that linked server above information is used to determine for each destination catalog weather it is on local instance or a linked instance the only problem with the above approach is that in the outbox mode it unnecessarily escalates to dtc implementing should prevent this
| 1
|
262,138
| 27,857,406,372
|
IssuesEvent
|
2023-03-21 00:56:08
|
aws/eks-distro-build-tooling
|
https://api.github.com/repos/aws/eks-distro-build-tooling
|
closed
|
Vulnerability in golang.org/x/text/language - CVE-2022-32149
|
security golang
|
From [Golang Security Announcement](https://groups.google.com/g/golang-announce/c/-hjNw559_tE/m/KlGTfid5CAAJ):
Version v0.3.8 of [golang.org/x/text](http://golang.org/x/text) fixes a vulnerability in the [golang.org/x/text/language](http://golang.org/x/text/language) package which could cause a denial of service.
An attacker can craft an Accept-Language header which ParseAcceptLanguage will take significant time to parse.
This issue was discovered by OSS-Fuzz and reported to us by Adam Korczynski (ADA Logics), and is tracked as CVE-2022-32149 and https://go.dev/issue/56152.
|
True
|
Vulnerability in golang.org/x/text/language - CVE-2022-32149 - From [Golang Security Announcement](https://groups.google.com/g/golang-announce/c/-hjNw559_tE/m/KlGTfid5CAAJ):
Version v0.3.8 of [golang.org/x/text](http://golang.org/x/text) fixes a vulnerability in the [golang.org/x/text/language](http://golang.org/x/text/language) package which could cause a denial of service.
An attacker can craft an Accept-Language header which ParseAcceptLanguage will take significant time to parse.
This issue was discovered by OSS-Fuzz and reported to us by Adam Korczynski (ADA Logics), and is tracked as CVE-2022-32149 and https://go.dev/issue/56152.
|
non_main
|
vulnerability in golang org x text language cve from version of fixes a vulnerability in the package which could cause a denial of service an attacker can craft an accept language header which parseacceptlanguage will take significant time to parse this issue was discovered by oss fuzz and reported to us by adam korczynski ada logics and is tracked as cve and
| 0
|
51,283
| 13,635,089,318
|
IssuesEvent
|
2020-09-25 01:51:38
|
nasifimtiazohi/openmrs-module-event-2.7.0
|
https://api.github.com/repos/nasifimtiazohi/openmrs-module-event-2.7.0
|
opened
|
CVE-2015-6524 (Medium) detected in activemq-core-5.4.3.jar
|
security vulnerability
|
## CVE-2015-6524 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>activemq-core-5.4.3.jar</b></p></summary>
<p>The ActiveMQ Message Broker and Client implementations</p>
<p>Path to dependency file: openmrs-module-event-2.7.0/omod/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/activemq/activemq-core/5.4.3/activemq-core-5.4.3.jar,openmrs-module-event-2.7.0/omod/target/event-2.7.0/lib/activemq-core-5.4.3.jar,canner/.m2/repository/org/apache/activemq/activemq-core/5.4.3/activemq-core-5.4.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **activemq-core-5.4.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-module-event-2.7.0/commit/fea757b7944c14045b34164a543090874667bcc4">fea757b7944c14045b34164a543090874667bcc4</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The LDAPLoginModule implementation in the Java Authentication and Authorization Service (JAAS) in Apache ActiveMQ 5.x before 5.10.1 allows wildcard operators in usernames, which allows remote attackers to obtain credentials via a brute force attack. NOTE: this identifier was SPLIT from CVE-2014-3612 per ADT2 due to different vulnerability types.
<p>Publish Date: 2015-08-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-6524>CVE-2015-6524</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-6524">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-6524</a></p>
<p>Release Date: 2015-08-24</p>
<p>Fix Resolution: org.apache.activemq:activemq-osgi:5.10.1,org.apache.activemq:activemq-all:5.10.1,org.apache.activemq:activemq-web:5.10.1,org.apache.activemq:activemq-jaas:5.10.1,org.apache.activemq:activemq-broker:5.10.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2015-6524 (Medium) detected in activemq-core-5.4.3.jar - ## CVE-2015-6524 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>activemq-core-5.4.3.jar</b></p></summary>
<p>The ActiveMQ Message Broker and Client implementations</p>
<p>Path to dependency file: openmrs-module-event-2.7.0/omod/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/activemq/activemq-core/5.4.3/activemq-core-5.4.3.jar,openmrs-module-event-2.7.0/omod/target/event-2.7.0/lib/activemq-core-5.4.3.jar,canner/.m2/repository/org/apache/activemq/activemq-core/5.4.3/activemq-core-5.4.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **activemq-core-5.4.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-module-event-2.7.0/commit/fea757b7944c14045b34164a543090874667bcc4">fea757b7944c14045b34164a543090874667bcc4</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The LDAPLoginModule implementation in the Java Authentication and Authorization Service (JAAS) in Apache ActiveMQ 5.x before 5.10.1 allows wildcard operators in usernames, which allows remote attackers to obtain credentials via a brute force attack. NOTE: this identifier was SPLIT from CVE-2014-3612 per ADT2 due to different vulnerability types.
<p>Publish Date: 2015-08-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-6524>CVE-2015-6524</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-6524">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-6524</a></p>
<p>Release Date: 2015-08-24</p>
<p>Fix Resolution: org.apache.activemq:activemq-osgi:5.10.1,org.apache.activemq:activemq-all:5.10.1,org.apache.activemq:activemq-web:5.10.1,org.apache.activemq:activemq-jaas:5.10.1,org.apache.activemq:activemq-broker:5.10.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_main
|
cve medium detected in activemq core jar cve medium severity vulnerability vulnerable library activemq core jar the activemq message broker and client implementations path to dependency file openmrs module event omod pom xml path to vulnerable library home wss scanner repository org apache activemq activemq core activemq core jar openmrs module event omod target event lib activemq core jar canner repository org apache activemq activemq core activemq core jar dependency hierarchy x activemq core jar vulnerable library found in head commit a href found in base branch master vulnerability details the ldaploginmodule implementation in the java authentication and authorization service jaas in apache activemq x before allows wildcard operators in usernames which allows remote attackers to obtain credentials via a brute force attack note this identifier was split from cve per due to different vulnerability types publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution org apache activemq activemq osgi org apache activemq activemq all org apache activemq activemq web org apache activemq activemq jaas org apache activemq activemq broker step up your open source security game with whitesource
| 0
|
226,083
| 17,947,865,460
|
IssuesEvent
|
2021-09-12 06:18:10
|
MetagaussInc/Blazeforms-Revamped-Frontend
|
https://api.github.com/repos/MetagaussInc/Blazeforms-Revamped-Frontend
|
closed
|
When control is resized to the biggest size then Make Bigger or Make smaller not appearing faded out.[Form-build]
|
bug high Ready For Retest
|
1. When control is resized to the biggest possible size either via contextual menu or via left side panel options, the contextual option for "Make Bigger" not appearing to fade out.
2. When control is resized to the smallest possible size either via contextual menu or via left side panel options, the contextual option for "Make Smaller" not appearing to fade out.
|
1.0
|
When control is resized to the biggest size then Make Bigger or Make smaller not appearing faded out.[Form-build] - 1. When control is resized to the biggest possible size either via contextual menu or via left side panel options, the contextual option for "Make Bigger" not appearing to fade out.
2. When control is resized to the smallest possible size either via contextual menu or via left side panel options, the contextual option for "Make Smaller" not appearing to fade out.
|
non_main
|
when control is resized to the biggest size then make bigger or make smaller not appearing faded out when control is resized to the biggest possible size either via contextual menu or via left side panel options the contextual option for make bigger not appearing to fade out when control is resized to the smallest possible size either via contextual menu or via left side panel options the contextual option for make smaller not appearing to fade out
| 0
|
18,055
| 3,664,439,003
|
IssuesEvent
|
2016-02-19 11:41:52
|
handsontable/handsontable
|
https://api.github.com/repos/handsontable/handsontable
|
closed
|
Getting 2 rows of data when pasting to one row in Chrome.
|
Answered Merged (ready for release) Tested
|
Hi There
There's an issue in Handsontable when it's used inside an Angular2 application.
When you copy a row and paste it into a different row on Chrome, 2 rows are inserted instead of one.

It only happens when Handsontable is running inside Angular2. I think this is somehow related to ZoneJS since this issue seems also be happening in MeteorJS when the ZoneJS Bundle is used there (https://github.com/awsp/handsontable-meteor/issues/17).
You can reproduce it here: https://valor-software.github.io/ng2-handsontable/
**Steps to reproduce:**
- Open example page in Chrome (47.0.2526.111)
- Click on a single row
- CMD+C (on Mac)
- Click on a different row (has to be writable)
- CMD+V
Two rows are inserted.
Any idea why ZoneJS might cause this issue on Chrome?
Best regards
Pascal Helfenstein
|
1.0
|
Getting 2 rows of data when pasting to one row in Chrome. - Hi There
There's an issue in Handsontable when it's used inside an Angular2 application.
When you copy a row and paste it into a different row on Chrome, 2 rows are inserted instead of one.

It only happens when Handsontable is running inside Angular2. I think this is somehow related to ZoneJS since this issue seems also be happening in MeteorJS when the ZoneJS Bundle is used there (https://github.com/awsp/handsontable-meteor/issues/17).
You can reproduce it here: https://valor-software.github.io/ng2-handsontable/
**Steps to reproduce:**
- Open example page in Chrome (47.0.2526.111)
- Click on a single row
- CMD+C (on Mac)
- Click on a different row (has to be writable)
- CMD+V
Two rows are inserted.
Any idea why ZoneJS might cause this issue on Chrome?
Best regards
Pascal Helfenstein
|
non_main
|
getting rows of data when pasting to one row in chrome hi there there s an issue in handsontable when it s used inside an application when you copy a row and paste it into a different row on chrome rows are inserted instead of one it only happens when handsontable is running inside i think this is somehow related to zonejs since this issue seems also be happening in meteorjs when the zonejs bundle is used there you can reproduce it here steps to reproduce open example page in chrome click on a single row cmd c on mac click on a different row has to be writable cmd v two rows are inserted any idea why zonejs might cause this issue on chrome best regards pascal helfenstein
| 0
|
4,027
| 18,797,941,499
|
IssuesEvent
|
2021-11-09 01:43:35
|
tgstation/tgstation
|
https://api.github.com/repos/tgstation/tgstation
|
closed
|
Admin logs on a ckey read from said mob's specific status, rather than all logs.
|
Maintainability/Hinders improvements Bug Administration Cleanup Flagged
|
This means that when looking at a ghost's logs through the player panel, it only shows ghost death logs, if you examine bodies it shows body combat logs, etc.
|
True
|
Admin logs on a ckey read from said mob's specific status, rather than all logs. - This means that when looking at a ghost's logs through the player panel, it only shows ghost death logs, if you examine bodies it shows body combat logs, etc.
|
main
|
admin logs on a ckey read from said mob s specific status rather than all logs this means that when looking at a ghost s logs through the player panel it only shows ghost death logs if you examine bodies it shows body combat logs etc
| 1
|
692
| 4,238,514,479
|
IssuesEvent
|
2016-07-06 04:18:04
|
duckduckgo/zeroclickinfo-spice
|
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
|
closed
|
Goodreads: IA not showing up
|
Maintainer Input Requested
|
When I performed a search the instant answer did not show up. I was using Microsoft Edge with DDG Beta. URL - https://beta.duckduckgo.com/?q=books+by+melanie+watt.

------
IA Page: http://duck.co/ia/view/goodreads
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @iammrigank
|
True
|
Goodreads: IA not showing up - When I performed a search the instant answer did not show up. I was using Microsoft Edge with DDG Beta. URL - https://beta.duckduckgo.com/?q=books+by+melanie+watt.

------
IA Page: http://duck.co/ia/view/goodreads
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @iammrigank
|
main
|
goodreads ia not showing up when i performed a search the instant answer did not show up i was using microsoft edge with ddg beta url ia page iammrigank
| 1
|
1,754
| 6,574,970,743
|
IssuesEvent
|
2017-09-11 14:39:06
|
ansible/ansible-modules-core
|
https://api.github.com/repos/ansible/ansible-modules-core
|
closed
|
ios_config changes some items which do not need to be changed
|
affects_2.3 bug_report networking waiting_on_maintainer
|
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ios_config
##### ANSIBLE VERSION
```
ansible 2.3.0 commit d9b570a
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
core modules commit ec4eebc
extras modules commit edbddf6
```
##### CONFIGURATION
inventory = ./hosts
gathering = explicit
roles_path = /home/actionmystique/Program-Files/Ubuntu/Ansible/git-Ansible/roles
private_role_vars = yes
log_path = /var/log/ansible.log
fact_caching = redis
fact_caching_timeout = 86400
retry_files_enabled = False
##### OS / ENVIRONMENT
- host: Ubuntu 16.04 4.4.0
- target: **CSR-1000v 16.2.2**
##### SUMMARY
There is a strange behavior when some commands are correctly applied multiple times in a row: the previous occurrences are not seen as changed.
For instance, the following tasks are always applied even though they have already been successfully run:
```
- name: Configuring location, contact & chassis ID information in new SNMPv{{ snmp.new_version }}
ios_config:
provider: "{{ connections.ssh }}"
lines:
- "snmp-server location {{ snmp.location }}"
- "snmp-server contact {{ snmp.contact }}"
- "snmp-server chassis-id {{ snmp.chassis_id }}"
- "snmp-server ifindex persist"
when: snmp.admin_state == "up"
register: result
```
```
TASK [ios_snmp : Configuring location, contact & chassis ID information in new SNMPv2c] ***
changed: [XEv_Spine_31] => {"changed": true, "updates": ["snmp-server ifindex persist"], "warnings": []}
changed: [XEv_Spine_32] => {"changed": true, "updates": ["snmp-server ifindex persist"], "warnings": []}
```
```
- name: Configuring RO/RW community string in new IPv4/SNMPv{{ snmp.new_version }}
ios_config:
provider: "{{ connections.ssh }}"
lines:
- "snmp-server community {{ snmp.ssh.community }} {{ snmp.community_type }} {{ snmp.acl_name }}"
when: snmp.ip_version == 'ip'
register: result
```
```
TASK [ios_snmp : Configuring RO/RW community string in new IPv4/SNMPv2c] *******
changed: [XEv_Spine_32] => {"changed": true, "updates": ["snmp-server community xxxxxxxx ro authorized-snmp-nms"], "warnings": []}
changed: [XEv_Spine_31] => {"changed": true, "updates": ["snmp-server community xxxxxxxx ro authorized-snmp-nms"], "warnings": []}
```
```
- name: Configuring which NMS is allowed to receive the traps in new SNMPv{{ snmp.new_version }}
ios_config:
provider: "{{ connections.ssh }}"
lines:
- "snmp-server host {{ nms_mgt_ip_address }} {{ snmp.notifications_type }} version {{ snmp.new_version }} {{ snmp.ssh.community }} udp-port {{ snmp.trap_port }}"
when: (snmp.notifications == 'all') or (snmp.notifications == 'list')
register: result
```
```
TASK [ios_snmp : Configuring which NMS is allowed to receive the traps in new SNMPv2c] ***
changed: [XEv_Spine_31] => {"changed": true, "updates": ["snmp-server host 172.21.100.1 traps version 2c xxxxxxxx udp-port 162"], "warnings": []}
changed: [XEv_Spine_32] => {"changed": true, "updates": ["snmp-server host 172.21.100.1 traps version 2c xxxxxxxx udp-port 162"], "warnings": []}
```
Not all commands are concerned with this glitch. For instance, the second passage does not lead to a change for:
```
- name: Configuring the ACL for authorized NMS in new SNMPv{{ snmp.new_version }}
ios_config:
provider: "{{ connections.ssh }}"
parents:
- "{{ snmp.ip_version }} access-list extended {{ snmp.acl_name }}"
lines:
- "permit {{ snmp.ip_version }} host {{ nms_mgt_ip_address }} host {{ ansible_host }} log"
when: snmp.admin_state == "up"
register: result
```
```
TASK [ios_snmp : Configuring the ACL for authorized NMS in new SNMPv2c] ********
ok: [XEv_Spine_32] => {"changed": false, "warnings": []}
ok: [XEv_Spine_31] => {"changed": false, "warnings": []}
```
```
- name: Configuring the list of notifications in new SNMPv{{ snmp.new_version }}
ios_config:
provider: "{{ connections.ssh }}"
lines:
- "snmp-server enable traps {{ item }}"
with_items: "{{ snmp.notifications_list.traps }}"
when: (snmp.notifications == 'list')
register: result
```
The previous task works as expected for all the notifications I have tested, except for a few:
- cpu
- memory
- ospf
- snmp
```
TASK [ios_snmp : Configuring the list of notifications in new SNMPv2c] *********
ok: [XEv_Spine_31] => (item=bfd) => {"changed": false, "item": "bfd", "warnings": []}
ok: [XEv_Spine_32] => (item=bfd) => {"changed": false, "item": "bfd", "warnings": []}
ok: [XEv_Spine_31] => (item=bgp) => {"changed": false, "item": "bgp", "warnings": []}
ok: [XEv_Spine_32] => (item=bgp) => {"changed": false, "item": "bgp", "warnings": []}
changed: [XEv_Spine_31] => (item=cpu) => {"changed": true, "item": "cpu", "updates": ["snmp-server enable traps cpu"], "warnings": []}
changed: [XEv_Spine_32] => (item=cpu) => {"changed": true, "item": "cpu", "updates": ["snmp-server enable traps cpu"], "warnings": []}
ok: [XEv_Spine_31] => (item=eigrp) => {"changed": false, "item": "eigrp", "warnings": []}
ok: [XEv_Spine_32] => (item=eigrp) => {"changed": false, "item": "eigrp", "warnings": []}
ok: [XEv_Spine_31] => (item=event-manager) => {"changed": false, "item": "event-manager", "warnings": []}
ok: [XEv_Spine_32] => (item=event-manager) => {"changed": false, "item": "event-manager", "warnings": []}
ok: [XEv_Spine_31] => (item=firewall serverstatus) => {"changed": false, "item": "firewall serverstatus", "warnings": []}
ok: [XEv_Spine_32] => (item=firewall serverstatus) => {"changed": false, "item": "firewall serverstatus", "warnings": []}
ok: [XEv_Spine_32] => (item=ike policy add) => {"changed": false, "item": "ike policy add", "warnings": []}
ok: [XEv_Spine_31] => (item=ike policy add) => {"changed": false, "item": "ike policy add", "warnings": []}
ok: [XEv_Spine_32] => (item=ike policy delete) => {"changed": false, "item": "ike policy delete", "warnings": []}
ok: [XEv_Spine_31] => (item=ike policy delete) => {"changed": false, "item": "ike policy delete", "warnings": []}
ok: [XEv_Spine_32] => (item=ike tunnel start) => {"changed": false, "item": "ike tunnel start", "warnings": []}
ok: [XEv_Spine_31] => (item=ike tunnel start) => {"changed": false, "item": "ike tunnel start", "warnings": []}
ok: [XEv_Spine_31] => (item=ike tunnel stop) => {"changed": false, "item": "ike tunnel stop", "warnings": []}
ok: [XEv_Spine_32] => (item=ike tunnel stop) => {"changed": false, "item": "ike tunnel stop", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec cryptomap add) => {"changed": false, "item": "ipsec cryptomap add", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec cryptomap add) => {"changed": false, "item": "ipsec cryptomap add", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec cryptomap attach) => {"changed": false, "item": "ipsec cryptomap attach", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec cryptomap attach) => {"changed": false, "item": "ipsec cryptomap attach", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec cryptomap delete) => {"changed": false, "item": "ipsec cryptomap delete", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec cryptomap delete) => {"changed": false, "item": "ipsec cryptomap delete", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec cryptomap detach) => {"changed": false, "item": "ipsec cryptomap detach", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec cryptomap detach) => {"changed": false, "item": "ipsec cryptomap detach", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec too-many-sas) => {"changed": false, "item": "ipsec too-many-sas", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec too-many-sas) => {"changed": false, "item": "ipsec too-many-sas", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec tunnel start) => {"changed": false, "item": "ipsec tunnel start", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec tunnel start) => {"changed": false, "item": "ipsec tunnel start", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec tunnel stop) => {"changed": false, "item": "ipsec tunnel stop", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec tunnel stop) => {"changed": false, "item": "ipsec tunnel stop", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsla) => {"changed": false, "item": "ipsla", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsla) => {"changed": false, "item": "ipsla", "warnings": []}
ok: [XEv_Spine_32] => (item=isis) => {"changed": false, "item": "isis", "warnings": []}
ok: [XEv_Spine_31] => (item=isis) => {"changed": false, "item": "isis", "warnings": []}
changed: [XEv_Spine_32] => (item=memory) => {"changed": true, "item": "memory", "updates": ["snmp-server enable traps memory"], "warnings": []}
changed: [XEv_Spine_31] => (item=memory) => {"changed": true, "item": "memory", "updates": ["snmp-server enable traps memory"], "warnings": []}
changed: [XEv_Spine_32] => (item=ospf) => {"changed": true, "item": "ospf", "updates": ["snmp-server enable traps ospf"], "warnings": []}
changed: [XEv_Spine_31] => (item=ospf) => {"changed": true, "item": "ospf", "updates": ["snmp-server enable traps ospf"], "warnings": []}
ok: [XEv_Spine_32] => (item=ospfv3 errors) => {"changed": false, "item": "ospfv3 errors", "warnings": []}
ok: [XEv_Spine_31] => (item=ospfv3 errors) => {"changed": false, "item": "ospfv3 errors", "warnings": []}
ok: [XEv_Spine_32] => (item=ospfv3 rate-limit 60 150) => {"changed": false, "item": "ospfv3 rate-limit 60 150", "warnings": []}
ok: [XEv_Spine_31] => (item=ospfv3 rate-limit 60 150) => {"changed": false, "item": "ospfv3 rate-limit 60 150", "warnings": []}
ok: [XEv_Spine_32] => (item=ospfv3 state-change) => {"changed": false, "item": "ospfv3 state-change", "warnings": []}
ok: [XEv_Spine_31] => (item=ospfv3 state-change) => {"changed": false, "item": "ospfv3 state-change", "warnings": []}
ok: [XEv_Spine_32] => (item=pfr) => {"changed": false, "item": "pfr", "warnings": []}
ok: [XEv_Spine_31] => (item=pfr) => {"changed": false, "item": "pfr", "warnings": []}
ok: [XEv_Spine_32] => (item=pki) => {"changed": false, "item": "pki", "warnings": []}
ok: [XEv_Spine_31] => (item=pki) => {"changed": false, "item": "pki", "warnings": []}
changed: [XEv_Spine_32] => (item=snmp) => {"changed": true, "item": "snmp", "updates": ["snmp-server enable traps snmp"], "warnings": []}
changed: [XEv_Spine_31] => (item=snmp) => {"changed": true, "item": "snmp", "updates": ["snmp-server enable traps snmp"], "warnings": []}
ok: [XEv_Spine_32] => (item=syslog) => {"changed": false, "item": "syslog", "warnings": []}
ok: [XEv_Spine_31] => (item=syslog) => {"changed": false, "item": "syslog", "warnings": []}
ok: [XEv_Spine_32] => (item=tty) => {"changed": false, "item": "tty", "warnings": []}
ok: [XEv_Spine_31] => (item=tty) => {"changed": false, "item": "tty", "warnings": []}
```
Saving the running-config in between multiple runs does not change the situation at all.
|
True
|
ios_config changes some items which do not need to be changed - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ios_config
##### ANSIBLE VERSION
```
ansible 2.3.0 commit d9b570a
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
core modules commit ec4eebc
extras modules commit edbddf6
```
##### CONFIGURATION
inventory = ./hosts
gathering = explicit
roles_path = /home/actionmystique/Program-Files/Ubuntu/Ansible/git-Ansible/roles
private_role_vars = yes
log_path = /var/log/ansible.log
fact_caching = redis
fact_caching_timeout = 86400
retry_files_enabled = False
##### OS / ENVIRONMENT
- host: Ubuntu 16.04 4.4.0
- target: **CSR-1000v 16.2.2**
##### SUMMARY
There is a strange behavior when some commands are correctly applied multiple times in a row: the previous occurrences are not seen as changed.
For instance, the following tasks are always applied even though they have already been successfully run:
```
- name: Configuring location, contact & chassis ID information in new SNMPv{{ snmp.new_version }}
ios_config:
provider: "{{ connections.ssh }}"
lines:
- "snmp-server location {{ snmp.location }}"
- "snmp-server contact {{ snmp.contact }}"
- "snmp-server chassis-id {{ snmp.chassis_id }}"
- "snmp-server ifindex persist"
when: snmp.admin_state == "up"
register: result
```
```
TASK [ios_snmp : Configuring location, contact & chassis ID information in new SNMPv2c] ***
changed: [XEv_Spine_31] => {"changed": true, "updates": ["snmp-server ifindex persist"], "warnings": []}
changed: [XEv_Spine_32] => {"changed": true, "updates": ["snmp-server ifindex persist"], "warnings": []}
```
```
- name: Configuring RO/RW community string in new IPv4/SNMPv{{ snmp.new_version }}
ios_config:
provider: "{{ connections.ssh }}"
lines:
- "snmp-server community {{ snmp.ssh.community }} {{ snmp.community_type }} {{ snmp.acl_name }}"
when: snmp.ip_version == 'ip'
register: result
```
```
TASK [ios_snmp : Configuring RO/RW community string in new IPv4/SNMPv2c] *******
changed: [XEv_Spine_32] => {"changed": true, "updates": ["snmp-server community xxxxxxxx ro authorized-snmp-nms"], "warnings": []}
changed: [XEv_Spine_31] => {"changed": true, "updates": ["snmp-server community xxxxxxxx ro authorized-snmp-nms"], "warnings": []}
```
```
- name: Configuring which NMS is allowed to receive the traps in new SNMPv{{ snmp.new_version }}
ios_config:
provider: "{{ connections.ssh }}"
lines:
- "snmp-server host {{ nms_mgt_ip_address }} {{ snmp.notifications_type }} version {{ snmp.new_version }} {{ snmp.ssh.community }} udp-port {{ snmp.trap_port }}"
when: (snmp.notifications == 'all') or (snmp.notifications == 'list')
register: result
```
```
TASK [ios_snmp : Configuring which NMS is allowed to receive the traps in new SNMPv2c] ***
changed: [XEv_Spine_31] => {"changed": true, "updates": ["snmp-server host 172.21.100.1 traps version 2c xxxxxxxx udp-port 162"], "warnings": []}
changed: [XEv_Spine_32] => {"changed": true, "updates": ["snmp-server host 172.21.100.1 traps version 2c xxxxxxxx udp-port 162"], "warnings": []}
```
Not all commands are concerned with this glitch. For instance, the second passage does not lead to a change for:
```
- name: Configuring the ACL for authorized NMS in new SNMPv{{ snmp.new_version }}
ios_config:
provider: "{{ connections.ssh }}"
parents:
- "{{ snmp.ip_version }} access-list extended {{ snmp.acl_name }}"
lines:
- "permit {{ snmp.ip_version }} host {{ nms_mgt_ip_address }} host {{ ansible_host }} log"
when: snmp.admin_state == "up"
register: result
```
```
TASK [ios_snmp : Configuring the ACL for authorized NMS in new SNMPv2c] ********
ok: [XEv_Spine_32] => {"changed": false, "warnings": []}
ok: [XEv_Spine_31] => {"changed": false, "warnings": []}
```
```
- name: Configuring the list of notifications in new SNMPv{{ snmp.new_version }}
ios_config:
provider: "{{ connections.ssh }}"
lines:
- "snmp-server enable traps {{ item }}"
with_items: "{{ snmp.notifications_list.traps }}"
when: (snmp.notifications == 'list')
register: result
```
The previous task works as expected for all the notifications I have tested, except for a few:
- cpu
- memory
- ospf
- snmp
```
TASK [ios_snmp : Configuring the list of notifications in new SNMPv2c] *********
ok: [XEv_Spine_31] => (item=bfd) => {"changed": false, "item": "bfd", "warnings": []}
ok: [XEv_Spine_32] => (item=bfd) => {"changed": false, "item": "bfd", "warnings": []}
ok: [XEv_Spine_31] => (item=bgp) => {"changed": false, "item": "bgp", "warnings": []}
ok: [XEv_Spine_32] => (item=bgp) => {"changed": false, "item": "bgp", "warnings": []}
changed: [XEv_Spine_31] => (item=cpu) => {"changed": true, "item": "cpu", "updates": ["snmp-server enable traps cpu"], "warnings": []}
changed: [XEv_Spine_32] => (item=cpu) => {"changed": true, "item": "cpu", "updates": ["snmp-server enable traps cpu"], "warnings": []}
ok: [XEv_Spine_31] => (item=eigrp) => {"changed": false, "item": "eigrp", "warnings": []}
ok: [XEv_Spine_32] => (item=eigrp) => {"changed": false, "item": "eigrp", "warnings": []}
ok: [XEv_Spine_31] => (item=event-manager) => {"changed": false, "item": "event-manager", "warnings": []}
ok: [XEv_Spine_32] => (item=event-manager) => {"changed": false, "item": "event-manager", "warnings": []}
ok: [XEv_Spine_31] => (item=firewall serverstatus) => {"changed": false, "item": "firewall serverstatus", "warnings": []}
ok: [XEv_Spine_32] => (item=firewall serverstatus) => {"changed": false, "item": "firewall serverstatus", "warnings": []}
ok: [XEv_Spine_32] => (item=ike policy add) => {"changed": false, "item": "ike policy add", "warnings": []}
ok: [XEv_Spine_31] => (item=ike policy add) => {"changed": false, "item": "ike policy add", "warnings": []}
ok: [XEv_Spine_32] => (item=ike policy delete) => {"changed": false, "item": "ike policy delete", "warnings": []}
ok: [XEv_Spine_31] => (item=ike policy delete) => {"changed": false, "item": "ike policy delete", "warnings": []}
ok: [XEv_Spine_32] => (item=ike tunnel start) => {"changed": false, "item": "ike tunnel start", "warnings": []}
ok: [XEv_Spine_31] => (item=ike tunnel start) => {"changed": false, "item": "ike tunnel start", "warnings": []}
ok: [XEv_Spine_31] => (item=ike tunnel stop) => {"changed": false, "item": "ike tunnel stop", "warnings": []}
ok: [XEv_Spine_32] => (item=ike tunnel stop) => {"changed": false, "item": "ike tunnel stop", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec cryptomap add) => {"changed": false, "item": "ipsec cryptomap add", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec cryptomap add) => {"changed": false, "item": "ipsec cryptomap add", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec cryptomap attach) => {"changed": false, "item": "ipsec cryptomap attach", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec cryptomap attach) => {"changed": false, "item": "ipsec cryptomap attach", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec cryptomap delete) => {"changed": false, "item": "ipsec cryptomap delete", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec cryptomap delete) => {"changed": false, "item": "ipsec cryptomap delete", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec cryptomap detach) => {"changed": false, "item": "ipsec cryptomap detach", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec cryptomap detach) => {"changed": false, "item": "ipsec cryptomap detach", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec too-many-sas) => {"changed": false, "item": "ipsec too-many-sas", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec too-many-sas) => {"changed": false, "item": "ipsec too-many-sas", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec tunnel start) => {"changed": false, "item": "ipsec tunnel start", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec tunnel start) => {"changed": false, "item": "ipsec tunnel start", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsec tunnel stop) => {"changed": false, "item": "ipsec tunnel stop", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsec tunnel stop) => {"changed": false, "item": "ipsec tunnel stop", "warnings": []}
ok: [XEv_Spine_32] => (item=ipsla) => {"changed": false, "item": "ipsla", "warnings": []}
ok: [XEv_Spine_31] => (item=ipsla) => {"changed": false, "item": "ipsla", "warnings": []}
ok: [XEv_Spine_32] => (item=isis) => {"changed": false, "item": "isis", "warnings": []}
ok: [XEv_Spine_31] => (item=isis) => {"changed": false, "item": "isis", "warnings": []}
changed: [XEv_Spine_32] => (item=memory) => {"changed": true, "item": "memory", "updates": ["snmp-server enable traps memory"], "warnings": []}
changed: [XEv_Spine_31] => (item=memory) => {"changed": true, "item": "memory", "updates": ["snmp-server enable traps memory"], "warnings": []}
changed: [XEv_Spine_32] => (item=ospf) => {"changed": true, "item": "ospf", "updates": ["snmp-server enable traps ospf"], "warnings": []}
changed: [XEv_Spine_31] => (item=ospf) => {"changed": true, "item": "ospf", "updates": ["snmp-server enable traps ospf"], "warnings": []}
ok: [XEv_Spine_32] => (item=ospfv3 errors) => {"changed": false, "item": "ospfv3 errors", "warnings": []}
ok: [XEv_Spine_31] => (item=ospfv3 errors) => {"changed": false, "item": "ospfv3 errors", "warnings": []}
ok: [XEv_Spine_32] => (item=ospfv3 rate-limit 60 150) => {"changed": false, "item": "ospfv3 rate-limit 60 150", "warnings": []}
ok: [XEv_Spine_31] => (item=ospfv3 rate-limit 60 150) => {"changed": false, "item": "ospfv3 rate-limit 60 150", "warnings": []}
ok: [XEv_Spine_32] => (item=ospfv3 state-change) => {"changed": false, "item": "ospfv3 state-change", "warnings": []}
ok: [XEv_Spine_31] => (item=ospfv3 state-change) => {"changed": false, "item": "ospfv3 state-change", "warnings": []}
ok: [XEv_Spine_32] => (item=pfr) => {"changed": false, "item": "pfr", "warnings": []}
ok: [XEv_Spine_31] => (item=pfr) => {"changed": false, "item": "pfr", "warnings": []}
ok: [XEv_Spine_32] => (item=pki) => {"changed": false, "item": "pki", "warnings": []}
ok: [XEv_Spine_31] => (item=pki) => {"changed": false, "item": "pki", "warnings": []}
changed: [XEv_Spine_32] => (item=snmp) => {"changed": true, "item": "snmp", "updates": ["snmp-server enable traps snmp"], "warnings": []}
changed: [XEv_Spine_31] => (item=snmp) => {"changed": true, "item": "snmp", "updates": ["snmp-server enable traps snmp"], "warnings": []}
ok: [XEv_Spine_32] => (item=syslog) => {"changed": false, "item": "syslog", "warnings": []}
ok: [XEv_Spine_31] => (item=syslog) => {"changed": false, "item": "syslog", "warnings": []}
ok: [XEv_Spine_32] => (item=tty) => {"changed": false, "item": "tty", "warnings": []}
ok: [XEv_Spine_31] => (item=tty) => {"changed": false, "item": "tty", "warnings": []}
```
Saving the running-config in between multiple runs does not change the situation at all.
|
main
|
ios config changes some items which do not need to be changed issue type bug report component name ios config ansible version ansible commit config file etc ansible ansible cfg configured module search path default w o overrides core modules commit extras modules commit configuration inventory hosts gathering explicit roles path home actionmystique program files ubuntu ansible git ansible roles private role vars yes log path var log ansible log fact caching redis fact caching timeout retry files enabled false os environment host ubuntu target csr summary there is a strange behavior when some commands are correctly applied multiple times in a row the previous occurrences are not seen as changed for instance the following tasks are always applied even though they have already been successfully run name configuring location contact chassis id information in new snmpv snmp new version ios config provider connections ssh lines snmp server location snmp location snmp server contact snmp contact snmp server chassis id snmp chassis id snmp server ifindex persist when snmp admin state up register result task changed changed true updates warnings changed changed true updates warnings name configuring ro rw community string in new snmpv snmp new version ios config provider connections ssh lines snmp server community snmp ssh community snmp community type snmp acl name when snmp ip version ip register result task changed changed true updates warnings changed changed true updates warnings name configuring which nms is allowed to receive the traps in new snmpv snmp new version ios config provider connections ssh lines snmp server host nms mgt ip address snmp notifications type version snmp new version snmp ssh community udp port snmp trap port when snmp notifications all or snmp notifications list register result task changed changed true updates warnings changed changed true updates warnings not all commands are concerned with this glitch for instance the second passage does not lead to a change for name configuring the acl for authorized nms in new snmpv snmp new version ios config provider connections ssh parents snmp ip version access list extended snmp acl name lines permit snmp ip version host nms mgt ip address host ansible host log when snmp admin state up register result task ok changed false warnings ok changed false warnings name configuring the list of notifications in new snmpv snmp new version ios config provider connections ssh lines snmp server enable traps item with items snmp notifications list traps when snmp notifications list register result the previous task works as expected for all the notifications i have tested except for a few cpu memory ospf snmp task ok item bfd changed false item bfd warnings ok item bfd changed false item bfd warnings ok item bgp changed false item bgp warnings ok item bgp changed false item bgp warnings changed item cpu changed true item cpu updates warnings changed item cpu changed true item cpu updates warnings ok item eigrp changed false item eigrp warnings ok item eigrp changed false item eigrp warnings ok item event manager changed false item event manager warnings ok item event manager changed false item event manager warnings ok item firewall serverstatus changed false item firewall serverstatus warnings ok item firewall serverstatus changed false item firewall serverstatus warnings ok item ike policy add changed false item ike policy add warnings ok item ike policy add changed false item ike policy add warnings ok item ike policy delete changed false item ike policy delete warnings ok item ike policy delete changed false item ike policy delete warnings ok item ike tunnel start changed false item ike tunnel start warnings ok item ike tunnel start changed false item ike tunnel start warnings ok item ike tunnel stop changed false item ike tunnel stop warnings ok item ike tunnel stop changed false item ike tunnel stop warnings ok item ipsec cryptomap add changed false item ipsec cryptomap add warnings ok item ipsec cryptomap add changed false item ipsec cryptomap add warnings ok item ipsec cryptomap attach changed false item ipsec cryptomap attach warnings ok item ipsec cryptomap attach changed false item ipsec cryptomap attach warnings ok item ipsec cryptomap delete changed false item ipsec cryptomap delete warnings ok item ipsec cryptomap delete changed false item ipsec cryptomap delete warnings ok item ipsec cryptomap detach changed false item ipsec cryptomap detach warnings ok item ipsec cryptomap detach changed false item ipsec cryptomap detach warnings ok item ipsec too many sas changed false item ipsec too many sas warnings ok item ipsec too many sas changed false item ipsec too many sas warnings ok item ipsec tunnel start changed false item ipsec tunnel start warnings ok item ipsec tunnel start changed false item ipsec tunnel start warnings ok item ipsec tunnel stop changed false item ipsec tunnel stop warnings ok item ipsec tunnel stop changed false item ipsec tunnel stop warnings ok item ipsla changed false item ipsla warnings ok item ipsla changed false item ipsla warnings ok item isis changed false item isis warnings ok item isis changed false item isis warnings changed item memory changed true item memory updates warnings changed item memory changed true item memory updates warnings changed item ospf changed true item ospf updates warnings changed item ospf changed true item ospf updates warnings ok item errors changed false item errors warnings ok item errors changed false item errors warnings ok item rate limit changed false item rate limit warnings ok item rate limit changed false item rate limit warnings ok item state change changed false item state change warnings ok item state change changed false item state change warnings ok item pfr changed false item pfr warnings ok item pfr changed false item pfr warnings ok item pki changed false item pki warnings ok item pki changed false item pki warnings changed item snmp changed true item snmp updates warnings changed item snmp changed true item snmp updates warnings ok item syslog changed false item syslog warnings ok item syslog changed false item syslog warnings ok item tty changed false item tty warnings ok item tty changed false item tty warnings saving the running config in between multiple runs does not change the situation at all
| 1
|
731,208
| 25,205,995,828
|
IssuesEvent
|
2022-11-13 17:37:17
|
windchime-yk/cinelog
|
https://api.github.com/repos/windchime-yk/cinelog
|
opened
|
SurrealDBと接続
|
Type: Feature Priority: High
|
Fly.ioにDockerで包んだSurrealDBをホスティングする記事をあげている人がいたので、これをもとにSurrealDBを立てて利用してみる。
https://www.frontend-devops.com/blog/hosting-surrealdb
Denoとの接続には[公式ドライバーライブラリ](https://surrealdb.com/docs/integration/libraries/deno)を利用する。
|
1.0
|
SurrealDBと接続 - Fly.ioにDockerで包んだSurrealDBをホスティングする記事をあげている人がいたので、これをもとにSurrealDBを立てて利用してみる。
https://www.frontend-devops.com/blog/hosting-surrealdb
Denoとの接続には[公式ドライバーライブラリ](https://surrealdb.com/docs/integration/libraries/deno)を利用する。
|
non_main
|
surrealdbと接続 fly ioにdockerで包んだsurrealdbをホスティングする記事をあげている人がいたので、これをもとにsurrealdbを立てて利用してみる。 denoとの接続には
| 0
|
2,642
| 8,970,970,280
|
IssuesEvent
|
2019-01-29 14:54:49
|
chocolatey/chocolatey-package-requests
|
https://api.github.com/repos/chocolatey/chocolatey-package-requests
|
closed
|
RFP - Pipenv
|
Status: Available For Maintainer(s)
|
https://pipenv.readthedocs.io/en/latest/
> Pipenv is a tool that aims to bring the best of all packaging worlds (bundler, composer, npm, cargo, yarn, etc.) to the Python world. Windows is a first-class citizen, in our world.
|
True
|
RFP - Pipenv - https://pipenv.readthedocs.io/en/latest/
> Pipenv is a tool that aims to bring the best of all packaging worlds (bundler, composer, npm, cargo, yarn, etc.) to the Python world. Windows is a first-class citizen, in our world.
|
main
|
rfp pipenv pipenv is a tool that aims to bring the best of all packaging worlds bundler composer npm cargo yarn etc to the python world windows is a first class citizen in our world
| 1
|
1,341
| 5,721,489,503
|
IssuesEvent
|
2017-04-20 06:47:49
|
tomchentw/react-google-maps
|
https://api.github.com/repos/tomchentw/react-google-maps
|
closed
|
How can I "edit" Polygons and update state to match?
|
CALL_FOR_MAINTAINERS
|
If I allow editable polygons, I can listen for `onMouseup` and get back some "stuff" but I don't see anything which tells me which polygon I've just updated.
I have `this.state.polygons` as an array of polygon objects which contain paths...
`[{"id":0,"editable":true,"path":[{"lat":38.25764366831546,"lng":-85.7516598701477},{"lat":38.25656530505969,"lng":-85.75185298919678},{"lat":38.25808174881664,"lng":-85.75474977493286}],"area":16623.009493033773,"options":{"fillColor":"#caebba"}}]`
What I'm looking for is _(I think)_ a way to update state from what happens to be drawn/edited... either for all Polygons, or for the edited one... but I can't seem to find a listener which tells me which one it is...
NOTE - I know that if I draw the polygons with drawing tools, I have events `onOverlaycomplete` and `onPolygoncomplete` to use and link into w/ more options (raw google map objects) but that doesn't help me when I create with `<Polygon />`
|
True
|
How can I "edit" Polygons and update state to match? - If I allow editable polygons, I can listen for `onMouseup` and get back some "stuff" but I don't see anything which tells me which polygon I've just updated.
I have `this.state.polygons` as an array of polygon objects which contain paths...
`[{"id":0,"editable":true,"path":[{"lat":38.25764366831546,"lng":-85.7516598701477},{"lat":38.25656530505969,"lng":-85.75185298919678},{"lat":38.25808174881664,"lng":-85.75474977493286}],"area":16623.009493033773,"options":{"fillColor":"#caebba"}}]`
What I'm looking for is _(I think)_ a way to update state from what happens to be drawn/edited... either for all Polygons, or for the edited one... but I can't seem to find a listener which tells me which one it is...
NOTE - I know that if I draw the polygons with drawing tools, I have events `onOverlaycomplete` and `onPolygoncomplete` to use and link into w/ more options (raw google map objects) but that doesn't help me when I create with `<Polygon />`
|
main
|
how can i edit polygons and update state to match if i allow editable polygons i can listen for onmouseup and get back some stuff but i don t see anything which tells me which polygon i ve just updated i have this state polygons as an array of polygon objects which contain paths area options fillcolor caebba what i m looking for is i think a way to update state from what happens to be drawn edited either for all polygons or for the edited one but i can t seem to find a listener which tells me which one it is note i know that if i draw the polygons with drawing tools i have events onoverlaycomplete and onpolygoncomplete to use and link into w more options raw google map objects but that doesn t help me when i create with
| 1
|
32,613
| 26,831,026,511
|
IssuesEvent
|
2023-02-02 16:01:32
|
cds-snc/url-shortener
|
https://api.github.com/repos/cds-snc/url-shortener
|
opened
|
Terraform: create `network` module
|
Infrastructure
|
## Description
Create a `network` module that builds the Virtual Private Cloud (VPC) that will hold the service resources. Scan Files has a good example of the resources you would need in this module with a few adjustments:
https://github.com/cds-snc/scan-files/blob/main/terragrunt/aws/api/vpc.tf
Here are the adjustments:
1. You can remove all `aws_vpc_endpoint` except for the `logs` and `s3`.
1. Add a `aws_vpc_endpoint` for `dynamodb`.
1. Add a [Route53 resolver firewall resource](https://github.com/cds-snc/scan-files/blob/ece4d21257be07ced3cb84e3d13c5ff28e675e49/terragrunt/aws/api/route53.tf#L31-L45). This could go in its own `.tf` file in the module. For now, you can just safelist outgoing requests to:
```
*.amazonaws.com.
*.canada.ca.
*.gc.ca.
```
## Notes
Security groups can be tricky to get the hang of - happy to chat more about this one if they're giving you trouble.
|
1.0
|
Terraform: create `network` module - ## Description
Create a `network` module that builds the Virtual Private Cloud (VPC) that will hold the service resources. Scan Files has a good example of the resources you would need in this module with a few adjustments:
https://github.com/cds-snc/scan-files/blob/main/terragrunt/aws/api/vpc.tf
Here are the adjustments:
1. You can remove all `aws_vpc_endpoint` except for the `logs` and `s3`.
1. Add a `aws_vpc_endpoint` for `dynamodb`.
1. Add a [Route53 resolver firewall resource](https://github.com/cds-snc/scan-files/blob/ece4d21257be07ced3cb84e3d13c5ff28e675e49/terragrunt/aws/api/route53.tf#L31-L45). This could go in its own `.tf` file in the module. For now, you can just safelist outgoing requests to:
```
*.amazonaws.com.
*.canada.ca.
*.gc.ca.
```
## Notes
Security groups can be tricky to get the hang of - happy to chat more about this one if they're giving you trouble.
|
non_main
|
terraform create network module description create a network module that builds the virtual private cloud vpc that will hold the service resources scan files has a good example of the resources you would need in this module with a few adjustments here are the adjustments you can remove all aws vpc endpoint except for the logs and add a aws vpc endpoint for dynamodb add a this could go in its own tf file in the module for now you can just safelist outgoing requests to amazonaws com canada ca gc ca notes security groups can be tricky to get the hang of happy to chat more about this one if they re giving you trouble
| 0
|
5,414
| 27,182,131,390
|
IssuesEvent
|
2023-02-18 19:15:51
|
Lissy93/dashy
|
https://api.github.com/repos/Lissy93/dashy
|
closed
|
[QUESTION] Build from source not working
|
🤷♂️ Question 👤 Awaiting Maintainer Response
|
Running RHEL9
When trying to build with yarn I get the following:
yarn install
yarn install v1.22.19
[1/5] Validating package.json...
[2/5] Resolving packages...
[3/5] Fetching packages...
[4/5] Linking dependencies...
warning " > @babel/preset-env@7.18.2" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/helper-compilation-targets@7.18.2" has unmet peer dependency "@babel/core@^7.0.0".
warning "@babel/preset-env > @babel/plugin-bugfix-safari-id-destructuring-collision-in-function-expression@7.17.12" has unmet peer dependency "@babel/core@^7.0.0".
warning "@babel/preset-env > @babel/plugin-bugfix-v8-spread-parameters-in-optional-chaining@7.17.12" has unmet peer dependency "@babel/core@^7.13.0".
warning "@babel/preset-env > @babel/plugin-proposal-async-generator-functions@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-class-properties@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-class-static-block@7.18.0" has unmet peer dependency "@babel/core@^7.12.0".
warning "@babel/preset-env > @babel/plugin-proposal-dynamic-import@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-export-namespace-from@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-json-strings@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-logical-assignment-operators@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-nullish-coalescing-operator@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-numeric-separator@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-object-rest-spread@7.18.0" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-optional-catch-binding@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-optional-chaining@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-private-methods@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-private-property-in-object@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-unicode-property-regex@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-async-generators@7.8.4" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-class-properties@7.12.13" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-class-static-block@7.14.5" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-dynamic-import@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-export-namespace-from@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-import-assertions@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-json-strings@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-logical-assignment-operators@7.10.4" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-nullish-coalescing-operator@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-numeric-separator@7.10.4" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-object-rest-spread@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-optional-catch-binding@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-optional-chaining@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-private-property-in-object@7.14.5" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-top-level-await@7.14.5" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-arrow-functions@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-async-to-generator@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-block-scoped-functions@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-block-scoping@7.18.4" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-classes@7.18.4" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-computed-properties@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-destructuring@7.18.0" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-dotall-regex@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-duplicate-keys@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-exponentiation-operator@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-for-of@7.18.1" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-function-name@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-literals@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-member-expression-literals@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-modules-amd@7.18.0" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-modules-commonjs@7.18.2" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-modules-systemjs@7.18.4" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-modules-umd@7.18.0" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-named-capturing-groups-regex@7.17.12" has unmet peer dependency "@babel/core@^7.0.0".
warning "@babel/preset-env > @babel/plugin-transform-new-target@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-object-super@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-parameters@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-property-literals@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-regenerator@7.18.0" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-reserved-words@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-shorthand-properties@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-spread@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-sticky-regex@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-template-literals@7.18.2" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-typeof-symbol@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-unicode-escapes@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-unicode-regex@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/preset-modules@0.1.5" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > babel-plugin-polyfill-corejs2@0.3.1" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > babel-plugin-polyfill-corejs3@0.5.2" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > babel-plugin-polyfill-regenerator@0.3.1" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-class-properties > @babel/helper-create-class-features-plugin@7.18.0" has unmet peer dependency "@babel/core@^7.0.0".
warning "@babel/preset-env > @babel/plugin-proposal-unicode-property-regex > @babel/helper-create-regexp-features-plugin@7.17.12" has unmet peer dependency "@babel/core@^7.0.0".
warning "@babel/preset-env > babel-plugin-polyfill-corejs2 > @babel/helper-define-polyfill-provider@0.3.1" has unmet peer dependency "@babel/core@^7.4.0-0".
warning " > @vue/cli-plugin-eslint@4.5.17" has incorrect peer dependency "eslint@>= 1.6.0 < 7.0.0".
warning "@vue/cli-plugin-eslint > eslint-loader@2.2.1" has incorrect peer dependency "eslint@>=1.6.0 <7.0.0".
warning " > eslint-config-airbnb@18.2.1" has unmet peer dependency "eslint-plugin-import@^2.22.1".
warning " > eslint-config-airbnb@18.2.1" has unmet peer dependency "eslint-plugin-jsx-a11y@^6.4.1".
warning " > eslint-config-airbnb@18.2.1" has unmet peer dependency "eslint-plugin-react@^7.21.5".
warning " > eslint-config-airbnb@18.2.1" has unmet peer dependency "eslint-plugin-react-hooks@^4 || ^3 || ^2.3.0 || ^1.7.0".
warning "eslint-config-airbnb > eslint-config-airbnb-base@14.2.1" has unmet peer dependency "eslint-plugin-import@^2.22.1".
warning " > progress-bar-webpack-plugin@2.1.0" has unmet peer dependency "webpack@^1.3.0 || ^2 || ^3 || ^4 || ^5".
warning " > sass-loader@7.3.1" has unmet peer dependency "webpack@^3.0.0 || ^4.0.0".
[5/5] Building fresh packages...
Done in 66.99s.
[User@Hostname dashy]$ yarn build
yarn run v1.22.19
$ vue-cli-service build
Browserslist: caniuse-lite is outdated. Please run:
npx browserslist@latest --update-db
Why you should do it regularly: https://github.com/browserslist/browserslist#browsers-data-updating
WARN A new version of sass-loader is available. Please upgrade for best experience.
Building Dashy [== ] 10% (0.1 seconds)Error: error:0308010C:digital envelope routines::unsupported
at new Hash (node:internal/crypto/hash:71:19)
at Object.createHash (node:crypto:130:10)
at module.exports (/home/user/dashy/node_modules/webpack/lib/util/createHash.js:135:53)
at NormalModule._initBuildHash (/home/user/dashy/node_modules/webpack/lib/NormalModule.js:417:16)
at handleParseError (/home/user/dashy/node_modules/webpack/lib/NormalModule.js:471:10)
at /home/user/dashy/node_modules/webpack/lib/NormalModule.js:503:5
at /home/user/dashy/node_modules/webpack/lib/NormalModule.js:358:12
at /home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:373:3
at iterateNormalLoaders (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:214:10)
at iterateNormalLoaders (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:221:10)
at /home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:236:3
at runSyncOrAsync (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:130:11)
at iterateNormalLoaders (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:232:2)
at Array. (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:205:4)
at Storage.finished (/home/user/dashy/node_modules/enhanced-resolve/lib/CachedInputFileSystem.js:55:16)
at /home/user/dashy/node_modules/enhanced-resolve/lib/CachedInputFileSystem.js:91:9
⠏ Building for production...Browserslist: caniuse-lite is outdated. Please run:
npx browserslist@latest --update-db
Why you should do it regularly: https://github.com/browserslist/browserslist#browsers-data-updating
⠧ Building for production.../home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:114
throw e;
^
Error: error:0308010C:digital envelope routines::unsupported
at new Hash (node:internal/crypto/hash:71:19)
at Object.createHash (node:crypto:130:10)
at module.exports (/home/user/dashy/node_modules/webpack/lib/util/createHash.js:135:53)
at NormalModule._initBuildHash (/home/user/dashy/node_modules/webpack/lib/NormalModule.js:417:16)
at handleParseError (/home/user/dashy/node_modules/webpack/lib/NormalModule.js:471:10)
at /home/user/dashy/node_modules/webpack/lib/NormalModule.js:503:5
at /home/user/dashy/node_modules/webpack/lib/NormalModule.js:358:12
at /home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:373:3
at iterateNormalLoaders (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:214:10)
at iterateNormalLoaders (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:221:10)
at /home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:236:3
at context.callback (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:111:13)
at /home/user/dashy/node_modules/cache-loader/dist/index.js:147:7
at /home/user/dashy/node_modules/graceful-fs/graceful-fs.js:61:14
at FSReqCallback.oncomplete (node:fs:192:23) {
opensslErrorStack: [ 'error:03000086:digital envelope routines::initialization error' ],
library: 'digital envelope routines',
reason: 'unsupported',
code: 'ERR_OSSL_EVP_UNSUPPORTED'
}
error Command failed with exit code 1.
info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
I then tried with npm and this was the result:
$ npm install
npm notice
npm notice New major version of npm available! 8.19.2 -> 9.4.2
npm notice Changelog: https://github.com/npm/cli/releases/tag/v9.4.2
npm notice Run npm install -g npm@9.4.2 to update!
npm notice
npm ERR! code ERESOLVE
npm ERR! ERESOLVE could not resolve
npm ERR!
npm ERR! While resolving: @vue/cli-plugin-eslint@4.5.17
npm ERR! Found: eslint@7.32.0
npm ERR! node_modules/eslint
npm ERR! dev eslint@"^7.24.0" from the root project
npm ERR! peer eslint@">= 4.12.1" from babel-eslint@10.1.0
npm ERR! node_modules/babel-eslint
npm ERR! dev babel-eslint@"^10.0.1" from the root project
npm ERR! 9 more (eslint-config-airbnb, eslint-config-airbnb-base, ...)
npm ERR!
npm ERR! Could not resolve dependency:
npm ERR! peer eslint@">= 1.6.0 < 7.0.0" from @vue/cli-plugin-eslint@4.5.17
npm ERR! node_modules/@vue/cli-plugin-eslint
npm ERR! dev @vue/cli-plugin-eslint@"^4.5.15" from the root project
npm ERR!
npm ERR! Conflicting peer dependency: eslint@6.8.0
npm ERR! node_modules/eslint
npm ERR! peer eslint@">= 1.6.0 < 7.0.0" from @vue/cli-plugin-eslint@4.5.17
npm ERR! node_modules/@vue/cli-plugin-eslint
npm ERR! dev @vue/cli-plugin-eslint@"^4.5.15" from the root project
npm ERR!
npm ERR! Fix the upstream dependency conflict, or retry
npm ERR! this command with --force, or --legacy-peer-deps
npm ERR! to accept an incorrect (and potentially broken) dependency resolution.
npm ERR!
npm ERR! See //.npm/eresolve-report.txt for a full report.
npm ERR! A complete log of this run can be found in:
npm ERR! //.npm/_logs/2023-02-09T09_34_38_698Z-debug-0.log
Any idea where I'm going wrong?
Thanks!
|
True
|
[QUESTION] Build from source not working - Running RHEL9
When trying to build with yarn I get the following:
yarn install
yarn install v1.22.19
[1/5] Validating package.json...
[2/5] Resolving packages...
[3/5] Fetching packages...
[4/5] Linking dependencies...
warning " > @babel/preset-env@7.18.2" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/helper-compilation-targets@7.18.2" has unmet peer dependency "@babel/core@^7.0.0".
warning "@babel/preset-env > @babel/plugin-bugfix-safari-id-destructuring-collision-in-function-expression@7.17.12" has unmet peer dependency "@babel/core@^7.0.0".
warning "@babel/preset-env > @babel/plugin-bugfix-v8-spread-parameters-in-optional-chaining@7.17.12" has unmet peer dependency "@babel/core@^7.13.0".
warning "@babel/preset-env > @babel/plugin-proposal-async-generator-functions@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-class-properties@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-class-static-block@7.18.0" has unmet peer dependency "@babel/core@^7.12.0".
warning "@babel/preset-env > @babel/plugin-proposal-dynamic-import@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-export-namespace-from@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-json-strings@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-logical-assignment-operators@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-nullish-coalescing-operator@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-numeric-separator@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-object-rest-spread@7.18.0" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-optional-catch-binding@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-optional-chaining@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-private-methods@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-private-property-in-object@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-unicode-property-regex@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-async-generators@7.8.4" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-class-properties@7.12.13" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-class-static-block@7.14.5" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-dynamic-import@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-export-namespace-from@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-import-assertions@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-json-strings@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-logical-assignment-operators@7.10.4" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-nullish-coalescing-operator@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-numeric-separator@7.10.4" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-object-rest-spread@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-optional-catch-binding@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-optional-chaining@7.8.3" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-private-property-in-object@7.14.5" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-syntax-top-level-await@7.14.5" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-arrow-functions@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-async-to-generator@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-block-scoped-functions@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-block-scoping@7.18.4" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-classes@7.18.4" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-computed-properties@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-destructuring@7.18.0" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-dotall-regex@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-duplicate-keys@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-exponentiation-operator@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-for-of@7.18.1" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-function-name@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-literals@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-member-expression-literals@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-modules-amd@7.18.0" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-modules-commonjs@7.18.2" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-modules-systemjs@7.18.4" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-modules-umd@7.18.0" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-named-capturing-groups-regex@7.17.12" has unmet peer dependency "@babel/core@^7.0.0".
warning "@babel/preset-env > @babel/plugin-transform-new-target@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-object-super@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-parameters@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-property-literals@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-regenerator@7.18.0" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-reserved-words@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-shorthand-properties@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-spread@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-sticky-regex@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-template-literals@7.18.2" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-typeof-symbol@7.17.12" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-unicode-escapes@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-transform-unicode-regex@7.16.7" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/preset-modules@0.1.5" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > babel-plugin-polyfill-corejs2@0.3.1" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > babel-plugin-polyfill-corejs3@0.5.2" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > babel-plugin-polyfill-regenerator@0.3.1" has unmet peer dependency "@babel/core@^7.0.0-0".
warning "@babel/preset-env > @babel/plugin-proposal-class-properties > @babel/helper-create-class-features-plugin@7.18.0" has unmet peer dependency "@babel/core@^7.0.0".
warning "@babel/preset-env > @babel/plugin-proposal-unicode-property-regex > @babel/helper-create-regexp-features-plugin@7.17.12" has unmet peer dependency "@babel/core@^7.0.0".
warning "@babel/preset-env > babel-plugin-polyfill-corejs2 > @babel/helper-define-polyfill-provider@0.3.1" has unmet peer dependency "@babel/core@^7.4.0-0".
warning " > @vue/cli-plugin-eslint@4.5.17" has incorrect peer dependency "eslint@>= 1.6.0 < 7.0.0".
warning "@vue/cli-plugin-eslint > eslint-loader@2.2.1" has incorrect peer dependency "eslint@>=1.6.0 <7.0.0".
warning " > eslint-config-airbnb@18.2.1" has unmet peer dependency "eslint-plugin-import@^2.22.1".
warning " > eslint-config-airbnb@18.2.1" has unmet peer dependency "eslint-plugin-jsx-a11y@^6.4.1".
warning " > eslint-config-airbnb@18.2.1" has unmet peer dependency "eslint-plugin-react@^7.21.5".
warning " > eslint-config-airbnb@18.2.1" has unmet peer dependency "eslint-plugin-react-hooks@^4 || ^3 || ^2.3.0 || ^1.7.0".
warning "eslint-config-airbnb > eslint-config-airbnb-base@14.2.1" has unmet peer dependency "eslint-plugin-import@^2.22.1".
warning " > progress-bar-webpack-plugin@2.1.0" has unmet peer dependency "webpack@^1.3.0 || ^2 || ^3 || ^4 || ^5".
warning " > sass-loader@7.3.1" has unmet peer dependency "webpack@^3.0.0 || ^4.0.0".
[5/5] Building fresh packages...
Done in 66.99s.
[User@Hostname dashy]$ yarn build
yarn run v1.22.19
$ vue-cli-service build
Browserslist: caniuse-lite is outdated. Please run:
npx browserslist@latest --update-db
Why you should do it regularly: https://github.com/browserslist/browserslist#browsers-data-updating
WARN A new version of sass-loader is available. Please upgrade for best experience.
Building Dashy [== ] 10% (0.1 seconds)Error: error:0308010C:digital envelope routines::unsupported
at new Hash (node:internal/crypto/hash:71:19)
at Object.createHash (node:crypto:130:10)
at module.exports (/home/user/dashy/node_modules/webpack/lib/util/createHash.js:135:53)
at NormalModule._initBuildHash (/home/user/dashy/node_modules/webpack/lib/NormalModule.js:417:16)
at handleParseError (/home/user/dashy/node_modules/webpack/lib/NormalModule.js:471:10)
at /home/user/dashy/node_modules/webpack/lib/NormalModule.js:503:5
at /home/user/dashy/node_modules/webpack/lib/NormalModule.js:358:12
at /home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:373:3
at iterateNormalLoaders (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:214:10)
at iterateNormalLoaders (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:221:10)
at /home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:236:3
at runSyncOrAsync (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:130:11)
at iterateNormalLoaders (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:232:2)
at Array. (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:205:4)
at Storage.finished (/home/user/dashy/node_modules/enhanced-resolve/lib/CachedInputFileSystem.js:55:16)
at /home/user/dashy/node_modules/enhanced-resolve/lib/CachedInputFileSystem.js:91:9
⠏ Building for production...Browserslist: caniuse-lite is outdated. Please run:
npx browserslist@latest --update-db
Why you should do it regularly: https://github.com/browserslist/browserslist#browsers-data-updating
⠧ Building for production.../home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:114
throw e;
^
Error: error:0308010C:digital envelope routines::unsupported
at new Hash (node:internal/crypto/hash:71:19)
at Object.createHash (node:crypto:130:10)
at module.exports (/home/user/dashy/node_modules/webpack/lib/util/createHash.js:135:53)
at NormalModule._initBuildHash (/home/user/dashy/node_modules/webpack/lib/NormalModule.js:417:16)
at handleParseError (/home/user/dashy/node_modules/webpack/lib/NormalModule.js:471:10)
at /home/user/dashy/node_modules/webpack/lib/NormalModule.js:503:5
at /home/user/dashy/node_modules/webpack/lib/NormalModule.js:358:12
at /home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:373:3
at iterateNormalLoaders (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:214:10)
at iterateNormalLoaders (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:221:10)
at /home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:236:3
at context.callback (/home/user/dashy/node_modules/loader-runner/lib/LoaderRunner.js:111:13)
at /home/user/dashy/node_modules/cache-loader/dist/index.js:147:7
at /home/user/dashy/node_modules/graceful-fs/graceful-fs.js:61:14
at FSReqCallback.oncomplete (node:fs:192:23) {
opensslErrorStack: [ 'error:03000086:digital envelope routines::initialization error' ],
library: 'digital envelope routines',
reason: 'unsupported',
code: 'ERR_OSSL_EVP_UNSUPPORTED'
}
error Command failed with exit code 1.
info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
I then tried with npm and this was the result:
$ npm install
npm notice
npm notice New major version of npm available! 8.19.2 -> 9.4.2
npm notice Changelog: https://github.com/npm/cli/releases/tag/v9.4.2
npm notice Run npm install -g npm@9.4.2 to update!
npm notice
npm ERR! code ERESOLVE
npm ERR! ERESOLVE could not resolve
npm ERR!
npm ERR! While resolving: @vue/cli-plugin-eslint@4.5.17
npm ERR! Found: eslint@7.32.0
npm ERR! node_modules/eslint
npm ERR! dev eslint@"^7.24.0" from the root project
npm ERR! peer eslint@">= 4.12.1" from babel-eslint@10.1.0
npm ERR! node_modules/babel-eslint
npm ERR! dev babel-eslint@"^10.0.1" from the root project
npm ERR! 9 more (eslint-config-airbnb, eslint-config-airbnb-base, ...)
npm ERR!
npm ERR! Could not resolve dependency:
npm ERR! peer eslint@">= 1.6.0 < 7.0.0" from @vue/cli-plugin-eslint@4.5.17
npm ERR! node_modules/@vue/cli-plugin-eslint
npm ERR! dev @vue/cli-plugin-eslint@"^4.5.15" from the root project
npm ERR!
npm ERR! Conflicting peer dependency: eslint@6.8.0
npm ERR! node_modules/eslint
npm ERR! peer eslint@">= 1.6.0 < 7.0.0" from @vue/cli-plugin-eslint@4.5.17
npm ERR! node_modules/@vue/cli-plugin-eslint
npm ERR! dev @vue/cli-plugin-eslint@"^4.5.15" from the root project
npm ERR!
npm ERR! Fix the upstream dependency conflict, or retry
npm ERR! this command with --force, or --legacy-peer-deps
npm ERR! to accept an incorrect (and potentially broken) dependency resolution.
npm ERR!
npm ERR! See //.npm/eresolve-report.txt for a full report.
npm ERR! A complete log of this run can be found in:
npm ERR! //.npm/_logs/2023-02-09T09_34_38_698Z-debug-0.log
Any idea where I'm going wrong?
Thanks!
|
main
|
build from source not working running when trying to build with yarn i get the following yarn install yarn install validating package json resolving packages fetching packages linking dependencies warning babel preset env has unmet peer dependency babel core warning babel preset env babel helper compilation targets has unmet peer dependency babel core warning babel preset env babel plugin bugfix safari id destructuring collision in function expression has unmet peer dependency babel core warning babel preset env babel plugin bugfix spread parameters in optional chaining has unmet peer dependency babel core warning babel preset env babel plugin proposal async generator functions has unmet peer dependency babel core warning babel preset env babel plugin proposal class properties has unmet peer dependency babel core warning babel preset env babel plugin proposal class static block has unmet peer dependency babel core warning babel preset env babel plugin proposal dynamic import has unmet peer dependency babel core warning babel preset env babel plugin proposal export namespace from has unmet peer dependency babel core warning babel preset env babel plugin proposal json strings has unmet peer dependency babel core warning babel preset env babel plugin proposal logical assignment operators has unmet peer dependency babel core warning babel preset env babel plugin proposal nullish coalescing operator has unmet peer dependency babel core warning babel preset env babel plugin proposal numeric separator has unmet peer dependency babel core warning babel preset env babel plugin proposal object rest spread has unmet peer dependency babel core warning babel preset env babel plugin proposal optional catch binding has unmet peer dependency babel core warning babel preset env babel plugin proposal optional chaining has unmet peer dependency babel core warning babel preset env babel plugin proposal private methods has unmet peer dependency babel core warning babel preset env babel plugin proposal private property in object has unmet peer dependency babel core warning babel preset env babel plugin proposal unicode property regex has unmet peer dependency babel core warning babel preset env babel plugin syntax async generators has unmet peer dependency babel core warning babel preset env babel plugin syntax class properties has unmet peer dependency babel core warning babel preset env babel plugin syntax class static block has unmet peer dependency babel core warning babel preset env babel plugin syntax dynamic import has unmet peer dependency babel core warning babel preset env babel plugin syntax export namespace from has unmet peer dependency babel core warning babel preset env babel plugin syntax import assertions has unmet peer dependency babel core warning babel preset env babel plugin syntax json strings has unmet peer dependency babel core warning babel preset env babel plugin syntax logical assignment operators has unmet peer dependency babel core warning babel preset env babel plugin syntax nullish coalescing operator has unmet peer dependency babel core warning babel preset env babel plugin syntax numeric separator has unmet peer dependency babel core warning babel preset env babel plugin syntax object rest spread has unmet peer dependency babel core warning babel preset env babel plugin syntax optional catch binding has unmet peer dependency babel core warning babel preset env babel plugin syntax optional chaining has unmet peer dependency babel core warning babel preset env babel plugin syntax private property in object has unmet peer dependency babel core warning babel preset env babel plugin syntax top level await has unmet peer dependency babel core warning babel preset env babel plugin transform arrow functions has unmet peer dependency babel core warning babel preset env babel plugin transform async to generator has unmet peer dependency babel core warning babel preset env babel plugin transform block scoped functions has unmet peer dependency babel core warning babel preset env babel plugin transform block scoping has unmet peer dependency babel core warning babel preset env babel plugin transform classes has unmet peer dependency babel core warning babel preset env babel plugin transform computed properties has unmet peer dependency babel core warning babel preset env babel plugin transform destructuring has unmet peer dependency babel core warning babel preset env babel plugin transform dotall regex has unmet peer dependency babel core warning babel preset env babel plugin transform duplicate keys has unmet peer dependency babel core warning babel preset env babel plugin transform exponentiation operator has unmet peer dependency babel core warning babel preset env babel plugin transform for of has unmet peer dependency babel core warning babel preset env babel plugin transform function name has unmet peer dependency babel core warning babel preset env babel plugin transform literals has unmet peer dependency babel core warning babel preset env babel plugin transform member expression literals has unmet peer dependency babel core warning babel preset env babel plugin transform modules amd has unmet peer dependency babel core warning babel preset env babel plugin transform modules commonjs has unmet peer dependency babel core warning babel preset env babel plugin transform modules systemjs has unmet peer dependency babel core warning babel preset env babel plugin transform modules umd has unmet peer dependency babel core warning babel preset env babel plugin transform named capturing groups regex has unmet peer dependency babel core warning babel preset env babel plugin transform new target has unmet peer dependency babel core warning babel preset env babel plugin transform object super has unmet peer dependency babel core warning babel preset env babel plugin transform parameters has unmet peer dependency babel core warning babel preset env babel plugin transform property literals has unmet peer dependency babel core warning babel preset env babel plugin transform regenerator has unmet peer dependency babel core warning babel preset env babel plugin transform reserved words has unmet peer dependency babel core warning babel preset env babel plugin transform shorthand properties has unmet peer dependency babel core warning babel preset env babel plugin transform spread has unmet peer dependency babel core warning babel preset env babel plugin transform sticky regex has unmet peer dependency babel core warning babel preset env babel plugin transform template literals has unmet peer dependency babel core warning babel preset env babel plugin transform typeof symbol has unmet peer dependency babel core warning babel preset env babel plugin transform unicode escapes has unmet peer dependency babel core warning babel preset env babel plugin transform unicode regex has unmet peer dependency babel core warning babel preset env babel preset modules has unmet peer dependency babel core warning babel preset env babel plugin polyfill has unmet peer dependency babel core warning babel preset env babel plugin polyfill has unmet peer dependency babel core warning babel preset env babel plugin polyfill regenerator has unmet peer dependency babel core warning babel preset env babel plugin proposal class properties babel helper create class features plugin has unmet peer dependency babel core warning babel preset env babel plugin proposal unicode property regex babel helper create regexp features plugin has unmet peer dependency babel core warning babel preset env babel plugin polyfill babel helper define polyfill provider has unmet peer dependency babel core warning vue cli plugin eslint has incorrect peer dependency eslint warning vue cli plugin eslint eslint loader has incorrect peer dependency eslint warning eslint config airbnb has unmet peer dependency eslint plugin import warning eslint config airbnb has unmet peer dependency eslint plugin jsx warning eslint config airbnb has unmet peer dependency eslint plugin react warning eslint config airbnb has unmet peer dependency eslint plugin react hooks warning eslint config airbnb eslint config airbnb base has unmet peer dependency eslint plugin import warning progress bar webpack plugin has unmet peer dependency webpack warning sass loader has unmet peer dependency webpack building fresh packages done in yarn build yarn run vue cli service build browserslist caniuse lite is outdated please run npx browserslist latest update db why you should do it regularly warn a new version of sass loader is available please upgrade for best experience building dashy seconds error error digital envelope routines unsupported at new hash node internal crypto hash at object createhash node crypto at module exports home user dashy node modules webpack lib util createhash js at normalmodule initbuildhash home user dashy node modules webpack lib normalmodule js at handleparseerror home user dashy node modules webpack lib normalmodule js at home user dashy node modules webpack lib normalmodule js at home user dashy node modules webpack lib normalmodule js at home user dashy node modules loader runner lib loaderrunner js at iteratenormalloaders home user dashy node modules loader runner lib loaderrunner js at iteratenormalloaders home user dashy node modules loader runner lib loaderrunner js at home user dashy node modules loader runner lib loaderrunner js at runsyncorasync home user dashy node modules loader runner lib loaderrunner js at iteratenormalloaders home user dashy node modules loader runner lib loaderrunner js at array home user dashy node modules loader runner lib loaderrunner js at storage finished home user dashy node modules enhanced resolve lib cachedinputfilesystem js at home user dashy node modules enhanced resolve lib cachedinputfilesystem js ⠏ building for production browserslist caniuse lite is outdated please run npx browserslist latest update db why you should do it regularly ⠧ building for production home user dashy node modules loader runner lib loaderrunner js throw e error error digital envelope routines unsupported at new hash node internal crypto hash at object createhash node crypto at module exports home user dashy node modules webpack lib util createhash js at normalmodule initbuildhash home user dashy node modules webpack lib normalmodule js at handleparseerror home user dashy node modules webpack lib normalmodule js at home user dashy node modules webpack lib normalmodule js at home user dashy node modules webpack lib normalmodule js at home user dashy node modules loader runner lib loaderrunner js at iteratenormalloaders home user dashy node modules loader runner lib loaderrunner js at iteratenormalloaders home user dashy node modules loader runner lib loaderrunner js at home user dashy node modules loader runner lib loaderrunner js at context callback home user dashy node modules loader runner lib loaderrunner js at home user dashy node modules cache loader dist index js at home user dashy node modules graceful fs graceful fs js at fsreqcallback oncomplete node fs opensslerrorstack library digital envelope routines reason unsupported code err ossl evp unsupported error command failed with exit code info visit for documentation about this command i then tried with npm and this was the result npm install npm notice npm notice new major version of npm available npm notice changelog npm notice run npm install g npm to update npm notice npm err code eresolve npm err eresolve could not resolve npm err npm err while resolving vue cli plugin eslint npm err found eslint npm err node modules eslint npm err dev eslint from the root project npm err peer eslint from babel eslint npm err node modules babel eslint npm err dev babel eslint from the root project npm err more eslint config airbnb eslint config airbnb base npm err npm err could not resolve dependency npm err peer eslint from vue cli plugin eslint npm err node modules vue cli plugin eslint npm err dev vue cli plugin eslint from the root project npm err npm err conflicting peer dependency eslint npm err node modules eslint npm err peer eslint from vue cli plugin eslint npm err node modules vue cli plugin eslint npm err dev vue cli plugin eslint from the root project npm err npm err fix the upstream dependency conflict or retry npm err this command with force or legacy peer deps npm err to accept an incorrect and potentially broken dependency resolution npm err npm err see npm eresolve report txt for a full report npm err a complete log of this run can be found in npm err npm logs debug log any idea where i m going wrong thanks
| 1
|
1,594
| 6,572,379,549
|
IssuesEvent
|
2017-09-11 01:51:46
|
ansible/ansible-modules-extras
|
https://api.github.com/repos/ansible/ansible-modules-extras
|
closed
|
ecs_taskdefinition new parameter for task role arn (iam role)
|
affects_2.2 aws cloud feature_idea waiting_on_maintainer
|
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
ecs_taskdefinition
##### ANSIBLE VERSION
2.2
##### CONFIGURATION
NA
##### OS / ENVIRONMENT
NA
##### SUMMARY
AWS have recently added the ability to assign an IAM role to a task definition by specifying the ARN of the role.
This PR adds that functionality by defining a new parameter - task_role_arn
|
True
|
ecs_taskdefinition new parameter for task role arn (iam role) - ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
ecs_taskdefinition
##### ANSIBLE VERSION
2.2
##### CONFIGURATION
NA
##### OS / ENVIRONMENT
NA
##### SUMMARY
AWS have recently added the ability to assign an IAM role to a task definition by specifying the ARN of the role.
This PR adds that functionality by defining a new parameter - task_role_arn
|
main
|
ecs taskdefinition new parameter for task role arn iam role issue type feature idea component name ecs taskdefinition ansible version configuration na os environment na summary aws have recently added the ability to assign an iam role to a task definition by specifying the arn of the role this pr adds that functionality by defining a new parameter task role arn
| 1
|
288,164
| 31,859,617,745
|
IssuesEvent
|
2023-09-15 09:56:32
|
nidhi7598/external_tcpdump-4.9.2_AOSP_10_r33_CVE-2018-14882
|
https://api.github.com/repos/nidhi7598/external_tcpdump-4.9.2_AOSP_10_r33_CVE-2018-14882
|
opened
|
CVE-2018-19519 (Medium) detected in tcpdumptcpdump-4.9.2
|
Mend: dependency security vulnerability
|
## CVE-2018-19519 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tcpdumptcpdump-4.9.2</b></p></summary>
<p>
<p>the TCPdump network dissector</p>
<p>Library home page: <a href=https://github.com/the-tcpdump-group/tcpdump.git>https://github.com/the-tcpdump-group/tcpdump.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/external_tcpdump-4.9.2_AOSP_10_r33_CVE-2018-14882/commit/0b22eff1082a699206667d20bbac5ebcddcd555d">0b22eff1082a699206667d20bbac5ebcddcd555d</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/print-hncp.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In tcpdump 4.9.2, a stack-based buffer over-read exists in the print_prefix function of print-hncp.c via crafted packet data because of missing initialization.
<p>Publish Date: 2018-11-25
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-19519>CVE-2018-19519</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-19519">https://nvd.nist.gov/vuln/detail/CVE-2018-19519</a></p>
<p>Release Date: 2018-11-25</p>
<p>Fix Resolution: tcpdump - 4.9.2-6,4.9.2-4,4.9.2-6,4.9.2-6,4.9.2-4,4.9.2-6,4.9.2-4,4.9.2-6,4.9.2-4;tcpdump-debugsource - 4.9.2-6;tcpdump-debuginfo - 4.9.2-6,4.9.2-4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-19519 (Medium) detected in tcpdumptcpdump-4.9.2 - ## CVE-2018-19519 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tcpdumptcpdump-4.9.2</b></p></summary>
<p>
<p>the TCPdump network dissector</p>
<p>Library home page: <a href=https://github.com/the-tcpdump-group/tcpdump.git>https://github.com/the-tcpdump-group/tcpdump.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/external_tcpdump-4.9.2_AOSP_10_r33_CVE-2018-14882/commit/0b22eff1082a699206667d20bbac5ebcddcd555d">0b22eff1082a699206667d20bbac5ebcddcd555d</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/print-hncp.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In tcpdump 4.9.2, a stack-based buffer over-read exists in the print_prefix function of print-hncp.c via crafted packet data because of missing initialization.
<p>Publish Date: 2018-11-25
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-19519>CVE-2018-19519</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-19519">https://nvd.nist.gov/vuln/detail/CVE-2018-19519</a></p>
<p>Release Date: 2018-11-25</p>
<p>Fix Resolution: tcpdump - 4.9.2-6,4.9.2-4,4.9.2-6,4.9.2-6,4.9.2-4,4.9.2-6,4.9.2-4,4.9.2-6,4.9.2-4;tcpdump-debugsource - 4.9.2-6;tcpdump-debuginfo - 4.9.2-6,4.9.2-4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_main
|
cve medium detected in tcpdumptcpdump cve medium severity vulnerability vulnerable library tcpdumptcpdump the tcpdump network dissector library home page a href found in head commit a href found in base branch main vulnerable source files print hncp c vulnerability details in tcpdump a stack based buffer over read exists in the print prefix function of print hncp c via crafted packet data because of missing initialization publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tcpdump tcpdump debugsource tcpdump debuginfo step up your open source security game with mend
| 0
|
662,969
| 22,158,169,297
|
IssuesEvent
|
2022-06-04 04:35:20
|
wickedManatee/WordleProBugs
|
https://api.github.com/repos/wickedManatee/WordleProBugs
|
opened
|
Add ability for user to track their historical plays
|
enhancement medium-priority
|
I can think of two ways to accomplish this
- just take screenshots at end of game (way easier to code, but probably less storage efficient)
- create a file that save/reads all the guesses and games (lot more effort to code, better storage)
|
1.0
|
Add ability for user to track their historical plays - I can think of two ways to accomplish this
- just take screenshots at end of game (way easier to code, but probably less storage efficient)
- create a file that save/reads all the guesses and games (lot more effort to code, better storage)
|
non_main
|
add ability for user to track their historical plays i can think of two ways to accomplish this just take screenshots at end of game way easier to code but probably less storage efficient create a file that save reads all the guesses and games lot more effort to code better storage
| 0
|
2,734
| 9,673,616,417
|
IssuesEvent
|
2019-05-22 08:00:38
|
RalfKoban/MiKo-Analyzers
|
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
|
opened
|
Public Methods should not return List<> or Dictionary<>
|
Area: analyzer Area: maintainability feature
|
Methods that are `public` visible should not return a `List<>` or `Dictionary<>`. Instead, they should return the interfaces.
Doing so allows to change the implementation when it's needed because otherwise the method is bound to always and forever return a `List<>` or `Dictionary<>`.
|
True
|
Public Methods should not return List<> or Dictionary<> - Methods that are `public` visible should not return a `List<>` or `Dictionary<>`. Instead, they should return the interfaces.
Doing so allows to change the implementation when it's needed because otherwise the method is bound to always and forever return a `List<>` or `Dictionary<>`.
|
main
|
public methods should not return list or dictionary methods that are public visible should not return a list or dictionary instead they should return the interfaces doing so allows to change the implementation when it s needed because otherwise the method is bound to always and forever return a list or dictionary
| 1
|
1,500
| 6,495,362,170
|
IssuesEvent
|
2017-08-22 04:38:23
|
backdrop-ops/contrib
|
https://api.github.com/repos/backdrop-ops/contrib
|
closed
|
Port of Twilio module; request to join the Backdrop contrib group
|
Maintainer application Port status
|
I would like to join the Backdrop contrib group here on Github. I have ported the Drupal 7 version of the twilio module:
https://github.com/mclinn/twilio/tree/1.11_local
I posted a "backdrop-port" issue on drupal.org on 8/9/17:
https://www.drupal.org/node/2901091
No reply yet. As I had already begun the work, I decided to go ahead with the port rather than wait -- I am ok with the gamble that the original author might want to do this himself. It sounds like he is involved with a port to Drupal 8 so may not be interested / available anyway. I think Backdrop is fantastic so I enjoyed the port as a way to begin to get familiar with Backdrop.
The Drupal version includes a second module, twilio_twiml, to make use of "twiml," the "Twilio Markup Language." I have not included this in the Backdrop port. I have found very little information about using this module; some of the little bit I've found implies that the PHP filter is required to make it work. I don't know if this is true but since the primary module works fine without this I decided to leave it for a possible later project.
I realize that the SMS Framework module, already ported, works well with Twilio and has some additional functionality. This module might be a good fit for someone already familiar with Twilio who just wanted to stick with that platform. The Rules integration seems to work well also.
One neat "use case" could be sending a text to put your site into or out of maintenance mode. It looks like a simple addition to rules could make this work. I will test this out when I get a chance, maybe add the rules hook in an update later on.
Anyway please let me know how this looks and what else I need to do. Thanks!
And thanks to all who created Backdrop, what a lifesaver. I am a textbook example of the kind of person who simply would not have been able to move to Drupal 8 for a variety of reasons. Backdrop has saved the day!
|
True
|
Port of Twilio module; request to join the Backdrop contrib group - I would like to join the Backdrop contrib group here on Github. I have ported the Drupal 7 version of the twilio module:
https://github.com/mclinn/twilio/tree/1.11_local
I posted a "backdrop-port" issue on drupal.org on 8/9/17:
https://www.drupal.org/node/2901091
No reply yet. As I had already begun the work, I decided to go ahead with the port rather than wait -- I am ok with the gamble that the original author might want to do this himself. It sounds like he is involved with a port to Drupal 8 so may not be interested / available anyway. I think Backdrop is fantastic so I enjoyed the port as a way to begin to get familiar with Backdrop.
The Drupal version includes a second module, twilio_twiml, to make use of "twiml," the "Twilio Markup Language." I have not included this in the Backdrop port. I have found very little information about using this module; some of the little bit I've found implies that the PHP filter is required to make it work. I don't know if this is true but since the primary module works fine without this I decided to leave it for a possible later project.
I realize that the SMS Framework module, already ported, works well with Twilio and has some additional functionality. This module might be a good fit for someone already familiar with Twilio who just wanted to stick with that platform. The Rules integration seems to work well also.
One neat "use case" could be sending a text to put your site into or out of maintenance mode. It looks like a simple addition to rules could make this work. I will test this out when I get a chance, maybe add the rules hook in an update later on.
Anyway please let me know how this looks and what else I need to do. Thanks!
And thanks to all who created Backdrop, what a lifesaver. I am a textbook example of the kind of person who simply would not have been able to move to Drupal 8 for a variety of reasons. Backdrop has saved the day!
|
main
|
port of twilio module request to join the backdrop contrib group i would like to join the backdrop contrib group here on github i have ported the drupal version of the twilio module i posted a backdrop port issue on drupal org on no reply yet as i had already begun the work i decided to go ahead with the port rather than wait i am ok with the gamble that the original author might want to do this himself it sounds like he is involved with a port to drupal so may not be interested available anyway i think backdrop is fantastic so i enjoyed the port as a way to begin to get familiar with backdrop the drupal version includes a second module twilio twiml to make use of twiml the twilio markup language i have not included this in the backdrop port i have found very little information about using this module some of the little bit i ve found implies that the php filter is required to make it work i don t know if this is true but since the primary module works fine without this i decided to leave it for a possible later project i realize that the sms framework module already ported works well with twilio and has some additional functionality this module might be a good fit for someone already familiar with twilio who just wanted to stick with that platform the rules integration seems to work well also one neat use case could be sending a text to put your site into or out of maintenance mode it looks like a simple addition to rules could make this work i will test this out when i get a chance maybe add the rules hook in an update later on anyway please let me know how this looks and what else i need to do thanks and thanks to all who created backdrop what a lifesaver i am a textbook example of the kind of person who simply would not have been able to move to drupal for a variety of reasons backdrop has saved the day
| 1
|
5,059
| 25,917,431,265
|
IssuesEvent
|
2022-12-15 18:35:38
|
ElasticPerch/mux
|
https://api.github.com/repos/ElasticPerch/mux
|
closed
|
Prep/rename module for Gorilla fork
|
maintainence
|
Update documents, Licenses, readme and Go modules for project renaming to continue on the Gorilla project.
|
True
|
Prep/rename module for Gorilla fork - Update documents, Licenses, readme and Go modules for project renaming to continue on the Gorilla project.
|
main
|
prep rename module for gorilla fork update documents licenses readme and go modules for project renaming to continue on the gorilla project
| 1
|
1,551
| 6,572,245,438
|
IssuesEvent
|
2017-09-11 00:32:52
|
ansible/ansible-modules-extras
|
https://api.github.com/repos/ansible/ansible-modules-extras
|
closed
|
known_hosts is not idempotent
|
affects_2.1 bug_report waiting_on_maintainer
|
<!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
known_hosts
##### ANSIBLE VERSION
```
ansible 2.1.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
<!--- Paste verbatim output from “ansible --version” between quotes -->
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Debian Jessie
##### SUMMARY
<!--- Explain the problem briefly -->
If I call known_hosts to add a given key it changes on every call
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
Easy to reproduce, insert new key via known_hosts module in your .ssh/known_hosts, check for it, retry it again, and see that it changes again even if the correct key is already in your known_hosts.
Example:
```
# Insert Key using known_hosts module
uhochholdinger@xelo:~$ ansible -i 'localhost,' all -m known_hosts -a 'key="xelo ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJzrHvJZ7bVys4f73Qr1B3qgLkX7sUP6JUpUgAJCCacTeelzQu2JbbOr6aho7K4k1iTZICfaEXEwFMAMtuchpw4= root@xelo" name="xelo"'
localhost | SUCCESS => {
"changed": true,
"gid": 10000,
"group": "np",
"key": "xelo ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJzrHvJZ7bVys4f73Qr1B3qgLkX7sUP6JUpUgAJCCacTeelzQu2JbbOr6aho7K4k1iTZICfaEXEwFMAMtuchpw4= root@xelo",
"mode": "0600",
"name": "xelo",
"owner": "uhochholdinger",
"path": "/home/uhochholdinger/.ssh/known_hosts",
"size": 300811,
"state": "file",
"uid": 1016
}
# Check if key is available, compare with added key!
uhochholdinger@xelo:~$ ssh-keygen -F xelo
# Host xelo found: line 978 type ECDSA
xelo ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJzrHvJZ7bVys4f73Qr1B3qgLkX7sUP6JUpUgAJCCacTeelzQu2JbbOr6aho7K4k1iTZICfaEXEwFMAMtuchpw4=
# Try to add key again
uhochholdinger@xelo:~$ ansible -i 'localhost,' all -m known_hosts -a 'key="xelo ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJzrHvJZ7bVys4f73Qr1B3qgLkX7sUP6JUpUgAJCCacTeelzQu2JbbOr6aho7K4k1iTZICfaEXEwFMAMtuchpw4= root@xelo" name="xelo"'
localhost | SUCCESS => {
"changed": true,
"gid": 10000,
"group": "np",
"key": "xelo ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJzrHvJZ7bVys4f73Qr1B3qgLkX7sUP6JUpUgAJCCacTeelzQu2JbbOr6aho7K4k1iTZICfaEXEwFMAMtuchpw4= root@xelo",
"mode": "0600",
"name": "xelo",
"owner": "uhochholdinger",
"path": "/home/uhochholdinger/.ssh/known_hosts",
"size": 300811,
"state": "file",
"uid": 1016
}
```
Cheers
Ulli
|
True
|
known_hosts is not idempotent - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
known_hosts
##### ANSIBLE VERSION
```
ansible 2.1.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
<!--- Paste verbatim output from “ansible --version” between quotes -->
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Debian Jessie
##### SUMMARY
<!--- Explain the problem briefly -->
If I call known_hosts to add a given key it changes on every call
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
Easy to reproduce, insert new key via known_hosts module in your .ssh/known_hosts, check for it, retry it again, and see that it changes again even if the correct key is already in your known_hosts.
Example:
```
# Insert Key using known_hosts module
uhochholdinger@xelo:~$ ansible -i 'localhost,' all -m known_hosts -a 'key="xelo ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJzrHvJZ7bVys4f73Qr1B3qgLkX7sUP6JUpUgAJCCacTeelzQu2JbbOr6aho7K4k1iTZICfaEXEwFMAMtuchpw4= root@xelo" name="xelo"'
localhost | SUCCESS => {
"changed": true,
"gid": 10000,
"group": "np",
"key": "xelo ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJzrHvJZ7bVys4f73Qr1B3qgLkX7sUP6JUpUgAJCCacTeelzQu2JbbOr6aho7K4k1iTZICfaEXEwFMAMtuchpw4= root@xelo",
"mode": "0600",
"name": "xelo",
"owner": "uhochholdinger",
"path": "/home/uhochholdinger/.ssh/known_hosts",
"size": 300811,
"state": "file",
"uid": 1016
}
# Check if key is available, compare with added key!
uhochholdinger@xelo:~$ ssh-keygen -F xelo
# Host xelo found: line 978 type ECDSA
xelo ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJzrHvJZ7bVys4f73Qr1B3qgLkX7sUP6JUpUgAJCCacTeelzQu2JbbOr6aho7K4k1iTZICfaEXEwFMAMtuchpw4=
# Try to add key again
uhochholdinger@xelo:~$ ansible -i 'localhost,' all -m known_hosts -a 'key="xelo ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJzrHvJZ7bVys4f73Qr1B3qgLkX7sUP6JUpUgAJCCacTeelzQu2JbbOr6aho7K4k1iTZICfaEXEwFMAMtuchpw4= root@xelo" name="xelo"'
localhost | SUCCESS => {
"changed": true,
"gid": 10000,
"group": "np",
"key": "xelo ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJzrHvJZ7bVys4f73Qr1B3qgLkX7sUP6JUpUgAJCCacTeelzQu2JbbOr6aho7K4k1iTZICfaEXEwFMAMtuchpw4= root@xelo",
"mode": "0600",
"name": "xelo",
"owner": "uhochholdinger",
"path": "/home/uhochholdinger/.ssh/known_hosts",
"size": 300811,
"state": "file",
"uid": 1016
}
```
Cheers
Ulli
|
main
|
known hosts is not idempotent issue type bug report component name known hosts ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific debian jessie summary if i call known hosts to add a given key it changes on every call steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used easy to reproduce insert new key via known hosts module in your ssh known hosts check for it retry it again and see that it changes again even if the correct key is already in your known hosts example insert key using known hosts module uhochholdinger xelo ansible i localhost all m known hosts a key xelo ecdsa root xelo name xelo localhost success changed true gid group np key xelo ecdsa root xelo mode name xelo owner uhochholdinger path home uhochholdinger ssh known hosts size state file uid check if key is available compare with added key uhochholdinger xelo ssh keygen f xelo host xelo found line type ecdsa xelo ecdsa try to add key again uhochholdinger xelo ansible i localhost all m known hosts a key xelo ecdsa root xelo name xelo localhost success changed true gid group np key xelo ecdsa root xelo mode name xelo owner uhochholdinger path home uhochholdinger ssh known hosts size state file uid cheers ulli
| 1
|
5,751
| 30,476,083,328
|
IssuesEvent
|
2023-07-17 16:36:30
|
MozillaFoundation/foundation.mozilla.org
|
https://api.github.com/repos/MozillaFoundation/foundation.mozilla.org
|
closed
|
Wagtail 4.1.4 security upgrade
|
engineering backend maintain security
|
## Description
Wagtail 4.1.4 security upgrade will be released on Monday, April 3 2023.
> Advance notice of upcoming Wagtail security releases - I'd like to announce that Wagtail 4.1.4 and 4.2.2 will be released on Monday 3rd April, around 1100 UTC. These will fix a security defect with severity "moderate", along with several other bugfixes already logged at https://docs.wagtail.org/en/latest/releases/4.1.4.html / https://docs.wagtail.org/en/latest/releases/4.2.2.html. For details of severity levels, see: https://docs.djangoproject.com/en/dev/internals/security/#how-django-discloses-security-issues
https://wagtailcms.slack.com/archives/C0K2WLTRS/p1679934916910999
## Acceptance criteria
- [ ] Wagtail is upgraded to version 4.1.4
|
True
|
Wagtail 4.1.4 security upgrade - ## Description
Wagtail 4.1.4 security upgrade will be released on Monday, April 3 2023.
> Advance notice of upcoming Wagtail security releases - I'd like to announce that Wagtail 4.1.4 and 4.2.2 will be released on Monday 3rd April, around 1100 UTC. These will fix a security defect with severity "moderate", along with several other bugfixes already logged at https://docs.wagtail.org/en/latest/releases/4.1.4.html / https://docs.wagtail.org/en/latest/releases/4.2.2.html. For details of severity levels, see: https://docs.djangoproject.com/en/dev/internals/security/#how-django-discloses-security-issues
https://wagtailcms.slack.com/archives/C0K2WLTRS/p1679934916910999
## Acceptance criteria
- [ ] Wagtail is upgraded to version 4.1.4
|
main
|
wagtail security upgrade description wagtail security upgrade will be released on monday april advance notice of upcoming wagtail security releases i d like to announce that wagtail and will be released on monday april around utc these will fix a security defect with severity moderate along with several other bugfixes already logged at for details of severity levels see acceptance criteria wagtail is upgraded to version
| 1
|
91,155
| 11,472,189,015
|
IssuesEvent
|
2020-02-09 15:52:34
|
Quaver/Quaver
|
https://api.github.com/repos/Quaver/Quaver
|
opened
|
Song select takes too long to scroll
|
Bug Song Select UI Redesign
|
When doing an F2 over all mapsets or Ctrl+PageDown it takes good 2 seconds depending on how far away the map is. Instead, it should always take a fixed short time, like 0.5 seconds.
|
1.0
|
Song select takes too long to scroll - When doing an F2 over all mapsets or Ctrl+PageDown it takes good 2 seconds depending on how far away the map is. Instead, it should always take a fixed short time, like 0.5 seconds.
|
non_main
|
song select takes too long to scroll when doing an over all mapsets or ctrl pagedown it takes good seconds depending on how far away the map is instead it should always take a fixed short time like seconds
| 0
|
745,114
| 25,970,830,842
|
IssuesEvent
|
2022-12-19 11:05:06
|
Lakitna/Rulebound
|
https://api.github.com/repos/Lakitna/Rulebound
|
closed
|
Allow rule to alias multiple rules
|
enhancement priority: low
|
Right now a rule can only be alias a single rule.
To allow more flexibility it's a good idea to make it so any number of rules can be defined as an alias. This will increase the complexity of enforcing aliases, so a refactor might be needed to implement this properly.
|
1.0
|
Allow rule to alias multiple rules - Right now a rule can only be alias a single rule.
To allow more flexibility it's a good idea to make it so any number of rules can be defined as an alias. This will increase the complexity of enforcing aliases, so a refactor might be needed to implement this properly.
|
non_main
|
allow rule to alias multiple rules right now a rule can only be alias a single rule to allow more flexibility it s a good idea to make it so any number of rules can be defined as an alias this will increase the complexity of enforcing aliases so a refactor might be needed to implement this properly
| 0
|
1,250
| 5,313,304,391
|
IssuesEvent
|
2017-02-13 11:45:57
|
ansible/ansible-modules-core
|
https://api.github.com/repos/ansible/ansible-modules-core
|
closed
|
win_file: Failed to delete folder with broken symlink
|
affects_2.2 bug_report waiting_on_maintainer windows
|
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ansible-modules-core/windows/win_file.ps1
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel 7e0074263d) last updated 2016/10/02 10:15:29 (GMT +1100)
lib/ansible/modules/core: (detached HEAD 18f710fe32) last updated 2016/10/02 10:15:41 (GMT +1100)
lib/ansible/modules/extras: (detached HEAD a58e1d59c0) last updated 2016/10/02 10:15:53 (GMT +1100)
config file = /etc/ansible/ansible.cfg
configured module search path = ['/usr/share/ansible']
```
##### OS / ENVIRONMENT
Running: Bash for Ubuntu on Windows
Managing: Windows Server 2012 R2
##### SUMMARY
When setting win_file state: absent for a directory that contains a symlink that is no longer pointing to a valid directory it fails with DirectoryNotFoundException
##### STEPS TO REPRODUCE
Inventory.ini
```
[windows]
host
[windows:vars]
ansible_user=.\User
ansible_password=Password
ansible_connection=winrm
ansible_port=5986
ansible_winrm_transport=ntlm
ansible_winrm_server_cert_validation=ignore
```
Playbook.yml
```
- name: test windows
hosts: windows
tasks:
- name: create test folder
win_file:
path: C:\temp\test
state: directory
- name: create symlink
raw: CMD.EXE /C mklink /d C:\temp\link C:\temp\test
- name: delete folder that link is pointing to
win_file:
path: C:\temp\test
state: absent
- name: delete parent folder containing broken link
win_file:
path: C:\temp
state: absent
```
##### EXPECTED RESULTS
Expected C:\temp to be deleted correctly without exception
##### ACTUAL RESULTS
```
An exception occurred during task execution. The full traceback is:
At C:\Users\User\AppData\Local\Temp\ansible-tmp-1475364906.0-88571632788975\win_file.ps1:317 char:9
+ Remove-Item -Recurse -Force $fileinfo
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
fatal: [192.168.1.13]: FAILED! => {
"changed": false,
"error_record": {
"CategoryInfo": {
"Activity": "Remove-Item",
"Category": 23,
"Reason": "DirectoryNotFoundException",
"TargetName": "C:\\temp",
"TargetType": "String"
},
"ErrorDetails": null,
"Exception": {
"Data": {},
"HResult": -2147024893,
"HelpLink": null,
"InnerException": null,
"Message": "Could not find a part of the path 'C:\\temp\\link'.",
"Source": "mscorlib",
"StackTrace": " at System.IO.__Error.WinIOError(Int32 errorCode, String maybeFullPath)\r\n at System.IO.FileSystemEnumerableIterator`1.CommonInit()\r\n at System.IO.FileSystemEnumerableIterator`1..ctor(String path, String originalUserPath, String searchPattern, SearchOption searchOption, SearchResultHandler`1 resultHandler, Boolean checkHost)\r\n at System.IO.FileSystemEnumerableFactory.CreateDirectoryInfoIterator(String path, String originalUserPath, String searchPattern, SearchOption searchOption)\r\n at Microsoft.PowerShell.Commands.FileSystemProvider.RemoveDirectoryInfoItem(DirectoryInfo directory, Boolean recurse, Boolean force, Boolean rootOfRemoval)\r\n at Microsoft.PowerShell.Commands.FileSystemProvider.RemoveDirectoryInfoItem(DirectoryInfo directory, Boolean recurse, Boolean force, Boolean rootOfRemoval)\r\n at Microsoft.PowerShell.Commands.FileSystemProvider.RemoveItem(String path, Boolean recurse)",
"TargetSite": {
"Attributes": 147,
"CallingConvention": 1,
"ContainsGenericParameters": false,
"CustomAttributes": "[System.Security.SecurityCriticalAttribute()]",
"DeclaringType": "System.IO.__Error, mscorlib, Version=4.0.0.0, Culture=neutral, PublicKeyToken=b77a5c561934e089",
"IsAbstract": false,
"IsAssembly": true,
"IsConstructor": false,
"IsFamily": false,
"IsFamilyAndAssembly": false,
"IsFamilyOrAssembly": false,
"IsFinal": false,
"IsGenericMethod": false,
"IsGenericMethodDefinition": false,
"IsHideBySig": true,
"IsPrivate": false,
"IsPublic": false,
"IsSecurityCritical": true,
"IsSecuritySafeCritical": false,
"IsSecurityTransparent": false,
"IsSpecialName": false,
"IsStatic": true,
"IsVirtual": false,
"MemberType": 8,
"MetadataToken": 100680987,
"MethodHandle": "System.RuntimeMethodHandle",
"MethodImplementationFlags": 0,
"Module": "CommonLanguageRuntimeLibrary",
"Name": "WinIOError",
"ReflectedType": "System.IO.__Error, mscorlib, Version=4.0.0.0, Culture=neutral, PublicKeyToken=b77a5c561934e089",
"ReturnParameter": "Void ",
"ReturnType": "void",
"ReturnTypeCustomAttributes": "Void "
}
},
"FullyQualifiedErrorId": "RemoveItemIOError,Microsoft.PowerShell.Commands.RemoveItemCommand",
"InvocationInfo": {
"BoundParameters": {},
"CommandOrigin": 1,
"DisplayScriptPosition": null,
"ExpectingInput": false,
"HistoryId": 1,
"InvocationName": "Remove-Item",
"Line": " Remove-Item -Recurse -Force $fileinfo\n",
"MyCommand": {
"CommandType": 8,
"DefaultParameterSet": "Path",
"Definition": "\r\nRemove-Item [-Path] <string[]> [-Filter <string>] [-Include <string[]>] [-Exclude <string[]>] [-Recurse] [-Force] [-Credential <pscredential>] [-WhatIf] [-Confirm] [-UseTransaction] [<CommonParameters>]\r\n\r\nRemove-Item -LiteralPath <string[]> [-Filter <string>] [-Include <string[]>] [-Exclude <string[]>] [-Recurse] [-Force] [-Credential <pscredential>] [-WhatIf] [-Confirm] [-UseTransaction] [<CommonParameters>]\r\n",
"HelpFile": "Microsoft.PowerShell.Commands.Management.dll-Help.xml",
"ImplementingType": "Microsoft.PowerShell.Commands.RemoveItemCommand",
"Module": "Microsoft.PowerShell.Management",
"ModuleName": "Microsoft.PowerShell.Management",
"Name": "Remove-Item",
"Noun": "Item",
"Options": 1,
"OutputType": "",
"PSSnapIn": null,
"ParameterSets": "[-Path] <string[]> [-Filter <string>] [-Include <string[]>] [-Exclude <string[]>] [-Recurse] [-Force] [-Credential <pscredential>] [-WhatIf] [-Confirm] [-UseTransaction] [<CommonParameters>] -LiteralPath <string[]> [-Filter <string>] [-Include <string[]>] [-Exclude <string[]>] [-Recurse] [-Force] [-Credential <pscredential>] [-WhatIf] [-Confirm] [-UseTransaction] [<CommonParameters>]",
"Parameters": "System.Collections.Generic.Dictionary`2[System.String,System.Management.Automation.ParameterMetadata]",
"RemotingCapability": 1,
"Verb": "Remove",
"Visibility": 0
},
"OffsetInLine": 9,
"PSCommandPath": "C:\\Users\\User\\AppData\\Local\\Temp\\ansible-tmp-1475364906.0-88571632788975\\win_file.ps1",
"PSScriptRoot": "C:\\Users\\User\\AppData\\Local\\Temp\\ansible-tmp-1475364906.0-88571632788975",
"PipelineLength": 0,
"PipelinePosition": 0,
"PositionMessage": "At C:\\Users\\User\\AppData\\Local\\Temp\\ansible-tmp-1475364906.0-88571632788975\\win_file.ps1:317 char:9\r\n+ Remove-Item -Recurse -Force $fileinfo\r\n+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~",
"ScriptLineNumber": 317,
"ScriptName": "C:\\Users\\User\\AppData\\Local\\Temp\\ansible-tmp-1475364906.0-88571632788975\\win_file.ps1",
"UnboundArguments": []
},
"PSMessageDetails": null,
"PipelineIterationInfo": [],
"ScriptStackTrace": "at <ScriptBlock>, C:\\Users\\User\\AppData\\Local\\Temp\\ansible-tmp-1475364906.0-88571632788975\\win_file.ps1: line 317\r\nat <ScriptBlock>, <No file>: line 4",
"TargetObject": "C:\\temp"
},
"failed": true,
"invocation": {
"module_name": "win_file"
},
"msg": "Could not find a part of the path 'C:\\temp\\link'."
}
```
|
True
|
win_file: Failed to delete folder with broken symlink - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ansible-modules-core/windows/win_file.ps1
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel 7e0074263d) last updated 2016/10/02 10:15:29 (GMT +1100)
lib/ansible/modules/core: (detached HEAD 18f710fe32) last updated 2016/10/02 10:15:41 (GMT +1100)
lib/ansible/modules/extras: (detached HEAD a58e1d59c0) last updated 2016/10/02 10:15:53 (GMT +1100)
config file = /etc/ansible/ansible.cfg
configured module search path = ['/usr/share/ansible']
```
##### OS / ENVIRONMENT
Running: Bash for Ubuntu on Windows
Managing: Windows Server 2012 R2
##### SUMMARY
When setting win_file state: absent for a directory that contains a symlink that is no longer pointing to a valid directory it fails with DirectoryNotFoundException
##### STEPS TO REPRODUCE
Inventory.ini
```
[windows]
host
[windows:vars]
ansible_user=.\User
ansible_password=Password
ansible_connection=winrm
ansible_port=5986
ansible_winrm_transport=ntlm
ansible_winrm_server_cert_validation=ignore
```
Playbook.yml
```
- name: test windows
hosts: windows
tasks:
- name: create test folder
win_file:
path: C:\temp\test
state: directory
- name: create symlink
raw: CMD.EXE /C mklink /d C:\temp\link C:\temp\test
- name: delete folder that link is pointing to
win_file:
path: C:\temp\test
state: absent
- name: delete parent folder containing broken link
win_file:
path: C:\temp
state: absent
```
##### EXPECTED RESULTS
Expected C:\temp to be deleted correctly without exception
##### ACTUAL RESULTS
```
An exception occurred during task execution. The full traceback is:
At C:\Users\User\AppData\Local\Temp\ansible-tmp-1475364906.0-88571632788975\win_file.ps1:317 char:9
+ Remove-Item -Recurse -Force $fileinfo
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
fatal: [192.168.1.13]: FAILED! => {
"changed": false,
"error_record": {
"CategoryInfo": {
"Activity": "Remove-Item",
"Category": 23,
"Reason": "DirectoryNotFoundException",
"TargetName": "C:\\temp",
"TargetType": "String"
},
"ErrorDetails": null,
"Exception": {
"Data": {},
"HResult": -2147024893,
"HelpLink": null,
"InnerException": null,
"Message": "Could not find a part of the path 'C:\\temp\\link'.",
"Source": "mscorlib",
"StackTrace": " at System.IO.__Error.WinIOError(Int32 errorCode, String maybeFullPath)\r\n at System.IO.FileSystemEnumerableIterator`1.CommonInit()\r\n at System.IO.FileSystemEnumerableIterator`1..ctor(String path, String originalUserPath, String searchPattern, SearchOption searchOption, SearchResultHandler`1 resultHandler, Boolean checkHost)\r\n at System.IO.FileSystemEnumerableFactory.CreateDirectoryInfoIterator(String path, String originalUserPath, String searchPattern, SearchOption searchOption)\r\n at Microsoft.PowerShell.Commands.FileSystemProvider.RemoveDirectoryInfoItem(DirectoryInfo directory, Boolean recurse, Boolean force, Boolean rootOfRemoval)\r\n at Microsoft.PowerShell.Commands.FileSystemProvider.RemoveDirectoryInfoItem(DirectoryInfo directory, Boolean recurse, Boolean force, Boolean rootOfRemoval)\r\n at Microsoft.PowerShell.Commands.FileSystemProvider.RemoveItem(String path, Boolean recurse)",
"TargetSite": {
"Attributes": 147,
"CallingConvention": 1,
"ContainsGenericParameters": false,
"CustomAttributes": "[System.Security.SecurityCriticalAttribute()]",
"DeclaringType": "System.IO.__Error, mscorlib, Version=4.0.0.0, Culture=neutral, PublicKeyToken=b77a5c561934e089",
"IsAbstract": false,
"IsAssembly": true,
"IsConstructor": false,
"IsFamily": false,
"IsFamilyAndAssembly": false,
"IsFamilyOrAssembly": false,
"IsFinal": false,
"IsGenericMethod": false,
"IsGenericMethodDefinition": false,
"IsHideBySig": true,
"IsPrivate": false,
"IsPublic": false,
"IsSecurityCritical": true,
"IsSecuritySafeCritical": false,
"IsSecurityTransparent": false,
"IsSpecialName": false,
"IsStatic": true,
"IsVirtual": false,
"MemberType": 8,
"MetadataToken": 100680987,
"MethodHandle": "System.RuntimeMethodHandle",
"MethodImplementationFlags": 0,
"Module": "CommonLanguageRuntimeLibrary",
"Name": "WinIOError",
"ReflectedType": "System.IO.__Error, mscorlib, Version=4.0.0.0, Culture=neutral, PublicKeyToken=b77a5c561934e089",
"ReturnParameter": "Void ",
"ReturnType": "void",
"ReturnTypeCustomAttributes": "Void "
}
},
"FullyQualifiedErrorId": "RemoveItemIOError,Microsoft.PowerShell.Commands.RemoveItemCommand",
"InvocationInfo": {
"BoundParameters": {},
"CommandOrigin": 1,
"DisplayScriptPosition": null,
"ExpectingInput": false,
"HistoryId": 1,
"InvocationName": "Remove-Item",
"Line": " Remove-Item -Recurse -Force $fileinfo\n",
"MyCommand": {
"CommandType": 8,
"DefaultParameterSet": "Path",
"Definition": "\r\nRemove-Item [-Path] <string[]> [-Filter <string>] [-Include <string[]>] [-Exclude <string[]>] [-Recurse] [-Force] [-Credential <pscredential>] [-WhatIf] [-Confirm] [-UseTransaction] [<CommonParameters>]\r\n\r\nRemove-Item -LiteralPath <string[]> [-Filter <string>] [-Include <string[]>] [-Exclude <string[]>] [-Recurse] [-Force] [-Credential <pscredential>] [-WhatIf] [-Confirm] [-UseTransaction] [<CommonParameters>]\r\n",
"HelpFile": "Microsoft.PowerShell.Commands.Management.dll-Help.xml",
"ImplementingType": "Microsoft.PowerShell.Commands.RemoveItemCommand",
"Module": "Microsoft.PowerShell.Management",
"ModuleName": "Microsoft.PowerShell.Management",
"Name": "Remove-Item",
"Noun": "Item",
"Options": 1,
"OutputType": "",
"PSSnapIn": null,
"ParameterSets": "[-Path] <string[]> [-Filter <string>] [-Include <string[]>] [-Exclude <string[]>] [-Recurse] [-Force] [-Credential <pscredential>] [-WhatIf] [-Confirm] [-UseTransaction] [<CommonParameters>] -LiteralPath <string[]> [-Filter <string>] [-Include <string[]>] [-Exclude <string[]>] [-Recurse] [-Force] [-Credential <pscredential>] [-WhatIf] [-Confirm] [-UseTransaction] [<CommonParameters>]",
"Parameters": "System.Collections.Generic.Dictionary`2[System.String,System.Management.Automation.ParameterMetadata]",
"RemotingCapability": 1,
"Verb": "Remove",
"Visibility": 0
},
"OffsetInLine": 9,
"PSCommandPath": "C:\\Users\\User\\AppData\\Local\\Temp\\ansible-tmp-1475364906.0-88571632788975\\win_file.ps1",
"PSScriptRoot": "C:\\Users\\User\\AppData\\Local\\Temp\\ansible-tmp-1475364906.0-88571632788975",
"PipelineLength": 0,
"PipelinePosition": 0,
"PositionMessage": "At C:\\Users\\User\\AppData\\Local\\Temp\\ansible-tmp-1475364906.0-88571632788975\\win_file.ps1:317 char:9\r\n+ Remove-Item -Recurse -Force $fileinfo\r\n+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~",
"ScriptLineNumber": 317,
"ScriptName": "C:\\Users\\User\\AppData\\Local\\Temp\\ansible-tmp-1475364906.0-88571632788975\\win_file.ps1",
"UnboundArguments": []
},
"PSMessageDetails": null,
"PipelineIterationInfo": [],
"ScriptStackTrace": "at <ScriptBlock>, C:\\Users\\User\\AppData\\Local\\Temp\\ansible-tmp-1475364906.0-88571632788975\\win_file.ps1: line 317\r\nat <ScriptBlock>, <No file>: line 4",
"TargetObject": "C:\\temp"
},
"failed": true,
"invocation": {
"module_name": "win_file"
},
"msg": "Could not find a part of the path 'C:\\temp\\link'."
}
```
|
main
|
win file failed to delete folder with broken symlink issue type bug report component name ansible modules core windows win file ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file etc ansible ansible cfg configured module search path os environment running bash for ubuntu on windows managing windows server summary when setting win file state absent for a directory that contains a symlink that is no longer pointing to a valid directory it fails with directorynotfoundexception steps to reproduce inventory ini host ansible user user ansible password password ansible connection winrm ansible port ansible winrm transport ntlm ansible winrm server cert validation ignore playbook yml name test windows hosts windows tasks name create test folder win file path c temp test state directory name create symlink raw cmd exe c mklink d c temp link c temp test name delete folder that link is pointing to win file path c temp test state absent name delete parent folder containing broken link win file path c temp state absent expected results expected c temp to be deleted correctly without exception actual results an exception occurred during task execution the full traceback is at c users user appdata local temp ansible tmp win file char remove item recurse force fileinfo fatal failed changed false error record categoryinfo activity remove item category reason directorynotfoundexception targetname c temp targettype string errordetails null exception data hresult helplink null innerexception null message could not find a part of the path c temp link source mscorlib stacktrace at system io error winioerror errorcode string maybefullpath r n at system io filesystemenumerableiterator commoninit r n at system io filesystemenumerableiterator ctor string path string originaluserpath string searchpattern searchoption searchoption searchresulthandler resulthandler boolean checkhost r n at system io filesystemenumerablefactory createdirectoryinfoiterator string path string originaluserpath string searchpattern searchoption searchoption r n at microsoft powershell commands filesystemprovider removedirectoryinfoitem directoryinfo directory boolean recurse boolean force boolean rootofremoval r n at microsoft powershell commands filesystemprovider removedirectoryinfoitem directoryinfo directory boolean recurse boolean force boolean rootofremoval r n at microsoft powershell commands filesystemprovider removeitem string path boolean recurse targetsite attributes callingconvention containsgenericparameters false customattributes declaringtype system io error mscorlib version culture neutral publickeytoken isabstract false isassembly true isconstructor false isfamily false isfamilyandassembly false isfamilyorassembly false isfinal false isgenericmethod false isgenericmethoddefinition false ishidebysig true isprivate false ispublic false issecuritycritical true issecuritysafecritical false issecuritytransparent false isspecialname false isstatic true isvirtual false membertype metadatatoken methodhandle system runtimemethodhandle methodimplementationflags module commonlanguageruntimelibrary name winioerror reflectedtype system io error mscorlib version culture neutral publickeytoken returnparameter void returntype void returntypecustomattributes void fullyqualifiederrorid removeitemioerror microsoft powershell commands removeitemcommand invocationinfo boundparameters commandorigin displayscriptposition null expectinginput false historyid invocationname remove item line remove item recurse force fileinfo n mycommand commandtype defaultparameterset path definition r nremove item r n r nremove item literalpath r n helpfile microsoft powershell commands management dll help xml implementingtype microsoft powershell commands removeitemcommand module microsoft powershell management modulename microsoft powershell management name remove item noun item options outputtype pssnapin null parametersets literalpath parameters system collections generic dictionary remotingcapability verb remove visibility offsetinline pscommandpath c users user appdata local temp ansible tmp win file psscriptroot c users user appdata local temp ansible tmp pipelinelength pipelineposition positionmessage at c users user appdata local temp ansible tmp win file char r n remove item recurse force fileinfo r n scriptlinenumber scriptname c users user appdata local temp ansible tmp win file unboundarguments psmessagedetails null pipelineiterationinfo scriptstacktrace at c users user appdata local temp ansible tmp win file line r nat line targetobject c temp failed true invocation module name win file msg could not find a part of the path c temp link
| 1
|
101,478
| 4,118,629,413
|
IssuesEvent
|
2016-06-08 12:18:19
|
isawnyu/isaw.web
|
https://api.github.com/repos/isawnyu/isaw.web
|
closed
|
checklist view template limited to 100 items
|
deploy high priority style
|
Question; what is causing the limitation of checklist view to show only the first 100 items in the folder? Is it something that we can change?
|
1.0
|
checklist view template limited to 100 items - Question; what is causing the limitation of checklist view to show only the first 100 items in the folder? Is it something that we can change?
|
non_main
|
checklist view template limited to items question what is causing the limitation of checklist view to show only the first items in the folder is it something that we can change
| 0
|
3,586
| 14,450,254,974
|
IssuesEvent
|
2020-12-08 09:20:15
|
IPVS-AS/MBP
|
https://api.github.com/repos/IPVS-AS/MBP
|
closed
|
Clean up index.html
|
maintainance
|
The index.html template includes a lot of libraries/javascript files/css files which are not all needed by the project. It would be helpful to clean this file up.
|
True
|
Clean up index.html - The index.html template includes a lot of libraries/javascript files/css files which are not all needed by the project. It would be helpful to clean this file up.
|
main
|
clean up index html the index html template includes a lot of libraries javascript files css files which are not all needed by the project it would be helpful to clean this file up
| 1
|
3,887
| 17,242,674,711
|
IssuesEvent
|
2021-07-21 02:25:56
|
LSSTDESC/BlendingToolKit
|
https://api.github.com/repos/LSSTDESC/BlendingToolKit
|
opened
|
poetry dynamical versioning
|
bug maintainer
|
I need to investigate more how to use this package, currently disabled since not sure how to use it and prevents me from bumping versions sometimes.
|
True
|
poetry dynamical versioning - I need to investigate more how to use this package, currently disabled since not sure how to use it and prevents me from bumping versions sometimes.
|
main
|
poetry dynamical versioning i need to investigate more how to use this package currently disabled since not sure how to use it and prevents me from bumping versions sometimes
| 1
|
416,788
| 28,100,442,797
|
IssuesEvent
|
2023-03-30 19:02:19
|
openpmix/openpmix
|
https://api.github.com/repos/openpmix/openpmix
|
closed
|
Document that OpenPMIx requires C99-style assignment of server module fns
|
documentation
|
Required as we cannot otherwise tell the difference between versions of the struct.
|
1.0
|
Document that OpenPMIx requires C99-style assignment of server module fns - Required as we cannot otherwise tell the difference between versions of the struct.
|
non_main
|
document that openpmix requires style assignment of server module fns required as we cannot otherwise tell the difference between versions of the struct
| 0
|
522,717
| 15,166,583,057
|
IssuesEvent
|
2021-02-12 16:33:54
|
ArctosDB/arctos
|
https://api.github.com/repos/ArctosDB/arctos
|
closed
|
control agent name type login
|
Enhancement Priority-High
|
don't allow INSERT or UPDATE unless
* name is arctos username
* requirements in https://handbook.arctosdb.org/documentation/users.html are met
|
1.0
|
control agent name type login - don't allow INSERT or UPDATE unless
* name is arctos username
* requirements in https://handbook.arctosdb.org/documentation/users.html are met
|
non_main
|
control agent name type login don t allow insert or update unless name is arctos username requirements in are met
| 0
|
145,552
| 22,709,366,256
|
IssuesEvent
|
2022-07-05 17:37:51
|
CarnegieLearningWeb/UpGrade
|
https://api.github.com/repos/CarnegieLearningWeb/UpGrade
|
closed
|
The defined metrics table's "Description" should be renamed to "Display Name"
|
feature design
|
**Describe the feature**
We've discussed this in the meeting and we think the defined metrics table's "Description" should be renamed to "Display Name" because this field is required unlike other description fields and it will be used as a display name on the overview page (Data tab in the latest design).
**To Reproduce**
Steps to reproduce the behavior:
1. Go to the Experiment Stepper's Metrics step and find the "Description" label and the placeholder.
**Expected behavior**
The defined metrics table's "Description" label and the placeholder should be renamed to "Display Name".
**Screenshots**
* Current design:
<img width="500" alt="Screen Shot 2022-07-01 at 11 39 34 AM" src="https://user-images.githubusercontent.com/90279765/176926493-e3f00fc2-2c46-46fc-911b-a9abac16c6f9.png">
* Expected design:
<img width="500" alt="Experiments - Add Experiment-04-01" src="https://user-images.githubusercontent.com/90279765/176926667-453b5c57-2828-4b31-9cd0-a7804ae35827.png">
|
1.0
|
The defined metrics table's "Description" should be renamed to "Display Name" - **Describe the feature**
We've discussed this in the meeting and we think the defined metrics table's "Description" should be renamed to "Display Name" because this field is required unlike other description fields and it will be used as a display name on the overview page (Data tab in the latest design).
**To Reproduce**
Steps to reproduce the behavior:
1. Go to the Experiment Stepper's Metrics step and find the "Description" label and the placeholder.
**Expected behavior**
The defined metrics table's "Description" label and the placeholder should be renamed to "Display Name".
**Screenshots**
* Current design:
<img width="500" alt="Screen Shot 2022-07-01 at 11 39 34 AM" src="https://user-images.githubusercontent.com/90279765/176926493-e3f00fc2-2c46-46fc-911b-a9abac16c6f9.png">
* Expected design:
<img width="500" alt="Experiments - Add Experiment-04-01" src="https://user-images.githubusercontent.com/90279765/176926667-453b5c57-2828-4b31-9cd0-a7804ae35827.png">
|
non_main
|
the defined metrics table s description should be renamed to display name describe the feature we ve discussed this in the meeting and we think the defined metrics table s description should be renamed to display name because this field is required unlike other description fields and it will be used as a display name on the overview page data tab in the latest design to reproduce steps to reproduce the behavior go to the experiment stepper s metrics step and find the description label and the placeholder expected behavior the defined metrics table s description label and the placeholder should be renamed to display name screenshots current design img width alt screen shot at am src expected design img width alt experiments add experiment src
| 0
|
238,065
| 26,085,545,126
|
IssuesEvent
|
2022-12-26 01:58:58
|
kapseliboi/bui
|
https://api.github.com/repos/kapseliboi/bui
|
opened
|
CVE-2022-46175 (High) detected in json5-2.2.0.tgz, json5-1.0.1.tgz
|
security vulnerability
|
## CVE-2022-46175 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>json5-2.2.0.tgz</b>, <b>json5-1.0.1.tgz</b></p></summary>
<p>
<details><summary><b>json5-2.2.0.tgz</b></p></summary>
<p>JSON for humans.</p>
<p>Library home page: <a href="https://registry.npmjs.org/json5/-/json5-2.2.0.tgz">https://registry.npmjs.org/json5/-/json5-2.2.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@babel/core/node_modules/json5/package.json</p>
<p>
Dependency Hierarchy:
- parcel-bundler-1.12.5.tgz (Root Library)
- core-7.14.3.tgz
- :x: **json5-2.2.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>json5-1.0.1.tgz</b></p></summary>
<p>JSON for humans.</p>
<p>Library home page: <a href="https://registry.npmjs.org/json5/-/json5-1.0.1.tgz">https://registry.npmjs.org/json5/-/json5-1.0.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/json5/package.json</p>
<p>
Dependency Hierarchy:
- parcel-bundler-1.12.5.tgz (Root Library)
- :x: **json5-1.0.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/bui/commit/e8304e6335e5d45f5599a6dd9950348f734192b7">e8304e6335e5d45f5599a6dd9950348f734192b7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
JSON5 is an extension to the popular JSON file format that aims to be easier to write and maintain by hand (e.g. for config files). The `parse` method of the JSON5 library before and including version `2.2.1` does not restrict parsing of keys named `__proto__`, allowing specially crafted strings to pollute the prototype of the resulting object. This vulnerability pollutes the prototype of the object returned by `JSON5.parse` and not the global Object prototype, which is the commonly understood definition of Prototype Pollution. However, polluting the prototype of a single object can have significant security impact for an application if the object is later used in trusted operations. This vulnerability could allow an attacker to set arbitrary and unexpected keys on the object returned from `JSON5.parse`. The actual impact will depend on how applications utilize the returned object and how they filter unwanted keys, but could include denial of service, cross-site scripting, elevation of privilege, and in extreme cases, remote code execution. `JSON5.parse` should restrict parsing of `__proto__` keys when parsing JSON strings to objects. As a point of reference, the `JSON.parse` method included in JavaScript ignores `__proto__` keys. Simply changing `JSON5.parse` to `JSON.parse` in the examples above mitigates this vulnerability. This vulnerability is patched in json5 version 2.2.2 and later.
<p>Publish Date: 2022-12-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-46175>CVE-2022-46175</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-46175">https://www.cve.org/CVERecord?id=CVE-2022-46175</a></p>
<p>Release Date: 2022-12-24</p>
<p>Fix Resolution: json5 - 2.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-46175 (High) detected in json5-2.2.0.tgz, json5-1.0.1.tgz - ## CVE-2022-46175 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>json5-2.2.0.tgz</b>, <b>json5-1.0.1.tgz</b></p></summary>
<p>
<details><summary><b>json5-2.2.0.tgz</b></p></summary>
<p>JSON for humans.</p>
<p>Library home page: <a href="https://registry.npmjs.org/json5/-/json5-2.2.0.tgz">https://registry.npmjs.org/json5/-/json5-2.2.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@babel/core/node_modules/json5/package.json</p>
<p>
Dependency Hierarchy:
- parcel-bundler-1.12.5.tgz (Root Library)
- core-7.14.3.tgz
- :x: **json5-2.2.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>json5-1.0.1.tgz</b></p></summary>
<p>JSON for humans.</p>
<p>Library home page: <a href="https://registry.npmjs.org/json5/-/json5-1.0.1.tgz">https://registry.npmjs.org/json5/-/json5-1.0.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/json5/package.json</p>
<p>
Dependency Hierarchy:
- parcel-bundler-1.12.5.tgz (Root Library)
- :x: **json5-1.0.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/bui/commit/e8304e6335e5d45f5599a6dd9950348f734192b7">e8304e6335e5d45f5599a6dd9950348f734192b7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
JSON5 is an extension to the popular JSON file format that aims to be easier to write and maintain by hand (e.g. for config files). The `parse` method of the JSON5 library before and including version `2.2.1` does not restrict parsing of keys named `__proto__`, allowing specially crafted strings to pollute the prototype of the resulting object. This vulnerability pollutes the prototype of the object returned by `JSON5.parse` and not the global Object prototype, which is the commonly understood definition of Prototype Pollution. However, polluting the prototype of a single object can have significant security impact for an application if the object is later used in trusted operations. This vulnerability could allow an attacker to set arbitrary and unexpected keys on the object returned from `JSON5.parse`. The actual impact will depend on how applications utilize the returned object and how they filter unwanted keys, but could include denial of service, cross-site scripting, elevation of privilege, and in extreme cases, remote code execution. `JSON5.parse` should restrict parsing of `__proto__` keys when parsing JSON strings to objects. As a point of reference, the `JSON.parse` method included in JavaScript ignores `__proto__` keys. Simply changing `JSON5.parse` to `JSON.parse` in the examples above mitigates this vulnerability. This vulnerability is patched in json5 version 2.2.2 and later.
<p>Publish Date: 2022-12-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-46175>CVE-2022-46175</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-46175">https://www.cve.org/CVERecord?id=CVE-2022-46175</a></p>
<p>Release Date: 2022-12-24</p>
<p>Fix Resolution: json5 - 2.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_main
|
cve high detected in tgz tgz cve high severity vulnerability vulnerable libraries tgz tgz tgz json for humans library home page a href path to dependency file package json path to vulnerable library node modules babel core node modules package json dependency hierarchy parcel bundler tgz root library core tgz x tgz vulnerable library tgz json for humans library home page a href path to dependency file package json path to vulnerable library node modules package json dependency hierarchy parcel bundler tgz root library x tgz vulnerable library found in head commit a href found in base branch master vulnerability details is an extension to the popular json file format that aims to be easier to write and maintain by hand e g for config files the parse method of the library before and including version does not restrict parsing of keys named proto allowing specially crafted strings to pollute the prototype of the resulting object this vulnerability pollutes the prototype of the object returned by parse and not the global object prototype which is the commonly understood definition of prototype pollution however polluting the prototype of a single object can have significant security impact for an application if the object is later used in trusted operations this vulnerability could allow an attacker to set arbitrary and unexpected keys on the object returned from parse the actual impact will depend on how applications utilize the returned object and how they filter unwanted keys but could include denial of service cross site scripting elevation of privilege and in extreme cases remote code execution parse should restrict parsing of proto keys when parsing json strings to objects as a point of reference the json parse method included in javascript ignores proto keys simply changing parse to json parse in the examples above mitigates this vulnerability this vulnerability is patched in version and later publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
99,603
| 11,149,658,284
|
IssuesEvent
|
2019-12-23 19:30:17
|
baielvegas/baielvegas.github.io
|
https://api.github.com/repos/baielvegas/baielvegas.github.io
|
closed
|
Documentation
|
documentation
|
@mammamoo please
1.сreate a glossary(dictionary) of terms: give the client and the team a clear understanding of the meaning of words in the context of the website or SRS.
2.describe the website - Tech and Science Fest: give the client and the team an idea of the site
3.describe the role system: show the client and team what roles the users have and in what scenario they can use the site
4.write functional requirements: show the team how the product should work and what features or characteristics it should have
|
1.0
|
Documentation - @mammamoo please
1.сreate a glossary(dictionary) of terms: give the client and the team a clear understanding of the meaning of words in the context of the website or SRS.
2.describe the website - Tech and Science Fest: give the client and the team an idea of the site
3.describe the role system: show the client and team what roles the users have and in what scenario they can use the site
4.write functional requirements: show the team how the product should work and what features or characteristics it should have
|
non_main
|
documentation mammamoo please сreate a glossary dictionary of terms give the client and the team a clear understanding of the meaning of words in the context of the website or srs describe the website tech and science fest give the client and the team an idea of the site describe the role system show the client and team what roles the users have and in what scenario they can use the site write functional requirements show the team how the product should work and what features or characteristics it should have
| 0
|
241,011
| 7,808,056,164
|
IssuesEvent
|
2018-06-11 18:53:48
|
cwrc/ontology
|
https://api.github.com/repos/cwrc/ontology
|
closed
|
Failure to parse xml of genre.rdf
|
priority:immediate project:Genre Ontology status:help wanted type:bug
|
Output from xmllint: rapper: Parsing URI genre.rdf with parser rdfxml
rapper: Error - URI genre.rdf - Resolving URI failed: Could not resolve host: genre.rdf
rapper: Failed to parse URI genre.rdf rdfxml content
rapper: Parsing returned 0 triples
, Output from rapper: warning: failed to load external entity 'genre.rdf'
|
1.0
|
Failure to parse xml of genre.rdf - Output from xmllint: rapper: Parsing URI genre.rdf with parser rdfxml
rapper: Error - URI genre.rdf - Resolving URI failed: Could not resolve host: genre.rdf
rapper: Failed to parse URI genre.rdf rdfxml content
rapper: Parsing returned 0 triples
, Output from rapper: warning: failed to load external entity 'genre.rdf'
|
non_main
|
failure to parse xml of genre rdf output from xmllint rapper parsing uri genre rdf with parser rdfxml rapper error uri genre rdf resolving uri failed could not resolve host genre rdf rapper failed to parse uri genre rdf rdfxml content rapper parsing returned triples output from rapper warning failed to load external entity genre rdf
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.