Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16 values | text_combine stringlengths 96 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
775,185 | 27,222,355,448 | IssuesEvent | 2023-02-21 06:55:14 | CodeforHawaii/HIERR | https://api.github.com/repos/CodeforHawaii/HIERR | opened | Add leaflet address widget | Medium Priority | It can't be assumed that the user knows their zip code.
This would add in a leaflet address widget that would zoom in to the location based on the user's street and house number | 1.0 | Add leaflet address widget - It can't be assumed that the user knows their zip code.
This would add in a leaflet address widget that would zoom in to the location based on the user's street and house number | priority | add leaflet address widget it can t be assumed that the user knows their zip code this would add in a leaflet address widget that would zoom in to the location based on the user s street and house number | 1 |
645,776 | 21,015,005,394 | IssuesEvent | 2022-03-30 10:09:52 | micronaut-projects/micronaut-core | https://api.github.com/repos/micronaut-projects/micronaut-core | closed | Native image builds do not respect Jackson directive in application.yml | type: improvement priority: low | ### Expected Behavior
In a java app, with a src/main/resources/application,yml of
`micronaut:
application:
name: mntest2
jackson:
property-naming-strategy: SNAKE_CASE`
the expectation is that the properties of any POJO are transposed to snake case.
`➜ mntest1 curl http://localhost:8080/v1/things/fred
{"thing_id":"fred","property_one":"hatstand"}% `
### Actual Behaviour
When the app is built with ./mvnw package -Dpackaging=native-image, we see this behaviour
`➜ mntest1 curl http://localhost:8080/v1/things/fred
{"thingId":"fred","propertyOne":"hatstand"}% `
### Steps To Reproduce
Clone the app at https://github.com/alanwhite/mntest1
Build & run with ./mvnw mn:run
Once ready run curl http://localhost:8080/v1/things/fred
and witness correct response of
`{"thing_id":"fred","property_one":"hatstand"}% `
Cancel the running app and rebuild using ./mvnw package -Dpackaging=native-image
Once built launch with ./target/mntest1 & and then run again curl http://localhost:8080/v1/things/fred
and witness incorrect response of
`{"thingId":"fred","propertyOne":"hatstand"}% `
### Environment Information
Linux ubuntu 5.11.0-41-generic #45~20.04.1-Ubuntu SMP Wed Nov 10 10:20:10 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux
java: 21.3.0.r17-grl
micronaut: 3.2.3
### Example Application
https://github.com/alanwhite/mntest1
### Version
3.2.3 | 1.0 | Native image builds do not respect Jackson directive in application.yml - ### Expected Behavior
In a java app, with a src/main/resources/application,yml of
`micronaut:
application:
name: mntest2
jackson:
property-naming-strategy: SNAKE_CASE`
the expectation is that the properties of any POJO are transposed to snake case.
`➜ mntest1 curl http://localhost:8080/v1/things/fred
{"thing_id":"fred","property_one":"hatstand"}% `
### Actual Behaviour
When the app is built with ./mvnw package -Dpackaging=native-image, we see this behaviour
`➜ mntest1 curl http://localhost:8080/v1/things/fred
{"thingId":"fred","propertyOne":"hatstand"}% `
### Steps To Reproduce
Clone the app at https://github.com/alanwhite/mntest1
Build & run with ./mvnw mn:run
Once ready run curl http://localhost:8080/v1/things/fred
and witness correct response of
`{"thing_id":"fred","property_one":"hatstand"}% `
Cancel the running app and rebuild using ./mvnw package -Dpackaging=native-image
Once built launch with ./target/mntest1 & and then run again curl http://localhost:8080/v1/things/fred
and witness incorrect response of
`{"thingId":"fred","propertyOne":"hatstand"}% `
### Environment Information
Linux ubuntu 5.11.0-41-generic #45~20.04.1-Ubuntu SMP Wed Nov 10 10:20:10 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux
java: 21.3.0.r17-grl
micronaut: 3.2.3
### Example Application
https://github.com/alanwhite/mntest1
### Version
3.2.3 | priority | native image builds do not respect jackson directive in application yml expected behavior in a java app with a src main resources application yml of micronaut application name jackson property naming strategy snake case the expectation is that the properties of any pojo are transposed to snake case ➜ curl thing id fred property one hatstand actual behaviour when the app is built with mvnw package dpackaging native image we see this behaviour ➜ curl thingid fred propertyone hatstand steps to reproduce clone the app at build run with mvnw mn run once ready run curl and witness correct response of thing id fred property one hatstand cancel the running app and rebuild using mvnw package dpackaging native image once built launch with target and then run again curl and witness incorrect response of thingid fred propertyone hatstand environment information linux ubuntu generic ubuntu smp wed nov utc gnu linux java grl micronaut example application version | 1 |
732,374 | 25,257,388,371 | IssuesEvent | 2022-11-15 19:20:26 | dtcenter/METplotpy | https://api.github.com/repos/dtcenter/METplotpy | closed | Create the METplotpy v2.0.0-beta5 release | type: task priority: high alert: NEED ACCOUNT KEY METplotpy: General |
## Describe the Task ##
Create the beta5 release following these instructions:
https://metplus.readthedocs.io/en/develop/Release_Guide/metplotpy_development.html
### Time Estimate ###
*Estimate the amount of work required here.*
*Issues should represent approximately 1 to 3 days of work.*
### Sub-Issues ###
Consider breaking the task down into sub-issues.
- [ ] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [ ] Select **component(s)**
- [ ] Select **priority**
- [ ] Select **requestor(s)**
### Projects and Milestone ###
- [ ] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label
- [ ] Select **Milestone** as the next official version or **Future Versions**
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Task Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| 1.0 | Create the METplotpy v2.0.0-beta5 release -
## Describe the Task ##
Create the beta5 release following these instructions:
https://metplus.readthedocs.io/en/develop/Release_Guide/metplotpy_development.html
### Time Estimate ###
*Estimate the amount of work required here.*
*Issues should represent approximately 1 to 3 days of work.*
### Sub-Issues ###
Consider breaking the task down into sub-issues.
- [ ] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [ ] Select **component(s)**
- [ ] Select **priority**
- [ ] Select **requestor(s)**
### Projects and Milestone ###
- [ ] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label
- [ ] Select **Milestone** as the next official version or **Future Versions**
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Task Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| priority | create the metplotpy release describe the task create the release following these instructions time estimate estimate the amount of work required here issues should represent approximately to days of work sub issues consider breaking the task down into sub issues add a checkbox for each sub issue here relevant deadlines list relevant project deadlines here or state none funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select repository and or organization level project s or add alert need project assignment label select milestone as the next official version or future versions define related issue s consider the impact to the other metplus components task checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue | 1 |
54,480 | 13,912,020,375 | IssuesEvent | 2020-10-20 18:14:25 | jgeraigery/LocalCatalogManager | https://api.github.com/repos/jgeraigery/LocalCatalogManager | opened | CVE-2019-16942 (High) detected in jackson-databind-2.8.5.jar | security vulnerability | ## CVE-2019-16942 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: LocalCatalogManager/lcm-server/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.5/jackson-databind-2.8.5.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.5/jackson-databind-2.8.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.5/jackson-databind-2.8.5.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.5/jackson-databind-2.8.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/LocalCatalogManager/commit/b8c24e199f2d440dea3ce3cc2c66ada102d5d922">b8c24e199f2d440dea3ce3cc2c66ada102d5d922</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the commons-dbcp (1.4) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of org.apache.commons.dbcp.datasources.SharedPoolDataSource and org.apache.commons.dbcp.datasources.PerUserPoolDataSource mishandling.
<p>Publish Date: 2019-10-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16942>CVE-2019-16942</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942</a></p>
<p>Release Date: 2019-10-01</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.5","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1"}],"vulnerabilityIdentifier":"CVE-2019-16942","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the commons-dbcp (1.4) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of org.apache.commons.dbcp.datasources.SharedPoolDataSource and org.apache.commons.dbcp.datasources.PerUserPoolDataSource mishandling.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16942","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-16942 (High) detected in jackson-databind-2.8.5.jar - ## CVE-2019-16942 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: LocalCatalogManager/lcm-server/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.5/jackson-databind-2.8.5.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.5/jackson-databind-2.8.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.5/jackson-databind-2.8.5.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.5/jackson-databind-2.8.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/LocalCatalogManager/commit/b8c24e199f2d440dea3ce3cc2c66ada102d5d922">b8c24e199f2d440dea3ce3cc2c66ada102d5d922</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the commons-dbcp (1.4) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of org.apache.commons.dbcp.datasources.SharedPoolDataSource and org.apache.commons.dbcp.datasources.PerUserPoolDataSource mishandling.
<p>Publish Date: 2019-10-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16942>CVE-2019-16942</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942</a></p>
<p>Release Date: 2019-10-01</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.5","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1"}],"vulnerabilityIdentifier":"CVE-2019-16942","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the commons-dbcp (1.4) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of org.apache.commons.dbcp.datasources.SharedPoolDataSource and org.apache.commons.dbcp.datasources.PerUserPoolDataSource mishandling.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16942","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file localcatalogmanager lcm server pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the commons dbcp jar in the classpath and an attacker can find an rmi service endpoint to access it is possible to make the service execute a malicious payload this issue exists because of org apache commons dbcp datasources sharedpooldatasource and org apache commons dbcp datasources peruserpooldatasource mishandling publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the commons dbcp jar in the classpath and an attacker can find an rmi service endpoint to access it is possible to make the service execute a malicious payload this issue exists because of org apache commons dbcp datasources sharedpooldatasource and org apache commons dbcp datasources peruserpooldatasource mishandling vulnerabilityurl | 0 |
37,875 | 5,146,880,217 | IssuesEvent | 2017-01-13 03:39:30 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Test failure: ReadAndWrite/OutputEncoding fail with "Assert+WrapperXunitException" | area-System.Console test bug test-run-desktop | Opened on behalf of @Jiayili1
The test `ReadAndWrite/OutputEncoding` has failed.
```
Assert+WrapperXunitException : File path: D:\A\_work\2\s\src\System.Console\tests\ReadAndWrite.cs. Line: 202\r
---- Assert.Equal() Failure\r
Expected: Byte[] []\r
Actual: Byte[] [239, 187, 191]
```
Stack Trace:
```
at Assert.WrapException(Exception e, String callerFilePath, Int32 callerLineNumber) in D:\A\_work\2\s\src\Common\tests\System\Diagnostics\AssertWithCallerAttributes.cs:line 583
at Assert.Equal[T](T expected, T actual, String path, Int32 line) in D:\A\_work\2\s\src\Common\tests\System\Diagnostics\AssertWithCallerAttributes.cs:line 172
at ReadAndWrite.ValidateConsoleEncoding(Encoding encoding) in D:\A\_work\2\s\src\System.Console\tests\ReadAndWrite.cs:line 202
at ReadAndWrite.OutputEncoding() in D:\A\_work\2\s\src\System.Console\tests\ReadAndWrite.cs:line 275
----- Inner Stack Trace -----
at Assert.Equal[T](T expected, T actual, String path, Int32 line) in D:\A\_work\2\s\src\Common\tests\System\Diagnostics\AssertWithCallerAttributes.cs:line 171
```
Failing configurations:
- Windows.10.Amd64
- AnyCPU-Debug
- AnyCPU-Release
link: https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster/type/test~2Ffunctional~2Fdesktop~2Fcli~2F/build/20160816.02/workItem/System.Console.Tests/analysis/xunit/ReadAndWrite~2FOutputEncoding
| 2.0 | Test failure: ReadAndWrite/OutputEncoding fail with "Assert+WrapperXunitException" - Opened on behalf of @Jiayili1
The test `ReadAndWrite/OutputEncoding` has failed.
```
Assert+WrapperXunitException : File path: D:\A\_work\2\s\src\System.Console\tests\ReadAndWrite.cs. Line: 202\r
---- Assert.Equal() Failure\r
Expected: Byte[] []\r
Actual: Byte[] [239, 187, 191]
```
Stack Trace:
```
at Assert.WrapException(Exception e, String callerFilePath, Int32 callerLineNumber) in D:\A\_work\2\s\src\Common\tests\System\Diagnostics\AssertWithCallerAttributes.cs:line 583
at Assert.Equal[T](T expected, T actual, String path, Int32 line) in D:\A\_work\2\s\src\Common\tests\System\Diagnostics\AssertWithCallerAttributes.cs:line 172
at ReadAndWrite.ValidateConsoleEncoding(Encoding encoding) in D:\A\_work\2\s\src\System.Console\tests\ReadAndWrite.cs:line 202
at ReadAndWrite.OutputEncoding() in D:\A\_work\2\s\src\System.Console\tests\ReadAndWrite.cs:line 275
----- Inner Stack Trace -----
at Assert.Equal[T](T expected, T actual, String path, Int32 line) in D:\A\_work\2\s\src\Common\tests\System\Diagnostics\AssertWithCallerAttributes.cs:line 171
```
Failing configurations:
- Windows.10.Amd64
- AnyCPU-Debug
- AnyCPU-Release
link: https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster/type/test~2Ffunctional~2Fdesktop~2Fcli~2F/build/20160816.02/workItem/System.Console.Tests/analysis/xunit/ReadAndWrite~2FOutputEncoding
| non_priority | test failure readandwrite outputencoding fail with assert wrapperxunitexception opened on behalf of the test readandwrite outputencoding has failed assert wrapperxunitexception file path d a work s src system console tests readandwrite cs line r assert equal failure r expected byte r actual byte stack trace at assert wrapexception exception e string callerfilepath callerlinenumber in d a work s src common tests system diagnostics assertwithcallerattributes cs line at assert equal t expected t actual string path line in d a work s src common tests system diagnostics assertwithcallerattributes cs line at readandwrite validateconsoleencoding encoding encoding in d a work s src system console tests readandwrite cs line at readandwrite outputencoding in d a work s src system console tests readandwrite cs line inner stack trace at assert equal t expected t actual string path line in d a work s src common tests system diagnostics assertwithcallerattributes cs line failing configurations windows anycpu debug anycpu release link | 0 |
172,805 | 27,331,237,448 | IssuesEvent | 2023-02-25 17:00:41 | SpenceKonde/megaTinyCore | https://api.github.com/repos/SpenceKonde/megaTinyCore | closed | Suggestion: provide an alternative boards-hv.txt file | enhancement design decision noplans | Currently you have to locate and uncomment some 20 lines in the **boards.txt** file to activate the **UPDI/Reset Pin Function** menu option on the **Tools** menu.
As HV programmers are becoming available, more people will need these options. How about including an alternative version of the **boards.txt** file, called boards-hv.txt, which has the lines already commented out? Someone then just has to rename **boards.txt** to **boards-old.txt**, and **boards-hv.txt** to **boards.txt**, with no risk of accidentally messing something up.
Also, how about making DxCore consistent, with the HV options commented out by default? | 1.0 | Suggestion: provide an alternative boards-hv.txt file - Currently you have to locate and uncomment some 20 lines in the **boards.txt** file to activate the **UPDI/Reset Pin Function** menu option on the **Tools** menu.
As HV programmers are becoming available, more people will need these options. How about including an alternative version of the **boards.txt** file, called boards-hv.txt, which has the lines already commented out? Someone then just has to rename **boards.txt** to **boards-old.txt**, and **boards-hv.txt** to **boards.txt**, with no risk of accidentally messing something up.
Also, how about making DxCore consistent, with the HV options commented out by default? | non_priority | suggestion provide an alternative boards hv txt file currently you have to locate and uncomment some lines in the boards txt file to activate the updi reset pin function menu option on the tools menu as hv programmers are becoming available more people will need these options how about including an alternative version of the boards txt file called boards hv txt which has the lines already commented out someone then just has to rename boards txt to boards old txt and boards hv txt to boards txt with no risk of accidentally messing something up also how about making dxcore consistent with the hv options commented out by default | 0 |
254,447 | 8,074,130,096 | IssuesEvent | 2018-08-06 21:47:54 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio-ui] options on right click options of wcm-assets-folder dont work when using multi-path configuration | CI bug priority: high | ### Expected behavior
right click options should function correctly
### Actual behavior
selecting a right click menu option loads the preview with a 404 (undefined)
### Steps to reproduce the problem
* set up content so that you have multiple roots to use
```
/static-assets/images
/static-assets/images/folder-a
/static-assets/images/folder-a/my-image.jpg
/static-assets/images/folder-b
/static-assets/images/folder-b/your-image.jpg
```
* configure the static assets plugin to use the multiple paths in the side bar config
```
<!-- Static Assets -->
<modulehook>
<name>wcm-assets-folder</name>
<params>
<label>Static Assets</label>
<path>/static-assets/images/folder-a</path>
<path>/static-assets/images/folder-b</path>
<showRootItem>true</showRootItem>
<onClick>none</onClick>
</params>
</modulehook>
```
* Open the folders, right click on one of the images and choose an option (cut, history etc), note 404 in preview.
### Log/stack trace (use https://gist.github.com)
N/A
### Specs
#### Version
3.0.13 - 3.0.15
Replicated in:
Studio Version Number: 3.0.15-SNAPSHOT-d62cff
Build Number: d62cff05198955c0ccc924267702b62e94b14dc4
Build Date/Time: 07-17-2018 09:49:21 -0400
#### OS
Any
#### Browser
Any | 1.0 | [studio-ui] options on right click options of wcm-assets-folder dont work when using multi-path configuration - ### Expected behavior
right click options should function correctly
### Actual behavior
selecting a right click menu option loads the preview with a 404 (undefined)
### Steps to reproduce the problem
* set up content so that you have multiple roots to use
```
/static-assets/images
/static-assets/images/folder-a
/static-assets/images/folder-a/my-image.jpg
/static-assets/images/folder-b
/static-assets/images/folder-b/your-image.jpg
```
* configure the static assets plugin to use the multiple paths in the side bar config
```
<!-- Static Assets -->
<modulehook>
<name>wcm-assets-folder</name>
<params>
<label>Static Assets</label>
<path>/static-assets/images/folder-a</path>
<path>/static-assets/images/folder-b</path>
<showRootItem>true</showRootItem>
<onClick>none</onClick>
</params>
</modulehook>
```
* Open the folders, right click on one of the images and choose an option (cut, history etc), note 404 in preview.
### Log/stack trace (use https://gist.github.com)
N/A
### Specs
#### Version
3.0.13 - 3.0.15
Replicated in:
Studio Version Number: 3.0.15-SNAPSHOT-d62cff
Build Number: d62cff05198955c0ccc924267702b62e94b14dc4
Build Date/Time: 07-17-2018 09:49:21 -0400
#### OS
Any
#### Browser
Any | priority | options on right click options of wcm assets folder dont work when using multi path configuration expected behavior right click options should function correctly actual behavior selecting a right click menu option loads the preview with a undefined steps to reproduce the problem set up content so that you have multiple roots to use static assets images static assets images folder a static assets images folder a my image jpg static assets images folder b static assets images folder b your image jpg configure the static assets plugin to use the multiple paths in the side bar config wcm assets folder static assets static assets images folder a static assets images folder b true none open the folders right click on one of the images and choose an option cut history etc note in preview log stack trace use n a specs version replicated in studio version number snapshot build number build date time os any browser any | 1 |
552,661 | 16,246,321,870 | IssuesEvent | 2021-05-07 14:59:25 | ethereum/sourcify | https://api.github.com/repos/ethereum/sourcify | closed | Skip storing metadata under IPFS hash if not available | medium priority | In verification, after all sources and metadata are stored, the metadata additionally gets stored under the its IPFS hash. If this hash is not available, error is reported. As this additionally stored metadata is not used anywhere, this step can be skipped in this case, only perhaps logging a message on the server. | 1.0 | Skip storing metadata under IPFS hash if not available - In verification, after all sources and metadata are stored, the metadata additionally gets stored under the its IPFS hash. If this hash is not available, error is reported. As this additionally stored metadata is not used anywhere, this step can be skipped in this case, only perhaps logging a message on the server. | priority | skip storing metadata under ipfs hash if not available in verification after all sources and metadata are stored the metadata additionally gets stored under the its ipfs hash if this hash is not available error is reported as this additionally stored metadata is not used anywhere this step can be skipped in this case only perhaps logging a message on the server | 1 |
435,848 | 30,522,733,528 | IssuesEvent | 2023-07-19 09:10:06 | dana-team/hns-nqs-plugin | https://api.github.com/repos/dana-team/hns-nqs-plugin | closed | Add documentation for API | kind/documentation | The API needs to be documented and each field in the API needs to be briefly explained so it can be understood by outside readers
https://github.com/dana-team/hns-nodeQuotaSync-plugin/blob/46ca5f5b968b1d8bf34ce7b63d38aac77662b6a1/api/v1alpha1/nodequotaconfig_types.go#L28-L55
/kind documentation | 1.0 | Add documentation for API - The API needs to be documented and each field in the API needs to be briefly explained so it can be understood by outside readers
https://github.com/dana-team/hns-nodeQuotaSync-plugin/blob/46ca5f5b968b1d8bf34ce7b63d38aac77662b6a1/api/v1alpha1/nodequotaconfig_types.go#L28-L55
/kind documentation | non_priority | add documentation for api the api needs to be documented and each field in the api needs to be briefly explained so it can be understood by outside readers kind documentation | 0 |
472,405 | 13,623,855,640 | IssuesEvent | 2020-09-24 07:06:47 | incognitochain/incognito-chain | https://api.github.com/repos/incognitochain/incognito-chain | closed | [Test][Testnet] branch master-temp-B-deploy-consensus-v2-optimized | Priority: High | - [x] Transaction
-- Privacy V1
-- No-privacy
-- Batching verify
-- Withdraw reward
- [x] Brigde
-- Centralize
-- Decentralize
- [ ] Portal V2
- [ ] Staking
- [ ] pDex
| 1.0 | [Test][Testnet] branch master-temp-B-deploy-consensus-v2-optimized - - [x] Transaction
-- Privacy V1
-- No-privacy
-- Batching verify
-- Withdraw reward
- [x] Brigde
-- Centralize
-- Decentralize
- [ ] Portal V2
- [ ] Staking
- [ ] pDex
| priority | branch master temp b deploy consensus optimized transaction privacy no privacy batching verify withdraw reward brigde centralize decentralize portal staking pdex | 1 |
5,409 | 12,451,970,866 | IssuesEvent | 2020-05-27 11:27:02 | corona-warn-app/cwa-documentation | https://api.github.com/repos/corona-warn-app/cwa-documentation | closed | Use Corona-Warn-Server instead of Corona-Warn-App-Server | architecture documentation | ## Where to find the issue
Complete Project documentation.
Especially visible in figure_7.svg
## Describe the issue
Corona-Warn-App-Server is not unambiguous
Example figure 7 Is the corona Warn App Database on the server or on the phone? It should be on the server but why is it not called Corona Warn Server Database?
An does the App-server run on the iPhone like App suggests.
No we got a
Corona-Warn-Server with a Corona-Warn-Server-Database
and a Corona-Warn-App
## Suggested change
Rename (now -early- lower cost) Corona-Warn-App-Server to Corona-Warn-Server
| 1.0 | Use Corona-Warn-Server instead of Corona-Warn-App-Server - ## Where to find the issue
Complete Project documentation.
Especially visible in figure_7.svg
## Describe the issue
Corona-Warn-App-Server is not unambiguous
Example figure 7 Is the corona Warn App Database on the server or on the phone? It should be on the server but why is it not called Corona Warn Server Database?
An does the App-server run on the iPhone like App suggests.
No we got a
Corona-Warn-Server with a Corona-Warn-Server-Database
and a Corona-Warn-App
## Suggested change
Rename (now -early- lower cost) Corona-Warn-App-Server to Corona-Warn-Server
| non_priority | use corona warn server instead of corona warn app server where to find the issue complete project documentation especially visible in figure svg describe the issue corona warn app server is not unambiguous example figure is the corona warn app database on the server or on the phone it should be on the server but why is it not called corona warn server database an does the app server run on the iphone like app suggests no we got a corona warn server with a corona warn server database and a corona warn app suggested change rename now early lower cost corona warn app server to corona warn server | 0 |
368,367 | 25,790,108,854 | IssuesEvent | 2022-12-10 02:33:42 | wesen/glazed | https://api.github.com/repos/wesen/glazed | reopened | doc: add example program | documentation | The example program should show how to:
- integrate the library with command line flag parsing
- register custom middlewares
| 1.0 | doc: add example program - The example program should show how to:
- integrate the library with command line flag parsing
- register custom middlewares
| non_priority | doc add example program the example program should show how to integrate the library with command line flag parsing register custom middlewares | 0 |
1,477 | 6,404,174,472 | IssuesEvent | 2017-08-07 01:23:43 | caskroom/homebrew-cask | https://api.github.com/repos/caskroom/homebrew-cask | closed | microsoft-office uninstall does not remove .app files from /Applications | awaiting maintainer feedback | #### General troubleshooting steps
- [X] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue.
- [X] None of the templates was appropriate for my issue, or I’m not sure.
- [X] I ran `brew update-reset && brew update` and retried my command.
- [X] I ran `brew doctor`, fixed as many issues as possible and retried my command.
- [X] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md).
#### Description of issue
`brew cask zap microsoft-office` should get rid of the apps, but the apps (microsoft-office is a suite) remain on /Applications/.
#### Output of your command with `--verbose --debug`
```
[I] gtklocker@schwarz ~> brew cask zap microsoft-office --verbose --debug
==> Zapping Cask microsoft-office
==> Implied "brew cask uninstall microsoft-office"
==> Un-installing artifacts
==> Determining which artifacts are present in Cask microsoft-office
==> 3 artifact/s defined
#<Hbc::Artifact::Uninstall:0x007fe6ad109b90>
#<Hbc::Artifact::Pkg:0x007fe6ad109758>
#<Hbc::Artifact::Zap:0x007fe6ad109460>
==> Un-installing artifact of class Hbc::Artifact::Uninstall
==> Running uninstall process for microsoft-office; your password may be necessary
==> Removing launchctl service com.microsoft.autoupdate.helpertool
==> Executing: ["/bin/launchctl", "list", "com.microsoft.autoupdate.helpertool"]
==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.autoupdate.helpertool"]
Password:
==> Removing launchctl service com.microsoft.office.licensing.helper
==> Executing: ["/bin/launchctl", "list", "com.microsoft.office.licensing.helper"]
==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.office.licensing.helper"]
==> Removing launchctl service com.microsoft.office.licensingV2.helper
==> Executing: ["/bin/launchctl", "list", "com.microsoft.office.licensingV2.helper"]
==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.office.licensingV2.helper"]
==> Uninstalling packages:
==> Executing: ["/usr/sbin/pkgutil", "--pkgs=com.microsoft.package.*"]
==> Executing: ["/usr/sbin/pkgutil", "--pkgs=com.microsoft.pkg.licensing"]
==> Dispatching zap stanza
==> Running zap process for microsoft-office; your password may be necessary
==> Removing files:
~/Library/Application Scripts/com.microsoft.Excel
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Office365ServiceV2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Outlook
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Powerpoint
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Word
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.errorreporting
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.onenote.mac
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.excel.sfl
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.powerpoint.sfl
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.word.sfl
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/Microsoft/uls/com.microsoft.autoupdate.fba
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/Microsoft/uls/com.microsoft.autoupdate2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/com.microsoft.autoupdate.fba
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/com.microsoft.autoupdate2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Excel
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Office365ServiceV2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Outlook
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Powerpoint
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Word
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.errorreporting
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.onenote.mac
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Cookies/com.microsoft.autoupdate.fba.binarycookies
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Cookies/com.microsoft.autoupdate2.binarycookies
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Group Containers/UBF8T346G9.Office
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Group Containers/UBF8T346G9.OfficeOsfWebHost
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Group Containers/UBF8T346G9.ms
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.Excel.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.Powerpoint.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.Word.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.autoupdate.fba.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.autoupdate2.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Saved Application State/com.microsoft.autoupdate2.savedState
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Saved Application State/com.microsoft.office.setupassistant.savedState
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
==> Removing directories if empty:
~/Library/Caches/Microsoft/uls
~/Library/Caches/Microsoft
==> Removing all staged versions of Cask 'microsoft-office'
==> Purging all staged versions of Cask microsoft-office
[I] gtklocker@schwarz ~> ls -d1 /Applications/Microsoft\ *
/Applications/Microsoft Excel.app
/Applications/Microsoft OneNote.app
/Applications/Microsoft Outlook.app
/Applications/Microsoft PowerPoint.app
/Applications/Microsoft Word.app
```
#### Output of `brew cask doctor`
```
[I] gtklocker@schwarz ~> brew cask doctor
==> Homebrew-Cask Version
Homebrew-Cask 1.3.0-39-gf57a172
caskroom/homebrew-cask (git revision ebc89; last commit 2017-08-07)
==> Homebrew-Cask Install Location
<NONE>
==> Homebrew-Cask Staging Location
/usr/local/Caskroom
==> Homebrew-Cask Cached Downloads
~/Library/Caches/Homebrew/Cask (47 files, 3.5GB)
==> Homebrew-Cask Taps:
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3676 casks)
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-fonts (1107 casks)
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-versions (160 casks)
==> Contents of $LOAD_PATH
/usr/local/Homebrew/Library/Homebrew/cask/lib
/usr/local/Homebrew/Library/Homebrew
/Library/Ruby/Site/2.0.0
/Library/Ruby/Site/2.0.0/x86_64-darwin16
/Library/Ruby/Site/2.0.0/universal-darwin16
/Library/Ruby/Site
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16
==> Environment Variables
LANG="en_GB.UTF-8"
PATH="/usr/local/bin:/usr/local/opt/fzf/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/Homebrew/Library/Homebrew/shims/scm"
SHELL="/usr/local/bin/fish"
```
| True | microsoft-office uninstall does not remove .app files from /Applications - #### General troubleshooting steps
- [X] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue.
- [X] None of the templates was appropriate for my issue, or I’m not sure.
- [X] I ran `brew update-reset && brew update` and retried my command.
- [X] I ran `brew doctor`, fixed as many issues as possible and retried my command.
- [X] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md).
#### Description of issue
`brew cask zap microsoft-office` should get rid of the apps, but the apps (microsoft-office is a suite) remain on /Applications/.
#### Output of your command with `--verbose --debug`
```
[I] gtklocker@schwarz ~> brew cask zap microsoft-office --verbose --debug
==> Zapping Cask microsoft-office
==> Implied "brew cask uninstall microsoft-office"
==> Un-installing artifacts
==> Determining which artifacts are present in Cask microsoft-office
==> 3 artifact/s defined
#<Hbc::Artifact::Uninstall:0x007fe6ad109b90>
#<Hbc::Artifact::Pkg:0x007fe6ad109758>
#<Hbc::Artifact::Zap:0x007fe6ad109460>
==> Un-installing artifact of class Hbc::Artifact::Uninstall
==> Running uninstall process for microsoft-office; your password may be necessary
==> Removing launchctl service com.microsoft.autoupdate.helpertool
==> Executing: ["/bin/launchctl", "list", "com.microsoft.autoupdate.helpertool"]
==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.autoupdate.helpertool"]
Password:
==> Removing launchctl service com.microsoft.office.licensing.helper
==> Executing: ["/bin/launchctl", "list", "com.microsoft.office.licensing.helper"]
==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.office.licensing.helper"]
==> Removing launchctl service com.microsoft.office.licensingV2.helper
==> Executing: ["/bin/launchctl", "list", "com.microsoft.office.licensingV2.helper"]
==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.office.licensingV2.helper"]
==> Uninstalling packages:
==> Executing: ["/usr/sbin/pkgutil", "--pkgs=com.microsoft.package.*"]
==> Executing: ["/usr/sbin/pkgutil", "--pkgs=com.microsoft.pkg.licensing"]
==> Dispatching zap stanza
==> Running zap process for microsoft-office; your password may be necessary
==> Removing files:
~/Library/Application Scripts/com.microsoft.Excel
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Office365ServiceV2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Outlook
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Powerpoint
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Word
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.errorreporting
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.onenote.mac
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.excel.sfl
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.powerpoint.sfl
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.word.sfl
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/Microsoft/uls/com.microsoft.autoupdate.fba
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/Microsoft/uls/com.microsoft.autoupdate2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/com.microsoft.autoupdate.fba
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/com.microsoft.autoupdate2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Excel
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Office365ServiceV2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Outlook
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Powerpoint
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Word
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.errorreporting
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.onenote.mac
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Cookies/com.microsoft.autoupdate.fba.binarycookies
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Cookies/com.microsoft.autoupdate2.binarycookies
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Group Containers/UBF8T346G9.Office
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Group Containers/UBF8T346G9.OfficeOsfWebHost
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Group Containers/UBF8T346G9.ms
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.Excel.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.Powerpoint.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.Word.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.autoupdate.fba.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.autoupdate2.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Saved Application State/com.microsoft.autoupdate2.savedState
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Saved Application State/com.microsoft.office.setupassistant.savedState
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
==> Removing directories if empty:
~/Library/Caches/Microsoft/uls
~/Library/Caches/Microsoft
==> Removing all staged versions of Cask 'microsoft-office'
==> Purging all staged versions of Cask microsoft-office
[I] gtklocker@schwarz ~> ls -d1 /Applications/Microsoft\ *
/Applications/Microsoft Excel.app
/Applications/Microsoft OneNote.app
/Applications/Microsoft Outlook.app
/Applications/Microsoft PowerPoint.app
/Applications/Microsoft Word.app
```
#### Output of `brew cask doctor`
```
[I] gtklocker@schwarz ~> brew cask doctor
==> Homebrew-Cask Version
Homebrew-Cask 1.3.0-39-gf57a172
caskroom/homebrew-cask (git revision ebc89; last commit 2017-08-07)
==> Homebrew-Cask Install Location
<NONE>
==> Homebrew-Cask Staging Location
/usr/local/Caskroom
==> Homebrew-Cask Cached Downloads
~/Library/Caches/Homebrew/Cask (47 files, 3.5GB)
==> Homebrew-Cask Taps:
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3676 casks)
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-fonts (1107 casks)
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-versions (160 casks)
==> Contents of $LOAD_PATH
/usr/local/Homebrew/Library/Homebrew/cask/lib
/usr/local/Homebrew/Library/Homebrew
/Library/Ruby/Site/2.0.0
/Library/Ruby/Site/2.0.0/x86_64-darwin16
/Library/Ruby/Site/2.0.0/universal-darwin16
/Library/Ruby/Site
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16
==> Environment Variables
LANG="en_GB.UTF-8"
PATH="/usr/local/bin:/usr/local/opt/fzf/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/Homebrew/Library/Homebrew/shims/scm"
SHELL="/usr/local/bin/fish"
```
| non_priority | microsoft office uninstall does not remove app files from applications general troubleshooting steps i have checked the instructions for or before opening the issue none of the templates was appropriate for my issue or i’m not sure i ran brew update reset brew update and retried my command i ran brew doctor fixed as many issues as possible and retried my command i understand that description of issue brew cask zap microsoft office should get rid of the apps but the apps microsoft office is a suite remain on applications output of your command with verbose debug gtklocker schwarz brew cask zap microsoft office verbose debug zapping cask microsoft office implied brew cask uninstall microsoft office un installing artifacts determining which artifacts are present in cask microsoft office artifact s defined un installing artifact of class hbc artifact uninstall running uninstall process for microsoft office your password may be necessary removing launchctl service com microsoft autoupdate helpertool executing executing password removing launchctl service com microsoft office licensing helper executing executing removing launchctl service com microsoft office helper executing executing uninstalling packages executing executing dispatching zap stanza running zap process for microsoft office your password may be necessary removing files library application scripts com microsoft excel executing library application scripts com microsoft executing library application scripts com microsoft outlook executing library application scripts com microsoft powerpoint executing library application scripts com microsoft word executing library application scripts com microsoft errorreporting executing library application scripts com microsoft onenote mac executing library application support com apple sharedfilelist com apple lssharedfilelist applicationrecentdocuments com microsoft excel sfl executing library application support com apple sharedfilelist com apple lssharedfilelist applicationrecentdocuments com microsoft powerpoint sfl executing library application support com apple sharedfilelist com apple lssharedfilelist applicationrecentdocuments com microsoft word sfl executing library caches microsoft uls com microsoft autoupdate fba executing library caches microsoft uls com microsoft executing library caches com microsoft autoupdate fba executing library caches com microsoft executing library containers com microsoft excel executing library containers com microsoft executing library containers com microsoft outlook executing library containers com microsoft powerpoint executing library containers com microsoft word executing library containers com microsoft errorreporting executing library containers com microsoft onenote mac executing library cookies com microsoft autoupdate fba binarycookies executing library cookies com microsoft binarycookies executing library group containers office executing library group containers officeosfwebhost executing library group containers ms executing library preferences com microsoft excel plist executing library preferences com microsoft powerpoint plist executing library preferences com microsoft word plist executing library preferences com microsoft autoupdate fba plist executing library preferences com microsoft plist executing library saved application state com microsoft savedstate executing library saved application state com microsoft office setupassistant savedstate executing removing directories if empty library caches microsoft uls library caches microsoft removing all staged versions of cask microsoft office purging all staged versions of cask microsoft office gtklocker schwarz ls applications microsoft applications microsoft excel app applications microsoft onenote app applications microsoft outlook app applications microsoft powerpoint app applications microsoft word app output of brew cask doctor gtklocker schwarz brew cask doctor homebrew cask version homebrew cask caskroom homebrew cask git revision last commit homebrew cask install location homebrew cask staging location usr local caskroom homebrew cask cached downloads library caches homebrew cask files homebrew cask taps usr local homebrew library taps caskroom homebrew cask casks usr local homebrew library taps caskroom homebrew fonts casks usr local homebrew library taps caskroom homebrew versions casks contents of load path usr local homebrew library homebrew cask lib usr local homebrew library homebrew library ruby site library ruby site library ruby site universal library ruby site system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby universal system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby universal environment variables lang en gb utf path usr local bin usr local opt fzf bin usr bin bin usr sbin sbin opt bin usr local homebrew library homebrew shims scm shell usr local bin fish | 0 |
434,868 | 30,473,189,939 | IssuesEvent | 2023-07-17 14:49:48 | Infisical/infisical | https://api.github.com/repos/Infisical/infisical | closed | Typo in Docs | 🐞 bug 📑 documentation good first issue | ### Describe the bug
There's a little typo that I encountered
### To Reproduce
Steps to reproduce the behavior:
1. Go to [here](https://infisical.com/docs/documentation/platform/token)
2. See the typo error
### Screenshots

| 1.0 | Typo in Docs - ### Describe the bug
There's a little typo that I encountered
### To Reproduce
Steps to reproduce the behavior:
1. Go to [here](https://infisical.com/docs/documentation/platform/token)
2. See the typo error
### Screenshots

| non_priority | typo in docs describe the bug there s a little typo that i encountered to reproduce steps to reproduce the behavior go to see the typo error screenshots | 0 |
232,823 | 17,796,042,525 | IssuesEvent | 2021-08-31 22:24:22 | mhostetter/galois | https://api.github.com/repos/mhostetter/galois | closed | Fix code completion | bug documentation | Some classes have obscure code completion. This may be due to multiple inheritance or the docstring needing to be placed inside `__new__()` or `__init__()` and not at the class level.
`LFSR` shows the arguments, but doesn't show the docstring.

`Poly` seems to work fine.

| 1.0 | Fix code completion - Some classes have obscure code completion. This may be due to multiple inheritance or the docstring needing to be placed inside `__new__()` or `__init__()` and not at the class level.
`LFSR` shows the arguments, but doesn't show the docstring.

`Poly` seems to work fine.

| non_priority | fix code completion some classes have obscure code completion this may be due to multiple inheritance or the docstring needing to be placed inside new or init and not at the class level lfsr shows the arguments but doesn t show the docstring poly seems to work fine | 0 |
50,269 | 21,047,069,382 | IssuesEvent | 2022-03-31 17:00:53 | Azure/azure-sdk-for-java | https://api.github.com/repos/Azure/azure-sdk-for-java | closed | Root causing protonj "no current delivery" | Event Hubs Service Bus pillar-reliability amqp | ## Background
Two customer tickets reported that the ProtonJ error "no current delivery" appeared in the logs, but when it happened in the lower layer, it activated a code path in the upper layer that triggered a reliability issue;[ this-pr](https://github.com/Azure/azure-sdk-for-java/pull/24141) addressed it.
This ticket is created so that we can focus on ProtonJ error "no current delivery" .
## Root causing
As part of [this](https://github.com/Azure/azure-sdk-for-java/issues/27482) investigation, we learned that it is possible that the ProtonJ object state can be mutated (e.g disposed) after the "scheduling" of work to the message-pump but before the actual "execution" of the work.
Sketching the control flow, it looks like the reason for "no current delivery" is such a mutation; the flow appears to be below -
1. ProtonJ Receiver received a delivery. It updated its __current_ pointer to refer to this delivery then handed it over to ReceiveLinkHandler.
2. ReceiveLinkHandler [enqueues](https://github.com/Azure/azure-sdk-for-java/blob/0ddca1176c70abd30ab69d71c219f487aef4d0d3/sdk/core/azure-core-amqp/src/main/java/com/azure/core/amqp/implementation/handler/ReceiveLinkHandler.java#L164) the delivery to deliveries-Flux.
3. The async-chain attached to deliveries-Flux [emits](https://github.com/Azure/azure-sdk-for-java/blob/0ddca1176c70abd30ab69d71c219f487aef4d0d3/sdk/core/azure-core-amqp/src/main/java/com/azure/core/amqp/implementation/ReactorReceiver.java#L82) the delivery to ReactorReceiver.
4. The ReactorReceiver "[schedule](https://github.com/Azure/azure-sdk-for-java/blob/0ddca1176c70abd30ab69d71c219f487aef4d0d3/sdk/core/azure-core-amqp/src/main/java/com/azure/core/amqp/implementation/ReactorReceiver.java#L85)" a work to message-pump thread to decode this delivery.
5. The async-chain is waiting asynchronously for the Sink to complete. The completion happens only when scheduled work in 4 "executes" in the future.
6. The service traffic closed the amqp-receive-link, inturn [frees](https://github.com/Azure/azure-sdk-for-java/blob/0ddca1176c70abd30ab69d71c219f487aef4d0d3/sdk/core/azure-core-amqp/src/main/java/com/azure/core/amqp/implementation/ReactorReceiver.java#L347) the ProtonJ Receiver using message-pump thread. Note: the user-work in 4 and proton-j work in 6 does not execute strictly in FIFO order (two different pipes drive it), but the serial execution is guaranteed.
7. As part of freeing the ProtonJ Receiver, the ProtonJ library frees the delivery (that 4 refers to), settling the delivery and advancing the __current_ pointer. The __current_ become null.
8. ReceiveLinkHandler [enqueues](https://github.com/Azure/azure-sdk-for-java/blob/0ddca1176c70abd30ab69d71c219f487aef4d0d3/sdk/core/azure-core-amqp/src/main/java/com/azure/core/amqp/implementation/handler/ReceiveLinkHandler.java#L237) completion-event to deliveries-Flux. The completion-event emission to the async-chain is deferred because the Sink for the last event (5) is yet to be complete.
9. Work scheduled in 4 executes, but since __current_ is null and not the one the work is referring to, ProtonJ throws "no current delivery"
The exception in 9 goes to global error handling as an unhandled exception. Three 3 issues -
a. The Sink in 5 never errored
b. The completion-event in 8 never emits (because enqueued after a)
c. The unhandled exception forces the parent amqp-connection to close
The good thing is, there is no one [downstream](https://github.com/Azure/azure-sdk-for-java/blob/0ddca1176c70abd30ab69d71c219f487aef4d0d3/sdk/eventhubs/azure-messaging-eventhubs/src/main/java/com/azure/messaging/eventhubs/implementation/AmqpReceiveLinkProcessor.java#L286) waiting to react to the "terminal event" from a (or b) to happen. But c is concerning because the closure of one amqp-receiver-link combined with the race resulted in amqp-connection to close, which means it causes all other healthy amqp-receiver-links in the same amqp-connection to disconnect from service.
## Fix
Again fix is simple as [this](https://github.com/Azure/azure-sdk-for-java/issues/27482) ticket, i.e. do a deferred check for disposal of the Receiver so that the side effect is limited to that specific Receiver and not bypassed to the parent connection.
## References
The customer tickets - https://github.com/Azure/azure-sdk-for-java/issues/24575 and https://github.com/Azure/azure-sdk-for-java/issues/26975
| 1.0 | Root causing protonj "no current delivery" - ## Background
Two customer tickets reported that the ProtonJ error "no current delivery" appeared in the logs, but when it happened in the lower layer, it activated a code path in the upper layer that triggered a reliability issue;[ this-pr](https://github.com/Azure/azure-sdk-for-java/pull/24141) addressed it.
This ticket is created so that we can focus on ProtonJ error "no current delivery" .
## Root causing
As part of [this](https://github.com/Azure/azure-sdk-for-java/issues/27482) investigation, we learned that it is possible that the ProtonJ object state can be mutated (e.g disposed) after the "scheduling" of work to the message-pump but before the actual "execution" of the work.
Sketching the control flow, it looks like the reason for "no current delivery" is such a mutation; the flow appears to be below -
1. ProtonJ Receiver received a delivery. It updated its __current_ pointer to refer to this delivery then handed it over to ReceiveLinkHandler.
2. ReceiveLinkHandler [enqueues](https://github.com/Azure/azure-sdk-for-java/blob/0ddca1176c70abd30ab69d71c219f487aef4d0d3/sdk/core/azure-core-amqp/src/main/java/com/azure/core/amqp/implementation/handler/ReceiveLinkHandler.java#L164) the delivery to deliveries-Flux.
3. The async-chain attached to deliveries-Flux [emits](https://github.com/Azure/azure-sdk-for-java/blob/0ddca1176c70abd30ab69d71c219f487aef4d0d3/sdk/core/azure-core-amqp/src/main/java/com/azure/core/amqp/implementation/ReactorReceiver.java#L82) the delivery to ReactorReceiver.
4. The ReactorReceiver "[schedule](https://github.com/Azure/azure-sdk-for-java/blob/0ddca1176c70abd30ab69d71c219f487aef4d0d3/sdk/core/azure-core-amqp/src/main/java/com/azure/core/amqp/implementation/ReactorReceiver.java#L85)" a work to message-pump thread to decode this delivery.
5. The async-chain is waiting asynchronously for the Sink to complete. The completion happens only when scheduled work in 4 "executes" in the future.
6. The service traffic closed the amqp-receive-link, inturn [frees](https://github.com/Azure/azure-sdk-for-java/blob/0ddca1176c70abd30ab69d71c219f487aef4d0d3/sdk/core/azure-core-amqp/src/main/java/com/azure/core/amqp/implementation/ReactorReceiver.java#L347) the ProtonJ Receiver using message-pump thread. Note: the user-work in 4 and proton-j work in 6 does not execute strictly in FIFO order (two different pipes drive it), but the serial execution is guaranteed.
7. As part of freeing the ProtonJ Receiver, the ProtonJ library frees the delivery (that 4 refers to), settling the delivery and advancing the __current_ pointer. The __current_ become null.
8. ReceiveLinkHandler [enqueues](https://github.com/Azure/azure-sdk-for-java/blob/0ddca1176c70abd30ab69d71c219f487aef4d0d3/sdk/core/azure-core-amqp/src/main/java/com/azure/core/amqp/implementation/handler/ReceiveLinkHandler.java#L237) completion-event to deliveries-Flux. The completion-event emission to the async-chain is deferred because the Sink for the last event (5) is yet to be complete.
9. Work scheduled in 4 executes, but since __current_ is null and not the one the work is referring to, ProtonJ throws "no current delivery"
The exception in 9 goes to global error handling as an unhandled exception. Three 3 issues -
a. The Sink in 5 never errored
b. The completion-event in 8 never emits (because enqueued after a)
c. The unhandled exception forces the parent amqp-connection to close
The good thing is, there is no one [downstream](https://github.com/Azure/azure-sdk-for-java/blob/0ddca1176c70abd30ab69d71c219f487aef4d0d3/sdk/eventhubs/azure-messaging-eventhubs/src/main/java/com/azure/messaging/eventhubs/implementation/AmqpReceiveLinkProcessor.java#L286) waiting to react to the "terminal event" from a (or b) to happen. But c is concerning because the closure of one amqp-receiver-link combined with the race resulted in amqp-connection to close, which means it causes all other healthy amqp-receiver-links in the same amqp-connection to disconnect from service.
## Fix
Again fix is simple as [this](https://github.com/Azure/azure-sdk-for-java/issues/27482) ticket, i.e. do a deferred check for disposal of the Receiver so that the side effect is limited to that specific Receiver and not bypassed to the parent connection.
## References
The customer tickets - https://github.com/Azure/azure-sdk-for-java/issues/24575 and https://github.com/Azure/azure-sdk-for-java/issues/26975
| non_priority | root causing protonj no current delivery background two customer tickets reported that the protonj error no current delivery appeared in the logs but when it happened in the lower layer it activated a code path in the upper layer that triggered a reliability issue addressed it this ticket is created so that we can focus on protonj error no current delivery root causing as part of investigation we learned that it is possible that the protonj object state can be mutated e g disposed after the scheduling of work to the message pump but before the actual execution of the work sketching the control flow it looks like the reason for no current delivery is such a mutation the flow appears to be below protonj receiver received a delivery it updated its current pointer to refer to this delivery then handed it over to receivelinkhandler receivelinkhandler the delivery to deliveries flux the async chain attached to deliveries flux the delivery to reactorreceiver the reactorreceiver a work to message pump thread to decode this delivery the async chain is waiting asynchronously for the sink to complete the completion happens only when scheduled work in executes in the future the service traffic closed the amqp receive link inturn the protonj receiver using message pump thread note the user work in and proton j work in does not execute strictly in fifo order two different pipes drive it but the serial execution is guaranteed as part of freeing the protonj receiver the protonj library frees the delivery that refers to settling the delivery and advancing the current pointer the current become null receivelinkhandler completion event to deliveries flux the completion event emission to the async chain is deferred because the sink for the last event is yet to be complete work scheduled in executes but since current is null and not the one the work is referring to protonj throws no current delivery the exception in goes to global error handling as an unhandled exception three issues a the sink in never errored b the completion event in never emits because enqueued after a c the unhandled exception forces the parent amqp connection to close the good thing is there is no one waiting to react to the terminal event from a or b to happen but c is concerning because the closure of one amqp receiver link combined with the race resulted in amqp connection to close which means it causes all other healthy amqp receiver links in the same amqp connection to disconnect from service fix again fix is simple as ticket i e do a deferred check for disposal of the receiver so that the side effect is limited to that specific receiver and not bypassed to the parent connection references the customer tickets and | 0 |
209,865 | 7,180,654,065 | IssuesEvent | 2018-02-01 00:22:05 | INN/umbrella-energynewsnetwork | https://api.github.com/repos/INN/umbrella-energynewsnetwork | closed | MORI isn't saving things properly | bug priority: normal | The following URL
> 'posts-from-site-id-64/in-michigan-solar-growth-meets-uncertainty-with-end-of-net-metering'
where `posts-from-site-id-64` is the slug of a term in the region taxonomy
Might match one of these rules from `$wp_rewrite->wp_rewrite_rules()`:
```
'region/([^/]+)/feed/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?region=$matches[1]&feed=$matches[2]',
'region/([^/]+)/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?region=$matches[1]&feed=$matches[2]',
'region/([^/]+)/embed/?$' => 'index.php?region=$matches[1]&embed=true',
'region/([^/]+)/page/?([0-9]{1,})/?$' => 'index.php?region=$matches[1]&paged=$matches[2]',
'region/([^/]+)/comment-page-([0-9]{1,})/?$' => 'index.php?region=$matches[1]&cpage=$matches[2]',
'region/([^/]+)/?$' => 'index.php?region=$matches[1]',
'([^/]+)/([^/]+)/embed/?$' => 'index.php?region=$matches[1]&name=$matches[2]&embed=true',
'([^/]+)/([^/]+)/trackback/?$' => 'index.php?region=$matches[1]&name=$matches[2]&tb=1',
'([^/]+)/([^/]+)/feed/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?region=$matches[1]&name=$matches[2]&feed=$matches[3]',
'([^/]+)/([^/]+)/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?region=$matches[1]&name=$matches[2]&feed=$matches[3]',
'([^/]+)/([^/]+)/page/?([0-9]{1,})/?$' => 'index.php?region=$matches[1]&name=$matches[2]&paged=$matches[3]',
'([^/]+)/([^/]+)/comment-page-([0-9]{1,})/?$' => 'index.php?region=$matches[1]&name=$matches[2]&cpage=$matches[3]',
'([^/]+)/([^/]+)(?:/([0-9]+))?/?$' => 'index.php?region=$matches[1]&name=$matches[2]&page=$matches[3]',
'[^/]+/[^/]+/([^/]+)/?$' => 'index.php?attachment=$matches[1]',
'[^/]+/[^/]+/([^/]+)/trackback/?$' => 'index.php?attachment=$matches[1]&tb=1',
'[^/]+/[^/]+/([^/]+)/feed/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?attachment=$matches[1]&feed=$matches[2]',
'[^/]+/[^/]+/([^/]+)/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?attachment=$matches[1]&feed=$matches[2]',
'[^/]+/[^/]+/([^/]+)/comment-page-([0-9]{1,})/?$' => 'index.php?attachment=$matches[1]&cpage=$matches[2]',
'[^/]+/[^/]+/([^/]+)/embed/?$' => 'index.php?attachment=$matches[1]&embed=true',
'([^/]+)/feed/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?region=$matches[1]&feed=$matches[2]',
'([^/]+)/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?region=$matches[1]&feed=$matches[2]',
'([^/]+)/embed/?$' => 'index.php?region=$matches[1]&embed=true',
'([^/]+)/page/?([0-9]{1,})/?$' => 'index.php?region=$matches[1]&paged=$matches[2]',
'([^/]+)/comment-page-([0-9]{1,})/?$' => 'index.php?region=$matches[1]&cpage=$matches[2]',
'([^/]+)/?$' => 'index.php?region=$matches[1]',
```
It's matching `'(.?.+?)(?:/([0-9]+))?/?$'` which should go into the pattern `'index.php?pagename=$matches[1]&page=$matches[2]'`, which is wrong - this isn't a page.
I have a sneaking suspicion that this is because of the page stuff in the rewrite rules firing at the wrong time.
| 1.0 | MORI isn't saving things properly - The following URL
> 'posts-from-site-id-64/in-michigan-solar-growth-meets-uncertainty-with-end-of-net-metering'
where `posts-from-site-id-64` is the slug of a term in the region taxonomy
Might match one of these rules from `$wp_rewrite->wp_rewrite_rules()`:
```
'region/([^/]+)/feed/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?region=$matches[1]&feed=$matches[2]',
'region/([^/]+)/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?region=$matches[1]&feed=$matches[2]',
'region/([^/]+)/embed/?$' => 'index.php?region=$matches[1]&embed=true',
'region/([^/]+)/page/?([0-9]{1,})/?$' => 'index.php?region=$matches[1]&paged=$matches[2]',
'region/([^/]+)/comment-page-([0-9]{1,})/?$' => 'index.php?region=$matches[1]&cpage=$matches[2]',
'region/([^/]+)/?$' => 'index.php?region=$matches[1]',
'([^/]+)/([^/]+)/embed/?$' => 'index.php?region=$matches[1]&name=$matches[2]&embed=true',
'([^/]+)/([^/]+)/trackback/?$' => 'index.php?region=$matches[1]&name=$matches[2]&tb=1',
'([^/]+)/([^/]+)/feed/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?region=$matches[1]&name=$matches[2]&feed=$matches[3]',
'([^/]+)/([^/]+)/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?region=$matches[1]&name=$matches[2]&feed=$matches[3]',
'([^/]+)/([^/]+)/page/?([0-9]{1,})/?$' => 'index.php?region=$matches[1]&name=$matches[2]&paged=$matches[3]',
'([^/]+)/([^/]+)/comment-page-([0-9]{1,})/?$' => 'index.php?region=$matches[1]&name=$matches[2]&cpage=$matches[3]',
'([^/]+)/([^/]+)(?:/([0-9]+))?/?$' => 'index.php?region=$matches[1]&name=$matches[2]&page=$matches[3]',
'[^/]+/[^/]+/([^/]+)/?$' => 'index.php?attachment=$matches[1]',
'[^/]+/[^/]+/([^/]+)/trackback/?$' => 'index.php?attachment=$matches[1]&tb=1',
'[^/]+/[^/]+/([^/]+)/feed/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?attachment=$matches[1]&feed=$matches[2]',
'[^/]+/[^/]+/([^/]+)/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?attachment=$matches[1]&feed=$matches[2]',
'[^/]+/[^/]+/([^/]+)/comment-page-([0-9]{1,})/?$' => 'index.php?attachment=$matches[1]&cpage=$matches[2]',
'[^/]+/[^/]+/([^/]+)/embed/?$' => 'index.php?attachment=$matches[1]&embed=true',
'([^/]+)/feed/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?region=$matches[1]&feed=$matches[2]',
'([^/]+)/(feed|rdf|rss|rss2|atom|fulltext|mailchimp)/?$' => 'index.php?region=$matches[1]&feed=$matches[2]',
'([^/]+)/embed/?$' => 'index.php?region=$matches[1]&embed=true',
'([^/]+)/page/?([0-9]{1,})/?$' => 'index.php?region=$matches[1]&paged=$matches[2]',
'([^/]+)/comment-page-([0-9]{1,})/?$' => 'index.php?region=$matches[1]&cpage=$matches[2]',
'([^/]+)/?$' => 'index.php?region=$matches[1]',
```
It's matching `'(.?.+?)(?:/([0-9]+))?/?$'` which should go into the pattern `'index.php?pagename=$matches[1]&page=$matches[2]'`, which is wrong - this isn't a page.
I have a sneaking suspicion that this is because of the page stuff in the rewrite rules firing at the wrong time.
| priority | mori isn t saving things properly the following url posts from site id in michigan solar growth meets uncertainty with end of net metering where posts from site id is the slug of a term in the region taxonomy might match one of these rules from wp rewrite wp rewrite rules region feed feed rdf rss atom fulltext mailchimp index php region matches feed matches region feed rdf rss atom fulltext mailchimp index php region matches feed matches region embed index php region matches embed true region page index php region matches paged matches region comment page index php region matches cpage matches region index php region matches embed index php region matches name matches embed true trackback index php region matches name matches tb feed feed rdf rss atom fulltext mailchimp index php region matches name matches feed matches feed rdf rss atom fulltext mailchimp index php region matches name matches feed matches page index php region matches name matches paged matches comment page index php region matches name matches cpage matches index php region matches name matches page matches index php attachment matches trackback index php attachment matches tb feed feed rdf rss atom fulltext mailchimp index php attachment matches feed matches feed rdf rss atom fulltext mailchimp index php attachment matches feed matches comment page index php attachment matches cpage matches embed index php attachment matches embed true feed feed rdf rss atom fulltext mailchimp index php region matches feed matches feed rdf rss atom fulltext mailchimp index php region matches feed matches embed index php region matches embed true page index php region matches paged matches comment page index php region matches cpage matches index php region matches it s matching which should go into the pattern index php pagename matches page matches which is wrong this isn t a page i have a sneaking suspicion that this is because of the page stuff in the rewrite rules firing at the wrong time | 1 |
6,160 | 2,584,002,073 | IssuesEvent | 2015-02-16 12:07:14 | TrinityCore/TrinityCore | https://api.github.com/repos/TrinityCore/TrinityCore | closed | [Bug]Tome of Mel'Thandris event missing | Comp-Database Feedback-FixOutdatedMissingWip Priority-Low Sub-Quests | How it should work:
1. Accept "The Howling Vale".
2. Use "Tome of Mel'Thandris".
3. "Velinde Starsong" will appear and say:
"The numbers of my companions dwindles, goddess, and my own power shall soon be insufficient to hold back the demons of Felwood."
"Goddess, grant me the power to overcome my enemies! Hear me, please, my need is desperate and my cause is just!"
"What... what is this? Could this be the answer to my prayers? Elune has granted me a weapon -- this scythe -- to defeat the demons."
Bug:
"Velinde Starsong" never appears.
Links:
The Howling Vale - http://www.wowwiki.com/Quest:The_Howling_Vale
Velinde Starsong - http://www.wowhead.com/npc=3946
Tome of Mel'Thandris - http://www.wowhead.com/object=19027
Core revision: d943b1d67980
Database version: TDB 335.53
Addons: Anticheat1 | 1.0 | [Bug]Tome of Mel'Thandris event missing - How it should work:
1. Accept "The Howling Vale".
2. Use "Tome of Mel'Thandris".
3. "Velinde Starsong" will appear and say:
"The numbers of my companions dwindles, goddess, and my own power shall soon be insufficient to hold back the demons of Felwood."
"Goddess, grant me the power to overcome my enemies! Hear me, please, my need is desperate and my cause is just!"
"What... what is this? Could this be the answer to my prayers? Elune has granted me a weapon -- this scythe -- to defeat the demons."
Bug:
"Velinde Starsong" never appears.
Links:
The Howling Vale - http://www.wowwiki.com/Quest:The_Howling_Vale
Velinde Starsong - http://www.wowhead.com/npc=3946
Tome of Mel'Thandris - http://www.wowhead.com/object=19027
Core revision: d943b1d67980
Database version: TDB 335.53
Addons: Anticheat1 | priority | tome of mel thandris event missing how it should work accept the howling vale use tome of mel thandris velinde starsong will appear and say the numbers of my companions dwindles goddess and my own power shall soon be insufficient to hold back the demons of felwood goddess grant me the power to overcome my enemies hear me please my need is desperate and my cause is just what what is this could this be the answer to my prayers elune has granted me a weapon this scythe to defeat the demons bug velinde starsong never appears links the howling vale velinde starsong tome of mel thandris core revision database version tdb addons | 1 |
5,601 | 2,955,394,197 | IssuesEvent | 2015-07-08 02:32:25 | abargnesi/rantly | https://api.github.com/repos/abargnesi/rantly | opened | Make tests runnable. | documentation test | - [x] Update to latest minitest gem.
- [x] Use minitest spec-style testing to match *shoulda* appeal.
- [ ] Update README documentation. | 1.0 | Make tests runnable. - - [x] Update to latest minitest gem.
- [x] Use minitest spec-style testing to match *shoulda* appeal.
- [ ] Update README documentation. | non_priority | make tests runnable update to latest minitest gem use minitest spec style testing to match shoulda appeal update readme documentation | 0 |
50,513 | 10,518,155,621 | IssuesEvent | 2019-09-29 08:46:52 | triplea-game/triplea | https://api.github.com/repos/triplea-game/triplea | closed | 'ta_users' Schema Update | Design Stale code discussion | Following up to a number of initiatives, I'd like to propose/discuss a schema update. An ER diagram is below:

A few items are forward looking and would be added in later iterations:
- chat_history table
- token column on user_session
- match_count on bad_word
Otherwise of note:
- schema is mostly normalized with most tables having FK constraints
- nullable user_id on user_session is the difference between anonymous and registered login
- tracking of all bans/ips is meant to facilitate banning
Very happy to answer any questions and curious for any feedback/thoughts. | 1.0 | 'ta_users' Schema Update - Following up to a number of initiatives, I'd like to propose/discuss a schema update. An ER diagram is below:

A few items are forward looking and would be added in later iterations:
- chat_history table
- token column on user_session
- match_count on bad_word
Otherwise of note:
- schema is mostly normalized with most tables having FK constraints
- nullable user_id on user_session is the difference between anonymous and registered login
- tracking of all bans/ips is meant to facilitate banning
Very happy to answer any questions and curious for any feedback/thoughts. | non_priority | ta users schema update following up to a number of initiatives i d like to propose discuss a schema update an er diagram is below a few items are forward looking and would be added in later iterations chat history table token column on user session match count on bad word otherwise of note schema is mostly normalized with most tables having fk constraints nullable user id on user session is the difference between anonymous and registered login tracking of all bans ips is meant to facilitate banning very happy to answer any questions and curious for any feedback thoughts | 0 |
360,999 | 10,700,471,126 | IssuesEvent | 2019-10-24 00:07:54 | octobercms/october | https://api.github.com/repos/octobercms/october | closed | Pivot data support in deferred binding | Priority: High Status: Archived Status: Review Needed Type: Enhancement | Is there any way to create relation manager which has some pivot data and works with deffered binding?
Are there any chances for October to support this anytime soon?
| 1.0 | Pivot data support in deferred binding - Is there any way to create relation manager which has some pivot data and works with deffered binding?
Are there any chances for October to support this anytime soon?
| priority | pivot data support in deferred binding is there any way to create relation manager which has some pivot data and works with deffered binding are there any chances for october to support this anytime soon | 1 |
43,137 | 5,522,092,432 | IssuesEvent | 2017-03-19 20:44:19 | rust-lang/rust | https://api.github.com/repos/rust-lang/rust | closed | ICE on with #[repr(..)] for single-variant enum | E-needstest I-ICE | ``` rust
#[repr(u8)]
enum Foo {
Foo(u8),
}
fn main() {
match Foo::Foo(1) {
_ => ()
}
}
```
Same error on all nightly and stable:
```
error: internal compiler error: unexpected panic
note: the compiler unexpectedly panicked. this is a bug.
note: we would appreciate a bug report: https://github.com/rust-lang/rust/blob/master/CONTRIBUTING.md#bug-reports
thread 'rustc' panicked at 'assertion failed: `(left == right)` (left: `ReprInt(<anon>:1:8: 1:10, UnsignedInt(u8))`, right: `ReprAny`)', ../src/librustc/ty/layout.rs:953
note: Run with `RUST_BACKTRACE=1` for a backtrace.
```
| 1.0 | ICE on with #[repr(..)] for single-variant enum - ``` rust
#[repr(u8)]
enum Foo {
Foo(u8),
}
fn main() {
match Foo::Foo(1) {
_ => ()
}
}
```
Same error on all nightly and stable:
```
error: internal compiler error: unexpected panic
note: the compiler unexpectedly panicked. this is a bug.
note: we would appreciate a bug report: https://github.com/rust-lang/rust/blob/master/CONTRIBUTING.md#bug-reports
thread 'rustc' panicked at 'assertion failed: `(left == right)` (left: `ReprInt(<anon>:1:8: 1:10, UnsignedInt(u8))`, right: `ReprAny`)', ../src/librustc/ty/layout.rs:953
note: Run with `RUST_BACKTRACE=1` for a backtrace.
```
| non_priority | ice on with for single variant enum rust enum foo foo fn main match foo foo same error on all nightly and stable error internal compiler error unexpected panic note the compiler unexpectedly panicked this is a bug note we would appreciate a bug report thread rustc panicked at assertion failed left right left reprint unsignedint right reprany src librustc ty layout rs note run with rust backtrace for a backtrace | 0 |
88,641 | 25,477,159,595 | IssuesEvent | 2022-11-25 15:34:37 | GoogleCloudPlatform/fda-mystudies | https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies | closed | [GCI] [SB] A popup message is getting displayed while adding non-organizational user account in the below scenario | Bug P2 Study builder Process: Fixed Process: Tested QA Process: Tested dev | **Steps:**
1. Login to SB
2. Click on Manage admins icon
3. Click on Add admin button
4. Complete first name, last name , select the role and enter non-organizational user email
5. Now enter invalid number in phone number field
6. Click on Add button and Observe.
**AR:** A popup message is getting displayed while adding non-organizational user account in the above scenario
**ER:** A popup message should get displayed once all the mandatory fields are completed without any error messages
https://user-images.githubusercontent.com/86007179/154673694-471cb419-d757-44ed-a2dc-6a935c8646cd.mp4
| 1.0 | [GCI] [SB] A popup message is getting displayed while adding non-organizational user account in the below scenario - **Steps:**
1. Login to SB
2. Click on Manage admins icon
3. Click on Add admin button
4. Complete first name, last name , select the role and enter non-organizational user email
5. Now enter invalid number in phone number field
6. Click on Add button and Observe.
**AR:** A popup message is getting displayed while adding non-organizational user account in the above scenario
**ER:** A popup message should get displayed once all the mandatory fields are completed without any error messages
https://user-images.githubusercontent.com/86007179/154673694-471cb419-d757-44ed-a2dc-6a935c8646cd.mp4
| non_priority | a popup message is getting displayed while adding non organizational user account in the below scenario steps login to sb click on manage admins icon click on add admin button complete first name last name select the role and enter non organizational user email now enter invalid number in phone number field click on add button and observe ar a popup message is getting displayed while adding non organizational user account in the above scenario er a popup message should get displayed once all the mandatory fields are completed without any error messages | 0 |
95,260 | 19,683,848,028 | IssuesEvent | 2022-01-11 19:39:37 | Pokecube-Development/Pokecube-Issues-and-Wiki | https://api.github.com/repos/Pokecube-Development/Pokecube-Issues-and-Wiki | closed | New worldgen chunks without NPCs | Bug - Code Fixed | #### Issue Description:
When generating brand new chunks in the world (including ultraspace), NPCs are not spawning in villages or structures. NPCs which had been previously spawned in villages still remain. Affects all NPC types from wandering trainers to stationary ones like N.
#### What happens:
Moving to new chunks, visiting villages or structures that have NPCs do not have them, when pokemobs spawn just fine.
#### What you expected to happen:
NPCs to spawn when I found cool new structures and villages
#### Steps to reproduce:
1. Move to brand new chunks in the world
2. Find places that NPCs/villagers spawn
3. No spawns
...
____
#### Affected Versions:
BiomesOPlenty-1.16.5-13.1.0.477-universal.jar
journeymap-1.16.5-5.7.3.jar
jei-1.16.5-7.7.1.139.jar
- Pokecube AIO: 1.16.5-3.14.0.jar
- Minecraft: 1.16.5
- Forge: 36.2.20
| 1.0 | New worldgen chunks without NPCs - #### Issue Description:
When generating brand new chunks in the world (including ultraspace), NPCs are not spawning in villages or structures. NPCs which had been previously spawned in villages still remain. Affects all NPC types from wandering trainers to stationary ones like N.
#### What happens:
Moving to new chunks, visiting villages or structures that have NPCs do not have them, when pokemobs spawn just fine.
#### What you expected to happen:
NPCs to spawn when I found cool new structures and villages
#### Steps to reproduce:
1. Move to brand new chunks in the world
2. Find places that NPCs/villagers spawn
3. No spawns
...
____
#### Affected Versions:
BiomesOPlenty-1.16.5-13.1.0.477-universal.jar
journeymap-1.16.5-5.7.3.jar
jei-1.16.5-7.7.1.139.jar
- Pokecube AIO: 1.16.5-3.14.0.jar
- Minecraft: 1.16.5
- Forge: 36.2.20
| non_priority | new worldgen chunks without npcs issue description when generating brand new chunks in the world including ultraspace npcs are not spawning in villages or structures npcs which had been previously spawned in villages still remain affects all npc types from wandering trainers to stationary ones like n what happens moving to new chunks visiting villages or structures that have npcs do not have them when pokemobs spawn just fine what you expected to happen npcs to spawn when i found cool new structures and villages steps to reproduce move to brand new chunks in the world find places that npcs villagers spawn no spawns affected versions biomesoplenty universal jar journeymap jar jei jar pokecube aio jar minecraft forge | 0 |
256,355 | 19,409,321,279 | IssuesEvent | 2021-12-20 07:36:47 | glific/glific | https://api.github.com/repos/glific/glific | opened | View contact profile and history | documentation | @abhi1203 Assign this document to you for review. Let me know if you have feedback or suggestions.
**View contact profile and history**
https://glific.slab.com/posts/12-view-contact-profile-and-history-v9zuhety | 1.0 | View contact profile and history - @abhi1203 Assign this document to you for review. Let me know if you have feedback or suggestions.
**View contact profile and history**
https://glific.slab.com/posts/12-view-contact-profile-and-history-v9zuhety | non_priority | view contact profile and history assign this document to you for review let me know if you have feedback or suggestions view contact profile and history | 0 |
226,481 | 24,947,191,808 | IssuesEvent | 2022-11-01 01:59:08 | saif-khan1211/First_real_scan | https://api.github.com/repos/saif-khan1211/First_real_scan | closed | CVE-2022-40156 (High) detected in xstream-1.4.8.jar - autoclosed | security vulnerability | ## CVE-2022-40156 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.8.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Path to vulnerable library: /target/libs/provided/xstream-1.4.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **xstream-1.4.8.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Those using Xstream to seralize XML data may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. This effect may support a denial of service attack.
<p>Publish Date: 2022-09-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-40156>CVE-2022-40156</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-40156 (High) detected in xstream-1.4.8.jar - autoclosed - ## CVE-2022-40156 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.8.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Path to vulnerable library: /target/libs/provided/xstream-1.4.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **xstream-1.4.8.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Those using Xstream to seralize XML data may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. This effect may support a denial of service attack.
<p>Publish Date: 2022-09-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-40156>CVE-2022-40156</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in xstream jar autoclosed cve high severity vulnerability vulnerable library xstream jar xstream is a serialization library from java objects to xml and back path to vulnerable library target libs provided xstream jar dependency hierarchy x xstream jar vulnerable library found in base branch main vulnerability details those using xstream to seralize xml data may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow this effect may support a denial of service attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend | 0 |
737,036 | 25,498,211,317 | IssuesEvent | 2022-11-27 23:05:04 | turgutcem/swe573 | https://api.github.com/repos/turgutcem/swe573 | opened | Develop User Recomendations | effort:medium priority:medium backend | Develop a generic user recomendation interface and implement this interface . | 1.0 | Develop User Recomendations - Develop a generic user recomendation interface and implement this interface . | priority | develop user recomendations develop a generic user recomendation interface and implement this interface | 1 |
567,767 | 16,891,711,744 | IssuesEvent | 2021-06-23 10:02:01 | googleapis/google-cloud-php | https://api.github.com/repos/googleapis/google-cloud-php | closed | Synthesis failed for analyticsdata | api: analyticsdata autosynth failure priority: p2 type: bug | Hello! Autosynth couldn't regenerate analyticsdata. :broken_heart:
Please investigate and fix this issue within 5 business days. While it remains broken,
this library cannot be updated with changes to the analyticsdata API, and the library grows
stale.
See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md
for trouble shooting tips.
Here's the output from running `synth.py`:
```
se.php.
2021-06-09 10:02:38,606 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Cohort.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Cohort.php.
2021-06-09 10:02:38,606 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/RunRealtimeReportResponse.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/RunRealtimeReportResponse.php.
2021-06-09 10:02:38,606 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/BatchRunReportsRequest.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/BatchRunReportsRequest.php.
2021-06-09 10:02:38,606 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/QuotaStatus.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/QuotaStatus.php.
2021-06-09 10:02:38,607 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/PropertyQuota.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/PropertyQuota.php.
2021-06-09 10:02:38,607 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/CohortReportSettings.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/CohortReportSettings.php.
2021-06-09 10:02:38,607 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/MetricHeader.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/MetricHeader.php.
2021-06-09 10:02:38,607 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/BatchRunPivotReportsResponse.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/BatchRunPivotReportsResponse.php.
2021-06-09 10:02:38,608 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/PivotOrderBy.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/PivotOrderBy.php.
2021-06-09 10:02:38,608 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/DimensionOrderBy.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/DimensionOrderBy.php.
2021-06-09 10:02:38,608 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/MetricOrderBy.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/MetricOrderBy.php.
2021-06-09 10:02:38,609 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/PivotOrderBy/PivotSelection.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/PivotOrderBy/PivotSelection.php.
2021-06-09 10:02:38,609 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/InListFilter.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/InListFilter.php.
2021-06-09 10:02:38,609 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/BetweenFilter.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/BetweenFilter.php.
2021-06-09 10:02:38,609 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/NumericFilter.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/NumericFilter.php.
2021-06-09 10:02:38,610 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/StringFilter.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/StringFilter.php.
2021-06-09 10:02:38,610 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/DimensionExpression/CaseExpression.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/DimensionExpression/CaseExpression.php.
2021-06-09 10:02:38,610 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/DimensionExpression/ConcatenateExpression.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/DimensionExpression/ConcatenateExpression.php.
2021-06-09 10:02:38,624 synthtool [WARNING] > No replacements made in src/**/V*/**/*.php for pattern final class, maybe replacement is no longer needed?
WARNING:synthtool:No replacements made in src/**/V*/**/*.php for pattern final class, maybe replacement is no longer needed?
2021-06-09 10:02:38,637 synthtool [WARNING] > No replacements made in src/**/V*/**/*.php for pattern public function ([s|g]\w{3,})Unwrapped, maybe replacement is no longer needed?
WARNING:synthtool:No replacements made in src/**/V*/**/*.php for pattern public function ([s|g]\w{3,})Unwrapped, maybe replacement is no longer needed?
2021-06-09 10:02:38,946 synthtool [WARNING] > No replacements made in src/**/V*/**/*.php for pattern (.{0,})\]\((/.{0,})\), maybe replacement is no longer needed?
WARNING:synthtool:No replacements made in src/**/V*/**/*.php for pattern (.{0,})\]\((/.{0,})\), maybe replacement is no longer needed?
Traceback (most recent call last):
File "/usr/local/lib/python3.8/runpy.py", line 194, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/usr/local/lib/python3.8/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "/workspace/synthtool/__main__.py", line 102, in <module>
main()
File "/workspace/env/lib/python3.8/site-packages/click/core.py", line 1137, in __call__
return self.main(*args, **kwargs)
File "/workspace/env/lib/python3.8/site-packages/click/core.py", line 1062, in main
rv = self.invoke(ctx)
File "/workspace/env/lib/python3.8/site-packages/click/core.py", line 1404, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/workspace/env/lib/python3.8/site-packages/click/core.py", line 763, in invoke
return __callback(*args, **kwargs)
File "/workspace/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 783, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/h/.cache/synthtool/google-cloud-php/AnalyticsData/synth.py", line 124, in <module>
subprocess.run([
File "/usr/local/lib/python3.8/subprocess.py", line 493, in run
with Popen(*popenargs, **kwargs) as process:
File "/usr/local/lib/python3.8/subprocess.py", line 858, in __init__
self._execute_child(args, executable, preexec_fn, close_fds,
File "/usr/local/lib/python3.8/subprocess.py", line 1704, in _execute_child
raise child_exception_type(errno_num, err_msg, err_filename)
FileNotFoundError: [Errno 2] No such file or directory: 'npm'
2021-06-09 10:02:38,990 autosynth [ERROR] > Synthesis failed
2021-06-09 10:02:38,990 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at 74f8fbf4 feat: update to latest AnalyticsAdmin surface (#3825)
2021-06-09 10:02:39,039 autosynth [DEBUG] > Running: git checkout autosynth-analyticsdata
Switched to branch 'autosynth-analyticsdata'
Traceback (most recent call last):
File "/usr/local/lib/python3.8/runpy.py", line 194, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/usr/local/lib/python3.8/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "/workspace/autosynth/synth.py", line 356, in <module>
main()
File "/workspace/autosynth/synth.py", line 191, in main
return _inner_main(temp_dir)
File "/workspace/autosynth/synth.py", line 336, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/workspace/autosynth/synth.py", line 68, in synthesize_loop
has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest)
File "/workspace/autosynth/synth_toolbox.py", line 259, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/workspace/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/usr/local/lib/python3.8/subprocess.py", line 448, in check_returncode
raise CalledProcessError(self.returncode, self.args, self.stdout,
subprocess.CalledProcessError: Command '['/workspace/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/7ff096b1-3667-4d78-a6f9-469ab369b5ab/targets/github%2Fsynthtool;config=default/tests;query=google-cloud-php;failed=false).
| 1.0 | Synthesis failed for analyticsdata - Hello! Autosynth couldn't regenerate analyticsdata. :broken_heart:
Please investigate and fix this issue within 5 business days. While it remains broken,
this library cannot be updated with changes to the analyticsdata API, and the library grows
stale.
See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md
for trouble shooting tips.
Here's the output from running `synth.py`:
```
se.php.
2021-06-09 10:02:38,606 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Cohort.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Cohort.php.
2021-06-09 10:02:38,606 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/RunRealtimeReportResponse.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/RunRealtimeReportResponse.php.
2021-06-09 10:02:38,606 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/BatchRunReportsRequest.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/BatchRunReportsRequest.php.
2021-06-09 10:02:38,606 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/QuotaStatus.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/QuotaStatus.php.
2021-06-09 10:02:38,607 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/PropertyQuota.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/PropertyQuota.php.
2021-06-09 10:02:38,607 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/CohortReportSettings.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/CohortReportSettings.php.
2021-06-09 10:02:38,607 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/MetricHeader.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/MetricHeader.php.
2021-06-09 10:02:38,607 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/BatchRunPivotReportsResponse.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/BatchRunPivotReportsResponse.php.
2021-06-09 10:02:38,608 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/PivotOrderBy.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/PivotOrderBy.php.
2021-06-09 10:02:38,608 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/DimensionOrderBy.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/DimensionOrderBy.php.
2021-06-09 10:02:38,608 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/MetricOrderBy.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/MetricOrderBy.php.
2021-06-09 10:02:38,609 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/PivotOrderBy/PivotSelection.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/OrderBy/PivotOrderBy/PivotSelection.php.
2021-06-09 10:02:38,609 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/InListFilter.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/InListFilter.php.
2021-06-09 10:02:38,609 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/BetweenFilter.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/BetweenFilter.php.
2021-06-09 10:02:38,609 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/NumericFilter.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/NumericFilter.php.
2021-06-09 10:02:38,610 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/StringFilter.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/Filter/StringFilter.php.
2021-06-09 10:02:38,610 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/DimensionExpression/CaseExpression.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/DimensionExpression/CaseExpression.php.
2021-06-09 10:02:38,610 synthtool [INFO] > Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/DimensionExpression/ConcatenateExpression.php.
INFO:synthtool:Replaced 'Generated from protobuf field ([^\\n]{0,})\\n\\s{5}\\*/\\n\\s{4}protected \\$' in src/V1beta/DimensionExpression/ConcatenateExpression.php.
2021-06-09 10:02:38,624 synthtool [WARNING] > No replacements made in src/**/V*/**/*.php for pattern final class, maybe replacement is no longer needed?
WARNING:synthtool:No replacements made in src/**/V*/**/*.php for pattern final class, maybe replacement is no longer needed?
2021-06-09 10:02:38,637 synthtool [WARNING] > No replacements made in src/**/V*/**/*.php for pattern public function ([s|g]\w{3,})Unwrapped, maybe replacement is no longer needed?
WARNING:synthtool:No replacements made in src/**/V*/**/*.php for pattern public function ([s|g]\w{3,})Unwrapped, maybe replacement is no longer needed?
2021-06-09 10:02:38,946 synthtool [WARNING] > No replacements made in src/**/V*/**/*.php for pattern (.{0,})\]\((/.{0,})\), maybe replacement is no longer needed?
WARNING:synthtool:No replacements made in src/**/V*/**/*.php for pattern (.{0,})\]\((/.{0,})\), maybe replacement is no longer needed?
Traceback (most recent call last):
File "/usr/local/lib/python3.8/runpy.py", line 194, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/usr/local/lib/python3.8/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "/workspace/synthtool/__main__.py", line 102, in <module>
main()
File "/workspace/env/lib/python3.8/site-packages/click/core.py", line 1137, in __call__
return self.main(*args, **kwargs)
File "/workspace/env/lib/python3.8/site-packages/click/core.py", line 1062, in main
rv = self.invoke(ctx)
File "/workspace/env/lib/python3.8/site-packages/click/core.py", line 1404, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/workspace/env/lib/python3.8/site-packages/click/core.py", line 763, in invoke
return __callback(*args, **kwargs)
File "/workspace/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 783, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/h/.cache/synthtool/google-cloud-php/AnalyticsData/synth.py", line 124, in <module>
subprocess.run([
File "/usr/local/lib/python3.8/subprocess.py", line 493, in run
with Popen(*popenargs, **kwargs) as process:
File "/usr/local/lib/python3.8/subprocess.py", line 858, in __init__
self._execute_child(args, executable, preexec_fn, close_fds,
File "/usr/local/lib/python3.8/subprocess.py", line 1704, in _execute_child
raise child_exception_type(errno_num, err_msg, err_filename)
FileNotFoundError: [Errno 2] No such file or directory: 'npm'
2021-06-09 10:02:38,990 autosynth [ERROR] > Synthesis failed
2021-06-09 10:02:38,990 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at 74f8fbf4 feat: update to latest AnalyticsAdmin surface (#3825)
2021-06-09 10:02:39,039 autosynth [DEBUG] > Running: git checkout autosynth-analyticsdata
Switched to branch 'autosynth-analyticsdata'
Traceback (most recent call last):
File "/usr/local/lib/python3.8/runpy.py", line 194, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/usr/local/lib/python3.8/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "/workspace/autosynth/synth.py", line 356, in <module>
main()
File "/workspace/autosynth/synth.py", line 191, in main
return _inner_main(temp_dir)
File "/workspace/autosynth/synth.py", line 336, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/workspace/autosynth/synth.py", line 68, in synthesize_loop
has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest)
File "/workspace/autosynth/synth_toolbox.py", line 259, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/workspace/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/usr/local/lib/python3.8/subprocess.py", line 448, in check_returncode
raise CalledProcessError(self.returncode, self.args, self.stdout,
subprocess.CalledProcessError: Command '['/workspace/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/7ff096b1-3667-4d78-a6f9-469ab369b5ab/targets/github%2Fsynthtool;config=default/tests;query=google-cloud-php;failed=false).
| priority | synthesis failed for analyticsdata hello autosynth couldn t regenerate analyticsdata broken heart please investigate and fix this issue within business days while it remains broken this library cannot be updated with changes to the analyticsdata api and the library grows stale see for trouble shooting tips here s the output from running synth py se php synthtool replaced generated from protobuf field n s n s protected in src cohort php info synthtool replaced generated from protobuf field n s n s protected in src cohort php synthtool replaced generated from protobuf field n s n s protected in src runrealtimereportresponse php info synthtool replaced generated from protobuf field n s n s protected in src runrealtimereportresponse php synthtool replaced generated from protobuf field n s n s protected in src batchrunreportsrequest php info synthtool replaced generated from protobuf field n s n s protected in src batchrunreportsrequest php synthtool replaced generated from protobuf field n s n s protected in src quotastatus php info synthtool replaced generated from protobuf field n s n s protected in src quotastatus php synthtool replaced generated from protobuf field n s n s protected in src propertyquota php info synthtool replaced generated from protobuf field n s n s protected in src propertyquota php synthtool replaced generated from protobuf field n s n s protected in src cohortreportsettings php info synthtool replaced generated from protobuf field n s n s protected in src cohortreportsettings php synthtool replaced generated from protobuf field n s n s protected in src metricheader php info synthtool replaced generated from protobuf field n s n s protected in src metricheader php synthtool replaced generated from protobuf field n s n s protected in src batchrunpivotreportsresponse php info synthtool replaced generated from protobuf field n s n s protected in src batchrunpivotreportsresponse php synthtool replaced generated from protobuf field n s n s protected in src orderby pivotorderby php info synthtool replaced generated from protobuf field n s n s protected in src orderby pivotorderby php synthtool replaced generated from protobuf field n s n s protected in src orderby dimensionorderby php info synthtool replaced generated from protobuf field n s n s protected in src orderby dimensionorderby php synthtool replaced generated from protobuf field n s n s protected in src orderby metricorderby php info synthtool replaced generated from protobuf field n s n s protected in src orderby metricorderby php synthtool replaced generated from protobuf field n s n s protected in src orderby pivotorderby pivotselection php info synthtool replaced generated from protobuf field n s n s protected in src orderby pivotorderby pivotselection php synthtool replaced generated from protobuf field n s n s protected in src filter inlistfilter php info synthtool replaced generated from protobuf field n s n s protected in src filter inlistfilter php synthtool replaced generated from protobuf field n s n s protected in src filter betweenfilter php info synthtool replaced generated from protobuf field n s n s protected in src filter betweenfilter php synthtool replaced generated from protobuf field n s n s protected in src filter numericfilter php info synthtool replaced generated from protobuf field n s n s protected in src filter numericfilter php synthtool replaced generated from protobuf field n s n s protected in src filter stringfilter php info synthtool replaced generated from protobuf field n s n s protected in src filter stringfilter php synthtool replaced generated from protobuf field n s n s protected in src dimensionexpression caseexpression php info synthtool replaced generated from protobuf field n s n s protected in src dimensionexpression caseexpression php synthtool replaced generated from protobuf field n s n s protected in src dimensionexpression concatenateexpression php info synthtool replaced generated from protobuf field n s n s protected in src dimensionexpression concatenateexpression php synthtool no replacements made in src v php for pattern final class maybe replacement is no longer needed warning synthtool no replacements made in src v php for pattern final class maybe replacement is no longer needed synthtool no replacements made in src v php for pattern public function w unwrapped maybe replacement is no longer needed warning synthtool no replacements made in src v php for pattern public function w unwrapped maybe replacement is no longer needed synthtool no replacements made in src v php for pattern maybe replacement is no longer needed warning synthtool no replacements made in src v php for pattern maybe replacement is no longer needed traceback most recent call last file usr local lib runpy py line in run module as main return run code code main globals none file usr local lib runpy py line in run code exec code run globals file workspace synthtool main py line in main file workspace env lib site packages click core py line in call return self main args kwargs file workspace env lib site packages click core py line in main rv self invoke ctx file workspace env lib site packages click core py line in invoke return ctx invoke self callback ctx params file workspace env lib site packages click core py line in invoke return callback args kwargs file workspace synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file h cache synthtool google cloud php analyticsdata synth py line in subprocess run file usr local lib subprocess py line in run with popen popenargs kwargs as process file usr local lib subprocess py line in init self execute child args executable preexec fn close fds file usr local lib subprocess py line in execute child raise child exception type errno num err msg err filename filenotfounderror no such file or directory npm autosynth synthesis failed autosynth running git reset hard head head is now at feat update to latest analyticsadmin surface autosynth running git checkout autosynth analyticsdata switched to branch autosynth analyticsdata traceback most recent call last file usr local lib runpy py line in run module as main return run code code main globals none file usr local lib runpy py line in run code exec code run globals file workspace autosynth synth py line in main file workspace autosynth synth py line in main return inner main temp dir file workspace autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file workspace autosynth synth py line in synthesize loop has changes toolbox synthesize version in new branch synthesizer youngest file workspace autosynth synth toolbox py line in synthesize version in new branch synthesizer synthesize synth log path self environ file workspace autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file usr local lib subprocess py line in check returncode raise calledprocesserror self returncode self args self stdout subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log | 1 |
585,226 | 17,483,157,202 | IssuesEvent | 2021-08-09 07:21:05 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | sniffies.com - Unable to get location | browser-firefox-mobile nsfw priority-normal severity-critical browser-focus-geckoview engine-gecko | <!-- @browser: Firefox Mobile 90.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:90.0) Gecko/90.0 Firefox/90.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/81320 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://sniffies.com/
**Browser / Version**: Firefox Mobile 90.0
**Operating System**: Android 11
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
It just doesn't work on this browser. It says it can't get my location.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | sniffies.com - Unable to get location - <!-- @browser: Firefox Mobile 90.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:90.0) Gecko/90.0 Firefox/90.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/81320 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://sniffies.com/
**Browser / Version**: Firefox Mobile 90.0
**Operating System**: Android 11
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
It just doesn't work on this browser. It says it can't get my location.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | sniffies com unable to get location url browser version firefox mobile operating system android tested another browser no problem type site is not usable description page not loading correctly steps to reproduce it just doesn t work on this browser it says it can t get my location browser configuration none from with ❤️ | 1 |
642,565 | 20,907,317,690 | IssuesEvent | 2022-03-24 04:39:31 | mikezimm/drilldown7 | https://api.github.com/repos/mikezimm/drilldown7 | closed | Bug: Crash when using refiner Codigo in Espanol | High Priority complete | ## Location: /sites/SolutionTesting/SitePages/CH-Drilldown-2.aspx
Getting the 2 errors in the attached screenshot.
It doesn't seem to be a fetch error because it is fetching 146 items.
But is throwing both an undefined an null error.
Similar item to #79 and #80. #79 was resolved which was adding this column to a view.
But this is adding it to a refiner.
## Here are the 2 errors:

## Here's an item showing it is getting the C_x00f3_digo column

| 1.0 | Bug: Crash when using refiner Codigo in Espanol - ## Location: /sites/SolutionTesting/SitePages/CH-Drilldown-2.aspx
Getting the 2 errors in the attached screenshot.
It doesn't seem to be a fetch error because it is fetching 146 items.
But is throwing both an undefined an null error.
Similar item to #79 and #80. #79 was resolved which was adding this column to a view.
But this is adding it to a refiner.
## Here are the 2 errors:

## Here's an item showing it is getting the C_x00f3_digo column

| priority | bug crash when using refiner codigo in espanol location sites solutiontesting sitepages ch drilldown aspx getting the errors in the attached screenshot it doesn t seem to be a fetch error because it is fetching items but is throwing both an undefined an null error similar item to and was resolved which was adding this column to a view but this is adding it to a refiner here are the errors here s an item showing it is getting the c digo column | 1 |
649,721 | 21,318,168,180 | IssuesEvent | 2022-04-16 16:51:03 | kubesphere/kubesphere | https://api.github.com/repos/kubesphere/kubesphere | closed | After service mesh of cluster is enabled, the application governance status of the application is wrong | kind/bug stale priority/medium | **Versions Used**
KubeSphere: `v3.2.1-rc.3`
**Precondition**
```
servicemesh:
enabled: false
```
**How To Reproduce**
Steps to reproduce the behavior:
1. Create application 'test' and select close of Application governance
2. After the application ‘test’ is successfully created

3. Enabled servicemesh of cluster
4. See resource status page of application 'test'

**Actual behavior**
The status of application governance will become 'On'
/kind bug
/priority medium
| 1.0 | After service mesh of cluster is enabled, the application governance status of the application is wrong - **Versions Used**
KubeSphere: `v3.2.1-rc.3`
**Precondition**
```
servicemesh:
enabled: false
```
**How To Reproduce**
Steps to reproduce the behavior:
1. Create application 'test' and select close of Application governance
2. After the application ‘test’ is successfully created

3. Enabled servicemesh of cluster
4. See resource status page of application 'test'

**Actual behavior**
The status of application governance will become 'On'
/kind bug
/priority medium
| priority | after service mesh of cluster is enabled the application governance status of the application is wrong versions used kubesphere rc precondition servicemesh enabled false how to reproduce steps to reproduce the behavior create application test and select close of application governance after the application ‘test’ is successfully created enabled servicemesh of cluster see resource status page of application test actual behavior the status of application governance will become on kind bug priority medium | 1 |
62,063 | 25,850,988,442 | IssuesEvent | 2022-12-13 10:22:51 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | EnableWorkloadIdentityPreview parameter is register but not showing on console | container-service/svc triaged cxp product-question Pri2 |
[Enter feedback here]
I have register EnableWorkloadIdentityPreview, but it is not showing on console in preview feature. Ideally, any preview feature should atleast view from console.
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: efde8151-ca72-0ce1-271a-57a14a8c7ccd
* Version Independent ID: f37efaf0-6573-6813-a3e4-3065e44666e4
* Content: [Tutorial - Use a workload identity with an application on Azure Kubernetes Service (AKS) - Azure Kubernetes Service](https://learn.microsoft.com/en-us/azure/aks/learn/tutorial-kubernetes-workload-identity)
* Content Source: [articles/aks/learn/tutorial-kubernetes-workload-identity.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/aks/learn/tutorial-kubernetes-workload-identity.md)
* Service: **container-service**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte** | 1.0 | EnableWorkloadIdentityPreview parameter is register but not showing on console -
[Enter feedback here]
I have register EnableWorkloadIdentityPreview, but it is not showing on console in preview feature. Ideally, any preview feature should atleast view from console.
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: efde8151-ca72-0ce1-271a-57a14a8c7ccd
* Version Independent ID: f37efaf0-6573-6813-a3e4-3065e44666e4
* Content: [Tutorial - Use a workload identity with an application on Azure Kubernetes Service (AKS) - Azure Kubernetes Service](https://learn.microsoft.com/en-us/azure/aks/learn/tutorial-kubernetes-workload-identity)
* Content Source: [articles/aks/learn/tutorial-kubernetes-workload-identity.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/aks/learn/tutorial-kubernetes-workload-identity.md)
* Service: **container-service**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte** | non_priority | enableworkloadidentitypreview parameter is register but not showing on console i have register enableworkloadidentitypreview but it is not showing on console in preview feature ideally any preview feature should atleast view from console document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service container service github login mgoedtel microsoft alias magoedte | 0 |
21,916 | 2,642,876,064 | IssuesEvent | 2015-03-12 05:20:23 | cs2103jan2015-t15-3j/main | https://api.github.com/repos/cs2103jan2015-t15-3j/main | closed | Allow user to be able to key in input into the textfield | priority.high | ... and able to press the 'Enter' button successfully. | 1.0 | Allow user to be able to key in input into the textfield - ... and able to press the 'Enter' button successfully. | priority | allow user to be able to key in input into the textfield and able to press the enter button successfully | 1 |
393,171 | 11,611,355,857 | IssuesEvent | 2020-02-26 06:18:13 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | reopened | www.pontofrio.com.br - Some page elements are missing | browser-fenix engine-gecko priority-normal severity-critical type-uaoverride | <!-- @browser: Firefox Mobile 73.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:73.0) Gecko/73.0 Firefox/73.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/48755 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.pontofrio.com.br/acessorioseinovacoes/Monitores/dell-monitor-alienware-de-34-quot-curvo-para-jogos-aw3418hw-preto-12732406.html?IdSku=12732406
**Browser / Version**: Firefox Mobile 73.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: site broken.
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.pontofrio.com.br - Some page elements are missing - <!-- @browser: Firefox Mobile 73.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:73.0) Gecko/73.0 Firefox/73.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/48755 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.pontofrio.com.br/acessorioseinovacoes/Monitores/dell-monitor-alienware-de-34-quot-curvo-para-jogos-aw3418hw-preto-12732406.html?IdSku=12732406
**Browser / Version**: Firefox Mobile 73.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: site broken.
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | some page elements are missing url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description site broken steps to reproduce browser configuration none from with ❤️ | 1 |
187,731 | 22,045,873,516 | IssuesEvent | 2022-05-30 01:35:25 | Nivaskumark/kernel_v4.1.15 | https://api.github.com/repos/Nivaskumark/kernel_v4.1.15 | closed | CVE-2017-16530 (Medium) detected in linuxlinux-4.6, linuxlinux-4.6 - autoclosed | security vulnerability | ## CVE-2017-16530 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.6</b>, <b>linuxlinux-4.6</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The uas driver in the Linux kernel before 4.13.6 allows local users to cause a denial of service (out-of-bounds read and system crash) or possibly have unspecified other impact via a crafted USB device, related to drivers/usb/storage/uas-detect.h and drivers/usb/storage/uas.c.
<p>Publish Date: 2017-11-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16530>CVE-2017-16530</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-16530">https://nvd.nist.gov/vuln/detail/CVE-2017-16530</a></p>
<p>Release Date: 2017-11-04</p>
<p>Fix Resolution: 4.13.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-16530 (Medium) detected in linuxlinux-4.6, linuxlinux-4.6 - autoclosed - ## CVE-2017-16530 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.6</b>, <b>linuxlinux-4.6</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The uas driver in the Linux kernel before 4.13.6 allows local users to cause a denial of service (out-of-bounds read and system crash) or possibly have unspecified other impact via a crafted USB device, related to drivers/usb/storage/uas-detect.h and drivers/usb/storage/uas.c.
<p>Publish Date: 2017-11-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16530>CVE-2017-16530</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-16530">https://nvd.nist.gov/vuln/detail/CVE-2017-16530</a></p>
<p>Release Date: 2017-11-04</p>
<p>Fix Resolution: 4.13.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linuxlinux linuxlinux autoclosed cve medium severity vulnerability vulnerable libraries linuxlinux linuxlinux vulnerability details the uas driver in the linux kernel before allows local users to cause a denial of service out of bounds read and system crash or possibly have unspecified other impact via a crafted usb device related to drivers usb storage uas detect h and drivers usb storage uas c publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
322,390 | 9,817,051,992 | IssuesEvent | 2019-06-13 15:51:31 | dice-group/Squirrel | https://api.github.com/repos/dice-group/Squirrel | closed | Improve seed file reader | component:frontier priority:high type:enhancement | ## Description
As a user, I would like to provide additional information for seed files, e.g., the type of a URI or some other data that could improve its processing. Therefore, the reader for the seed file should be improved to be able to process this additional information.
## Solution
* To keep the solution general, it would be good if the reader would accept a CSV file with a head line which defines the keys for the values. In the following example `uri` is used to mark the column with the URIs while the other columns contain additional data which should be added to the URIs object.
```
"uri","key1","key2"
"http://example1.org","value1-1","value1-2"
"http://example2.org","","value2-2"
"http://example3.org",,
```
* Note that `""` should be used as an empty value, i.e., the key should be added with an empty string as value.
* Note that the key value pair (or the key with an empty value) should only be added if the cell is _not_ empty. The line of `"http://example3.org"` contains two empty cells which should make sure that both keys are not added to this URI. | 1.0 | Improve seed file reader - ## Description
As a user, I would like to provide additional information for seed files, e.g., the type of a URI or some other data that could improve its processing. Therefore, the reader for the seed file should be improved to be able to process this additional information.
## Solution
* To keep the solution general, it would be good if the reader would accept a CSV file with a head line which defines the keys for the values. In the following example `uri` is used to mark the column with the URIs while the other columns contain additional data which should be added to the URIs object.
```
"uri","key1","key2"
"http://example1.org","value1-1","value1-2"
"http://example2.org","","value2-2"
"http://example3.org",,
```
* Note that `""` should be used as an empty value, i.e., the key should be added with an empty string as value.
* Note that the key value pair (or the key with an empty value) should only be added if the cell is _not_ empty. The line of `"http://example3.org"` contains two empty cells which should make sure that both keys are not added to this URI. | priority | improve seed file reader description as a user i would like to provide additional information for seed files e g the type of a uri or some other data that could improve its processing therefore the reader for the seed file should be improved to be able to process this additional information solution to keep the solution general it would be good if the reader would accept a csv file with a head line which defines the keys for the values in the following example uri is used to mark the column with the uris while the other columns contain additional data which should be added to the uris object uri note that should be used as an empty value i e the key should be added with an empty string as value note that the key value pair or the key with an empty value should only be added if the cell is not empty the line of contains two empty cells which should make sure that both keys are not added to this uri | 1 |
15,611 | 10,325,516,963 | IssuesEvent | 2019-09-01 18:00:38 | microsoft/BotBuilder-Samples | https://api.github.com/repos/microsoft/BotBuilder-Samples | closed | OAuthPrompt popup is not showing for login in MS-Teams. | Bot Services customer-replied-to customer-reported | I have use OAuthPrompt card to login user for bot authentication. It is working in Emulator but when i test in Teams then pop-up is not showing to login.
sample code:
==========
return await stepContext.BeginDialogAsync(nameof(OAuthPrompt), null, cancellationToken);
Bot Framework version: 4.5.1
| 1.0 | OAuthPrompt popup is not showing for login in MS-Teams. - I have use OAuthPrompt card to login user for bot authentication. It is working in Emulator but when i test in Teams then pop-up is not showing to login.
sample code:
==========
return await stepContext.BeginDialogAsync(nameof(OAuthPrompt), null, cancellationToken);
Bot Framework version: 4.5.1
| non_priority | oauthprompt popup is not showing for login in ms teams i have use oauthprompt card to login user for bot authentication it is working in emulator but when i test in teams then pop up is not showing to login sample code return await stepcontext begindialogasync nameof oauthprompt null cancellationtoken bot framework version | 0 |
162,296 | 6,150,306,351 | IssuesEvent | 2017-06-27 22:11:49 | geoff-maddock/events-tracker | https://api.github.com/repos/geoff-maddock/events-tracker | closed | Search > Add tag summary as result | low priority | When a user searches for a term, if it matches a tag, display a tag summary
Include
- name
- button to follow (or unfollow)
- number of related events / entities? | 1.0 | Search > Add tag summary as result - When a user searches for a term, if it matches a tag, display a tag summary
Include
- name
- button to follow (or unfollow)
- number of related events / entities? | priority | search add tag summary as result when a user searches for a term if it matches a tag display a tag summary include name button to follow or unfollow number of related events entities | 1 |
451,501 | 32,030,631,594 | IssuesEvent | 2023-09-22 12:06:01 | swedenconnect/bankid-saml-idp | https://api.github.com/repos/swedenconnect/bankid-saml-idp | closed | Add documentation about how to join Sweden Connect federation | documentation | Add documentation about how to join Sweden Connect federation. | 1.0 | Add documentation about how to join Sweden Connect federation - Add documentation about how to join Sweden Connect federation. | non_priority | add documentation about how to join sweden connect federation add documentation about how to join sweden connect federation | 0 |
272,954 | 8,519,556,788 | IssuesEvent | 2018-11-01 15:00:15 | desktop/desktop | https://api.github.com/repos/desktop/desktop | closed | Aborting a new merge doesn't update the changed files list | bug priority-2 | ## Description
After aborting a merge, the status doesn't update. I continue to see all the conflicted files, etc. If I check `git status` on the command line, it reflects that I aborted the merge. If I switch away from the app and back, the changed files updates as I'd expect.
## Version
* GitHub Desktop: 1.5.0-beta2
* Operating system: macOS 10.13.6
## Steps to Reproduce
1. Merge two branches that introduce a conflict.
1. Abort the merge using the ✨ fancy new dialog ✨
### Expected Behavior
1. All the changed files and conflicts are gone.
### Actual Behavior
1. I still see the changed files and conflicts. | 1.0 | Aborting a new merge doesn't update the changed files list - ## Description
After aborting a merge, the status doesn't update. I continue to see all the conflicted files, etc. If I check `git status` on the command line, it reflects that I aborted the merge. If I switch away from the app and back, the changed files updates as I'd expect.
## Version
* GitHub Desktop: 1.5.0-beta2
* Operating system: macOS 10.13.6
## Steps to Reproduce
1. Merge two branches that introduce a conflict.
1. Abort the merge using the ✨ fancy new dialog ✨
### Expected Behavior
1. All the changed files and conflicts are gone.
### Actual Behavior
1. I still see the changed files and conflicts. | priority | aborting a new merge doesn t update the changed files list description after aborting a merge the status doesn t update i continue to see all the conflicted files etc if i check git status on the command line it reflects that i aborted the merge if i switch away from the app and back the changed files updates as i d expect version github desktop operating system macos steps to reproduce merge two branches that introduce a conflict abort the merge using the ✨ fancy new dialog ✨ expected behavior all the changed files and conflicts are gone actual behavior i still see the changed files and conflicts | 1 |
157,183 | 5,996,375,119 | IssuesEvent | 2017-06-03 13:44:27 | TASVideos/BizHawk | https://api.github.com/repos/TASVideos/BizHawk | closed | Disconnect SNES Controllers | auto-migrated Core-BSNES Priority-Medium Type-Enhancement | ```
What steps will reproduce the problem?
1. Play Tiny Toon Adventures Wacky Sports Challenge
2. Select Start Game
3. You will need to push the A button on Player 2 to start playing
What is the expected output? What do you see instead?
I would like a way to turn off SNES Controllers that I don't want/need.
What version of the product are you using? On what operating system?
SVN Build r7446
Please provide any additional information below.
This applies to all SNES Cores only.
```
Original issue reported on code.google.com by `hegyak` on 29 Jul 2014 at 1:10
| 1.0 | Disconnect SNES Controllers - ```
What steps will reproduce the problem?
1. Play Tiny Toon Adventures Wacky Sports Challenge
2. Select Start Game
3. You will need to push the A button on Player 2 to start playing
What is the expected output? What do you see instead?
I would like a way to turn off SNES Controllers that I don't want/need.
What version of the product are you using? On what operating system?
SVN Build r7446
Please provide any additional information below.
This applies to all SNES Cores only.
```
Original issue reported on code.google.com by `hegyak` on 29 Jul 2014 at 1:10
| priority | disconnect snes controllers what steps will reproduce the problem play tiny toon adventures wacky sports challenge select start game you will need to push the a button on player to start playing what is the expected output what do you see instead i would like a way to turn off snes controllers that i don t want need what version of the product are you using on what operating system svn build please provide any additional information below this applies to all snes cores only original issue reported on code google com by hegyak on jul at | 1 |
125,051 | 17,795,927,974 | IssuesEvent | 2021-08-31 22:13:07 | ghc-dev/Ashley-Campos | https://api.github.com/repos/ghc-dev/Ashley-Campos | opened | CVE-2017-1000048 (High) detected in qs-4.0.0.tgz | security vulnerability | ## CVE-2017-1000048 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-4.0.0.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-4.0.0.tgz">https://registry.npmjs.org/qs/-/qs-4.0.0.tgz</a></p>
<p>Path to dependency file: Ashley-Campos/package.json</p>
<p>Path to vulnerable library: Ashley-Campos/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- body-parser-1.13.3.tgz (Root Library)
- :x: **qs-4.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Ashley-Campos/commit/93bde7a441bf435e2b02a4e5c2c7e6e2e2ebb8e0">93bde7a441bf435e2b02a4e5c2c7e6e2e2ebb8e0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.
<p>Publish Date: 2017-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048</a></p>
<p>Release Date: 2017-07-17</p>
<p>Fix Resolution: qs - 6.0.4,6.1.2,6.2.3,6.3.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"4.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"body-parser:1.13.3;qs:4.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-1000048","vulnerabilityDetails":"the web framework using ljharb\u0027s qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2017-1000048 (High) detected in qs-4.0.0.tgz - ## CVE-2017-1000048 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-4.0.0.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-4.0.0.tgz">https://registry.npmjs.org/qs/-/qs-4.0.0.tgz</a></p>
<p>Path to dependency file: Ashley-Campos/package.json</p>
<p>Path to vulnerable library: Ashley-Campos/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- body-parser-1.13.3.tgz (Root Library)
- :x: **qs-4.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Ashley-Campos/commit/93bde7a441bf435e2b02a4e5c2c7e6e2e2ebb8e0">93bde7a441bf435e2b02a4e5c2c7e6e2e2ebb8e0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.
<p>Publish Date: 2017-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048</a></p>
<p>Release Date: 2017-07-17</p>
<p>Fix Resolution: qs - 6.0.4,6.1.2,6.2.3,6.3.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"4.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"body-parser:1.13.3;qs:4.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-1000048","vulnerabilityDetails":"the web framework using ljharb\u0027s qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in qs tgz cve high severity vulnerability vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file ashley campos package json path to vulnerable library ashley campos node modules qs package json dependency hierarchy body parser tgz root library x qs tgz vulnerable library found in head commit a href found in base branch master vulnerability details the web framework using ljharb s qs module older than and is vulnerable to a dos a malicious user can send a evil request to cause the web framework crash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree body parser qs isminimumfixversionavailable true minimumfixversion qs basebranches vulnerabilityidentifier cve vulnerabilitydetails the web framework using ljharb qs module older than and is vulnerable to a dos a malicious user can send a evil request to cause the web framework crash vulnerabilityurl | 0 |
16,609 | 2,615,119,834 | IssuesEvent | 2015-03-01 05:45:54 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | closed | Gzipped Request Content Unsupported by Some Servers | auto-migrated Component-HTTP Milestone-Version1.4.0 Priority-Medium Type-Enhancement | ```
Version of google-api-java-client (e.g. 1.3.1-alpha)?
1.4.0-SNAPSHOT
Java environment (e.g. Java 6, Android 2.3, App Engine 1.4.2)?
Java 6
Describe the problem.
The library currently gzips all content > 256 bytes. There is no guarantee
that all servers will accept gzipped content, nor any mechanism to negotiate
this with the server automatically.
How would you expect it to be fixed?
The library should conservatively not use gzip encoding and allow the user to
turn it on as an option. An optimization would be for the library to use it by
default on servers that it knows support gzip, although this may be difficult
to generalize.
```
Original issue reported on code.google.com by `moshe...@google.com` on 12 Apr 2011 at 7:04 | 1.0 | Gzipped Request Content Unsupported by Some Servers - ```
Version of google-api-java-client (e.g. 1.3.1-alpha)?
1.4.0-SNAPSHOT
Java environment (e.g. Java 6, Android 2.3, App Engine 1.4.2)?
Java 6
Describe the problem.
The library currently gzips all content > 256 bytes. There is no guarantee
that all servers will accept gzipped content, nor any mechanism to negotiate
this with the server automatically.
How would you expect it to be fixed?
The library should conservatively not use gzip encoding and allow the user to
turn it on as an option. An optimization would be for the library to use it by
default on servers that it knows support gzip, although this may be difficult
to generalize.
```
Original issue reported on code.google.com by `moshe...@google.com` on 12 Apr 2011 at 7:04 | priority | gzipped request content unsupported by some servers version of google api java client e g alpha snapshot java environment e g java android app engine java describe the problem the library currently gzips all content bytes there is no guarantee that all servers will accept gzipped content nor any mechanism to negotiate this with the server automatically how would you expect it to be fixed the library should conservatively not use gzip encoding and allow the user to turn it on as an option an optimization would be for the library to use it by default on servers that it knows support gzip although this may be difficult to generalize original issue reported on code google com by moshe google com on apr at | 1 |
489,542 | 14,107,904,085 | IssuesEvent | 2020-11-06 16:58:19 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | samples: net: sockets: tcp: tcp2 server not accepting with ipv6 bsd sockets | area: Networking bug priority: medium | **Describe the bug**
Prior to 70dae094ba2dd3e1f02684a8d673d434966b8b76 it was possible to run an IPv6 TCP server using the default TCP implementation and the BSD sockets API over IEEE 802.15.4. However, since TCP2 became the default TCP implementation, it has not been possible (with the same configuration).
Furthermore, it is unclear if and when it was actually possible to do so (with the same configuration). I went back as far as I could compile in a `git bisect` with TCP2 enabled by default and this configuration has never seemed to work.
It's not clear to me if the problem only affects applications that use the BSD sockets API, but that is how I noticed.
**To Reproduce**
Steps to reproduce the behavior:
1. apply the [attached patch](https://github.com/zephyrproject-rtos/zephyr/files/5485176/ieee802154-ipv6.patch.txt) for the `samples/net/sockets/echo` application (adds IPv6 / 15.4)
2. select a board that is supports IEEE 802.15.4 (e.g. nrf52840dk_nrf52840)
3. `west build -b `nrf52840dk_nrf52840` -t flash samples/net/sockets/echo -- -DOVERLAY_CONFIG=overlay-802154.conf`
4. telnet to `2001:db8::1` and try to see if echo works (it does not even connect for me.
5. adjust samples/net/sockets/echo/overlay-802154.conf and swap the values for `CONFIG_NET_TCP1=n` and `CONFIG_NET_TCP2=y` to enable `TCP1`.
6. repeat from step 3
I've tried this using `nrf52840dk_nrf52840`, `cc1352r1_launchxl`, and the `cc1352r_sensortag` using IEEE 802.15.4. I've not tried with BLE / 6lowpan yet, but would be surprised if it were unaffected.
**Expected behavior**
The server should echo each line of text back to the client.
**Impact**
While it's possible to explicitly select `CONFIG_NET_TCP1`, it's probably something that should be fixed for TCP2 given that it's the default.
**Logs and console output**
With TCP2 (`telnet` never connected):
```bash
$ ping6 2001:db8::1
PING 2001:db8::1(2001:db8::1) 56 data bytes
64 bytes from 2001:db8::1: icmp_seq=1 ttl=64 time=27.0 ms
64 bytes from 2001:db8::1: icmp_seq=2 ttl=64 time=16.1 ms
64 bytes from 2001:db8::1: icmp_seq=3 ttl=64 time=13.8 ms
^C
--- 2001:db8::1 ping statistics ---
3 packets transmitted, 3 received, 0% packet loss, time 2003ms
rtt min/avg/max/mdev = 13.807/18.974/27.021/5.766 ms
cfriedt@cfriedt-MacBookPro:~/workspace/zephyrproject/zephyr$ telnet 2001:db8::1 4242
Trying 2001:db8::1...
^C
```
With TCP1:
```bash
$ ping6 2001:db8::1
PING 2001:db8::1(2001:db8::1) 56 data bytes
64 bytes from 2001:db8::1: icmp_seq=1 ttl=64 time=12.4 ms
64 bytes from 2001:db8::1: icmp_seq=2 ttl=64 time=13.2 ms
^C
--- 2001:db8::1 ping statistics ---
2 packets transmitted, 2 received, 0% packet loss, time 1002ms
rtt min/avg/max/mdev = 12.412/12.802/13.193/0.390 ms
cfriedt@cfriedt-MacBookPro:~/workspace/zephyrproject/zephyr$ telnet 2001:db8::1 4242
Trying 2001:db8::1...
Connected to 2001:db8::1.
Escape character is '^]'.
hello?
hello?
```
**Environment (please complete the following information):**
- OS: Ubuntu Linux (Focal Fossa)
- Toolchain: Zephyr SDK 0.11.3
- Commit SHA: 761f1c4bedc4fc617920949c168f75b9d148badc
**Additional context**
Seems to be related to #28577
cc @jukkar @MaureenHelm | 1.0 | samples: net: sockets: tcp: tcp2 server not accepting with ipv6 bsd sockets - **Describe the bug**
Prior to 70dae094ba2dd3e1f02684a8d673d434966b8b76 it was possible to run an IPv6 TCP server using the default TCP implementation and the BSD sockets API over IEEE 802.15.4. However, since TCP2 became the default TCP implementation, it has not been possible (with the same configuration).
Furthermore, it is unclear if and when it was actually possible to do so (with the same configuration). I went back as far as I could compile in a `git bisect` with TCP2 enabled by default and this configuration has never seemed to work.
It's not clear to me if the problem only affects applications that use the BSD sockets API, but that is how I noticed.
**To Reproduce**
Steps to reproduce the behavior:
1. apply the [attached patch](https://github.com/zephyrproject-rtos/zephyr/files/5485176/ieee802154-ipv6.patch.txt) for the `samples/net/sockets/echo` application (adds IPv6 / 15.4)
2. select a board that is supports IEEE 802.15.4 (e.g. nrf52840dk_nrf52840)
3. `west build -b `nrf52840dk_nrf52840` -t flash samples/net/sockets/echo -- -DOVERLAY_CONFIG=overlay-802154.conf`
4. telnet to `2001:db8::1` and try to see if echo works (it does not even connect for me.
5. adjust samples/net/sockets/echo/overlay-802154.conf and swap the values for `CONFIG_NET_TCP1=n` and `CONFIG_NET_TCP2=y` to enable `TCP1`.
6. repeat from step 3
I've tried this using `nrf52840dk_nrf52840`, `cc1352r1_launchxl`, and the `cc1352r_sensortag` using IEEE 802.15.4. I've not tried with BLE / 6lowpan yet, but would be surprised if it were unaffected.
**Expected behavior**
The server should echo each line of text back to the client.
**Impact**
While it's possible to explicitly select `CONFIG_NET_TCP1`, it's probably something that should be fixed for TCP2 given that it's the default.
**Logs and console output**
With TCP2 (`telnet` never connected):
```bash
$ ping6 2001:db8::1
PING 2001:db8::1(2001:db8::1) 56 data bytes
64 bytes from 2001:db8::1: icmp_seq=1 ttl=64 time=27.0 ms
64 bytes from 2001:db8::1: icmp_seq=2 ttl=64 time=16.1 ms
64 bytes from 2001:db8::1: icmp_seq=3 ttl=64 time=13.8 ms
^C
--- 2001:db8::1 ping statistics ---
3 packets transmitted, 3 received, 0% packet loss, time 2003ms
rtt min/avg/max/mdev = 13.807/18.974/27.021/5.766 ms
cfriedt@cfriedt-MacBookPro:~/workspace/zephyrproject/zephyr$ telnet 2001:db8::1 4242
Trying 2001:db8::1...
^C
```
With TCP1:
```bash
$ ping6 2001:db8::1
PING 2001:db8::1(2001:db8::1) 56 data bytes
64 bytes from 2001:db8::1: icmp_seq=1 ttl=64 time=12.4 ms
64 bytes from 2001:db8::1: icmp_seq=2 ttl=64 time=13.2 ms
^C
--- 2001:db8::1 ping statistics ---
2 packets transmitted, 2 received, 0% packet loss, time 1002ms
rtt min/avg/max/mdev = 12.412/12.802/13.193/0.390 ms
cfriedt@cfriedt-MacBookPro:~/workspace/zephyrproject/zephyr$ telnet 2001:db8::1 4242
Trying 2001:db8::1...
Connected to 2001:db8::1.
Escape character is '^]'.
hello?
hello?
```
**Environment (please complete the following information):**
- OS: Ubuntu Linux (Focal Fossa)
- Toolchain: Zephyr SDK 0.11.3
- Commit SHA: 761f1c4bedc4fc617920949c168f75b9d148badc
**Additional context**
Seems to be related to #28577
cc @jukkar @MaureenHelm | priority | samples net sockets tcp server not accepting with bsd sockets describe the bug prior to it was possible to run an tcp server using the default tcp implementation and the bsd sockets api over ieee however since became the default tcp implementation it has not been possible with the same configuration furthermore it is unclear if and when it was actually possible to do so with the same configuration i went back as far as i could compile in a git bisect with enabled by default and this configuration has never seemed to work it s not clear to me if the problem only affects applications that use the bsd sockets api but that is how i noticed to reproduce steps to reproduce the behavior apply the for the samples net sockets echo application adds select a board that is supports ieee e g west build b t flash samples net sockets echo doverlay config overlay conf telnet to and try to see if echo works it does not even connect for me adjust samples net sockets echo overlay conf and swap the values for config net n and config net y to enable repeat from step i ve tried this using launchxl and the sensortag using ieee i ve not tried with ble yet but would be surprised if it were unaffected expected behavior the server should echo each line of text back to the client impact while it s possible to explicitly select config net it s probably something that should be fixed for given that it s the default logs and console output with telnet never connected bash ping data bytes bytes from icmp seq ttl time ms bytes from icmp seq ttl time ms bytes from icmp seq ttl time ms c ping statistics packets transmitted received packet loss time rtt min avg max mdev ms cfriedt cfriedt macbookpro workspace zephyrproject zephyr telnet trying c with bash ping data bytes bytes from icmp seq ttl time ms bytes from icmp seq ttl time ms c ping statistics packets transmitted received packet loss time rtt min avg max mdev ms cfriedt cfriedt macbookpro workspace zephyrproject zephyr telnet trying connected to escape character is hello hello environment please complete the following information os ubuntu linux focal fossa toolchain zephyr sdk commit sha additional context seems to be related to cc jukkar maureenhelm | 1 |
322,565 | 9,819,497,695 | IssuesEvent | 2019-06-13 22:13:38 | clearlinux/distribution | https://api.github.com/repos/clearlinux/distribution | closed | xdg-open (and other XDG tools) missing | bug medium priority | **Describe the bug**
```
$ xdg-open
bash: xdg-open: command not found
```
Many scripts assume the XDG tools to be present on any desktop system.
**To Reproduce**
Steps to reproduce the behavior:
1. Install the XFCE desktop
2. open `xfce-terminal`
3. open `xdg-open`
4. Try to close it
**Expected behavior**
`xdg-open` should be expected to be present on every desktop system.
**Environment (please complete the following information):**
```
user@clr-... ~ $ cat /usr/lib/os-release
NAME="Clear Linux OS"
VERSION=1
ID=clear-linux-os
ID_LIKE=clear-linux-os
VERSION_ID=26200
PRETTY_NAME="Clear Linux OS"
ANSI_COLOR="1;35"
HOME_URL="https://clearlinux.org"
SUPPORT_URL="https://clearlinux.org"
BUG_REPORT_URL="mailto:dev@lists.clearlinux.org"
PRIVACY_POLICY_URL="http://www.intel.com/privacy"
user@clr-... ~ $ sudo swupd bundle-list
Password:
bootloader
desktop
desktop-apps
desktop-assets
desktop-autostart
desktop-gnomelibs
desktop-locales
editors
emacs
ethtool
iproute2
joe
kernel-native
libX11client
network-basic
openssh-server
openssl
os-core
os-core-update
p11-kit
perl-basic
python3-basic
sudo
sysadmin-basic
vim
which
wpa_supplicant
x11-server
xemacs
xfce4-desktop
``` | 1.0 | xdg-open (and other XDG tools) missing - **Describe the bug**
```
$ xdg-open
bash: xdg-open: command not found
```
Many scripts assume the XDG tools to be present on any desktop system.
**To Reproduce**
Steps to reproduce the behavior:
1. Install the XFCE desktop
2. open `xfce-terminal`
3. open `xdg-open`
4. Try to close it
**Expected behavior**
`xdg-open` should be expected to be present on every desktop system.
**Environment (please complete the following information):**
```
user@clr-... ~ $ cat /usr/lib/os-release
NAME="Clear Linux OS"
VERSION=1
ID=clear-linux-os
ID_LIKE=clear-linux-os
VERSION_ID=26200
PRETTY_NAME="Clear Linux OS"
ANSI_COLOR="1;35"
HOME_URL="https://clearlinux.org"
SUPPORT_URL="https://clearlinux.org"
BUG_REPORT_URL="mailto:dev@lists.clearlinux.org"
PRIVACY_POLICY_URL="http://www.intel.com/privacy"
user@clr-... ~ $ sudo swupd bundle-list
Password:
bootloader
desktop
desktop-apps
desktop-assets
desktop-autostart
desktop-gnomelibs
desktop-locales
editors
emacs
ethtool
iproute2
joe
kernel-native
libX11client
network-basic
openssh-server
openssl
os-core
os-core-update
p11-kit
perl-basic
python3-basic
sudo
sysadmin-basic
vim
which
wpa_supplicant
x11-server
xemacs
xfce4-desktop
``` | priority | xdg open and other xdg tools missing describe the bug xdg open bash xdg open command not found many scripts assume the xdg tools to be present on any desktop system to reproduce steps to reproduce the behavior install the xfce desktop open xfce terminal open xdg open try to close it expected behavior xdg open should be expected to be present on every desktop system environment please complete the following information user clr cat usr lib os release name clear linux os version id clear linux os id like clear linux os version id pretty name clear linux os ansi color home url support url bug report url mailto dev lists clearlinux org privacy policy url user clr sudo swupd bundle list password bootloader desktop desktop apps desktop assets desktop autostart desktop gnomelibs desktop locales editors emacs ethtool joe kernel native network basic openssh server openssl os core os core update kit perl basic basic sudo sysadmin basic vim which wpa supplicant server xemacs desktop | 1 |
263,178 | 28,026,453,322 | IssuesEvent | 2023-03-28 09:26:42 | Dima2021/cargo-audit | https://api.github.com/repos/Dima2021/cargo-audit | closed | CVE-2020-25575 (High) detected in failure-0.1.8.crate - autoclosed | Mend: dependency security vulnerability | ## CVE-2020-25575 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>failure-0.1.8.crate</b></p></summary>
<p>Experimental error handling abstraction.</p>
<p>Library home page: <a href="https://crates.io/api/v1/crates/failure/0.1.8/download">https://crates.io/api/v1/crates/failure/0.1.8/download</a></p>
<p>
Dependency Hierarchy:
- rustsec-0.23.3.crate (Root Library)
- cargo-edit-0.7.0.crate
- :x: **failure-0.1.8.crate** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** An issue was discovered in the failure crate through 0.1.5 for Rust. It may introduce "compatibility hazards" in some applications, and has a type confusion flaw when downcasting. NOTE: This vulnerability only affects products that are no longer supported by the maintainer. NOTE: This may overlap CVE-2019-25010.
<p>Publish Date: 2020-09-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25575>CVE-2020-25575</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
| True | CVE-2020-25575 (High) detected in failure-0.1.8.crate - autoclosed - ## CVE-2020-25575 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>failure-0.1.8.crate</b></p></summary>
<p>Experimental error handling abstraction.</p>
<p>Library home page: <a href="https://crates.io/api/v1/crates/failure/0.1.8/download">https://crates.io/api/v1/crates/failure/0.1.8/download</a></p>
<p>
Dependency Hierarchy:
- rustsec-0.23.3.crate (Root Library)
- cargo-edit-0.7.0.crate
- :x: **failure-0.1.8.crate** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** An issue was discovered in the failure crate through 0.1.5 for Rust. It may introduce "compatibility hazards" in some applications, and has a type confusion flaw when downcasting. NOTE: This vulnerability only affects products that are no longer supported by the maintainer. NOTE: This may overlap CVE-2019-25010.
<p>Publish Date: 2020-09-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25575>CVE-2020-25575</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
| non_priority | cve high detected in failure crate autoclosed cve high severity vulnerability vulnerable library failure crate experimental error handling abstraction library home page a href dependency hierarchy rustsec crate root library cargo edit crate x failure crate vulnerable library found in base branch main vulnerability details unsupported when assigned an issue was discovered in the failure crate through for rust it may introduce compatibility hazards in some applications and has a type confusion flaw when downcasting note this vulnerability only affects products that are no longer supported by the maintainer note this may overlap cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href | 0 |
525,442 | 15,253,561,020 | IssuesEvent | 2021-02-20 08:16:57 | FaultyFunctions/Crochet | https://api.github.com/repos/FaultyFunctions/Crochet | closed | Crochet for Yarn 2.0 | Priority: High Status: In-Progress Type: Feature | - [x] File tags editing
- [x] File tags saving
- [x] File tags loading
- [x] Syntax highlighting for editor
- [x] Title validation
- [x] Prevent errors and fix loading if yarn file doesn't have Crochet specific, node-level metadata like position.
- [x] Header tags
- [x] File Tags don't get reset when opening new files | 1.0 | Crochet for Yarn 2.0 - - [x] File tags editing
- [x] File tags saving
- [x] File tags loading
- [x] Syntax highlighting for editor
- [x] Title validation
- [x] Prevent errors and fix loading if yarn file doesn't have Crochet specific, node-level metadata like position.
- [x] Header tags
- [x] File Tags don't get reset when opening new files | priority | crochet for yarn file tags editing file tags saving file tags loading syntax highlighting for editor title validation prevent errors and fix loading if yarn file doesn t have crochet specific node level metadata like position header tags file tags don t get reset when opening new files | 1 |
55,170 | 13,535,947,597 | IssuesEvent | 2020-09-16 08:20:52 | googleapis/nodejs-iot | https://api.github.com/repos/googleapis/nodejs-iot | opened | The build failed | buildcop: issue priority: p1 type: bug | This test failed!
To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop).
If I'm commenting on this issue too often, add the `buildcop: quiet` label and
I will stop commenting.
---
commit: cd7bf85083bade2f6522d5484fd7635b38344ef6
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/a3d01de0-f74a-4cd1-826d-7cc508c11c06), [Sponge](http://sponge2/a3d01de0-f74a-4cd1-826d-7cc508c11c06)
status: failed
<details><summary>Test output</summary><br><pre>The expression evaluated to a falsy value:
assert.ok(devices.includes(gatewayId))
+ expected - actual
-false
+true
AssertionError [ERR_ASSERTION]: The expression evaluated to a falsy value:
assert.ok(devices.includes(gatewayId))
at Context.<anonymous> (test/manager.test.js:420:10)</pre></details> | 1.0 | The build failed - This test failed!
To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop).
If I'm commenting on this issue too often, add the `buildcop: quiet` label and
I will stop commenting.
---
commit: cd7bf85083bade2f6522d5484fd7635b38344ef6
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/a3d01de0-f74a-4cd1-826d-7cc508c11c06), [Sponge](http://sponge2/a3d01de0-f74a-4cd1-826d-7cc508c11c06)
status: failed
<details><summary>Test output</summary><br><pre>The expression evaluated to a falsy value:
assert.ok(devices.includes(gatewayId))
+ expected - actual
-false
+true
AssertionError [ERR_ASSERTION]: The expression evaluated to a falsy value:
assert.ok(devices.includes(gatewayId))
at Context.<anonymous> (test/manager.test.js:420:10)</pre></details> | non_priority | the build failed this test failed to configure my behavior see if i m commenting on this issue too often add the buildcop quiet label and i will stop commenting commit buildurl status failed test output the expression evaluated to a falsy value assert ok devices includes gatewayid expected actual false true assertionerror the expression evaluated to a falsy value assert ok devices includes gatewayid at context test manager test js | 0 |
451,276 | 13,032,702,608 | IssuesEvent | 2020-07-28 05:00:53 | kubesphere/kubesphere | https://api.github.com/repos/kubesphere/kubesphere | closed | one email can be used for different user names | area/iam kind/bug kind/need-to-verify priority/high | **Describe the Bug**
I might accidentally create a new user with the email that is already used by another user. But it is successful created.
<img width="322" alt="Screen Shot 2020-07-24 at 8 50 51 PM" src="https://user-images.githubusercontent.com/28859385/88393041-9d68f280-cdef-11ea-9aee-32ae3bcb88fd.png">
<img width="383" alt="Screen Shot 2020-07-24 at 8 50 43 PM" src="https://user-images.githubusercontent.com/28859385/88393058-a35ed380-cdef-11ea-87d7-cc46f5c09e77.png">
**Versions Used**
KubeSphere: 3.0.0-dev
Kubernetes: 1.18.6 | 1.0 | one email can be used for different user names - **Describe the Bug**
I might accidentally create a new user with the email that is already used by another user. But it is successful created.
<img width="322" alt="Screen Shot 2020-07-24 at 8 50 51 PM" src="https://user-images.githubusercontent.com/28859385/88393041-9d68f280-cdef-11ea-9aee-32ae3bcb88fd.png">
<img width="383" alt="Screen Shot 2020-07-24 at 8 50 43 PM" src="https://user-images.githubusercontent.com/28859385/88393058-a35ed380-cdef-11ea-87d7-cc46f5c09e77.png">
**Versions Used**
KubeSphere: 3.0.0-dev
Kubernetes: 1.18.6 | priority | one email can be used for different user names describe the bug i might accidentally create a new user with the email that is already used by another user but it is successful created img width alt screen shot at pm src img width alt screen shot at pm src versions used kubesphere dev kubernetes | 1 |
41,032 | 21,394,439,071 | IssuesEvent | 2022-04-21 10:11:47 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | [APM] Add experimental probability setting for aggregations | Team:apm apm:performance v8.3.0 | with the `random_sampler` aggregation being available in ES (currently behind a feature flag), we can add a simple, experimental `probability` setting that takes any value between 0 and 1. In the future we can come up with something better, but to expose and test the feature a simple setting would be a good start.
We'd need to wrap our aggregations in a `random_sampler` aggregation where possible. It might be a good idea to have it as an API parameter, which would make API testing easier. It's something we also want to do with `searchAggregatedTransactions` at some point. | True | [APM] Add experimental probability setting for aggregations - with the `random_sampler` aggregation being available in ES (currently behind a feature flag), we can add a simple, experimental `probability` setting that takes any value between 0 and 1. In the future we can come up with something better, but to expose and test the feature a simple setting would be a good start.
We'd need to wrap our aggregations in a `random_sampler` aggregation where possible. It might be a good idea to have it as an API parameter, which would make API testing easier. It's something we also want to do with `searchAggregatedTransactions` at some point. | non_priority | add experimental probability setting for aggregations with the random sampler aggregation being available in es currently behind a feature flag we can add a simple experimental probability setting that takes any value between and in the future we can come up with something better but to expose and test the feature a simple setting would be a good start we d need to wrap our aggregations in a random sampler aggregation where possible it might be a good idea to have it as an api parameter which would make api testing easier it s something we also want to do with searchaggregatedtransactions at some point | 0 |
50,090 | 3,006,179,338 | IssuesEvent | 2015-07-27 08:41:25 | Itseez/opencv | https://api.github.com/repos/Itseez/opencv | opened | Error in Python Find Contours | affected: 2.4 auto-transferred bug category: python bindings priority: low | Transferred from http://code.opencv.org/issues/1440
```
|| kscottz - on 2011-10-20 16:25
|| Priority: Low
|| Affected: 2.4.3
|| Category: python bindings
|| Tracker: Bug
|| Difficulty: None
|| PR:
|| Platform: None / None
```
Error in Python Find Contours
-----------
```
I found a subtle bug in the cv.findContours method. The error occurs when a binary image is passed in where all of the image is black, or all of the image is black and a single pixel is white along the side of the image.
Using opencv 2.3 on OSX, compiled using the brew install default settings.
For the single pixel case this should re-create the error
seq = cv.FindContours(binaryImg, self.mMemStorage, cv.CV_RETR_TREE,
cv.CV_CHAIN_APPROX_SIMPLE)
# do some operations on the sequence
nextBlob = seq.h_next()
You can use this error to recreate the bug:
https://github.com/ingenuitas/SimpleCV/blob/master/SimpleCV/sampleimages/justapixel.png
See also:
https://github.com/ingenuitas/SimpleCV/blob/master/SimpleCV/BlobMaker.py
about line 55 onward.
```
History
-------
##### Alexander Shishkov on 2012-03-21 11:39
```
- Target version deleted ()
```
##### Alexander Shishkov on 2012-03-25 20:43
```
- Assignee deleted (Vadim Pisarevsky)
```
##### Alexander Shishkov on 2012-04-05 12:45
```
- Target version deleted ()
```
##### Vadim Pisarevsky on 2013-01-22 20:08
```
please, use cv2.findContours instead. cv.FindContours will be removed soon.
- Affected version set to 2.4.3
- Priority changed from Normal to Low
``` | 1.0 | Error in Python Find Contours - Transferred from http://code.opencv.org/issues/1440
```
|| kscottz - on 2011-10-20 16:25
|| Priority: Low
|| Affected: 2.4.3
|| Category: python bindings
|| Tracker: Bug
|| Difficulty: None
|| PR:
|| Platform: None / None
```
Error in Python Find Contours
-----------
```
I found a subtle bug in the cv.findContours method. The error occurs when a binary image is passed in where all of the image is black, or all of the image is black and a single pixel is white along the side of the image.
Using opencv 2.3 on OSX, compiled using the brew install default settings.
For the single pixel case this should re-create the error
seq = cv.FindContours(binaryImg, self.mMemStorage, cv.CV_RETR_TREE,
cv.CV_CHAIN_APPROX_SIMPLE)
# do some operations on the sequence
nextBlob = seq.h_next()
You can use this error to recreate the bug:
https://github.com/ingenuitas/SimpleCV/blob/master/SimpleCV/sampleimages/justapixel.png
See also:
https://github.com/ingenuitas/SimpleCV/blob/master/SimpleCV/BlobMaker.py
about line 55 onward.
```
History
-------
##### Alexander Shishkov on 2012-03-21 11:39
```
- Target version deleted ()
```
##### Alexander Shishkov on 2012-03-25 20:43
```
- Assignee deleted (Vadim Pisarevsky)
```
##### Alexander Shishkov on 2012-04-05 12:45
```
- Target version deleted ()
```
##### Vadim Pisarevsky on 2013-01-22 20:08
```
please, use cv2.findContours instead. cv.FindContours will be removed soon.
- Affected version set to 2.4.3
- Priority changed from Normal to Low
``` | priority | error in python find contours transferred from kscottz on priority low affected category python bindings tracker bug difficulty none pr platform none none error in python find contours i found a subtle bug in the cv findcontours method the error occurs when a binary image is passed in where all of the image is black or all of the image is black and a single pixel is white along the side of the image using opencv on osx compiled using the brew install default settings for the single pixel case this should re create the error seq cv findcontours binaryimg self mmemstorage cv cv retr tree cv cv chain approx simple do some operations on the sequence nextblob seq h next you can use this error to recreate the bug see also about line onward history alexander shishkov on target version deleted alexander shishkov on assignee deleted vadim pisarevsky alexander shishkov on target version deleted vadim pisarevsky on please use findcontours instead cv findcontours will be removed soon affected version set to priority changed from normal to low | 1 |
313,337 | 9,559,795,641 | IssuesEvent | 2019-05-03 17:44:14 | GOTO-OBS/goto-alert | https://api.github.com/repos/GOTO-OBS/goto-alert | opened | [WIP] Lessons from S190425z and S190426c | -effort 1 -priority 1 -type proposal GW_events database event_handler | So we've had two actual BNS/NSBH events that we tried to follow up with GOTO. For S190425z and S190426c I ended up doing a fair amount of database insertion myself, from the EMMA workshop in Baltimore. There were a few reasons for that:
* The sentinel failed a couple of times thanks to the way skymaps are stored in the XML VOEvents. See #24 and #25, should be fixed going forward.
* There was a later LVC skymap that also failed due to a problem on LIGO's end (for ivo://gwnet/LVC#S190426c-4-Update). Nothing to do with me, and they did release an update ~45 minutes later. However due to the way the code is ordered the previous tiles were deleted before creating the new ones, so had we been observing we would have had a very annoying gap.
* The first event, 25z, had an incredibly large and wide skymap. The sentinel has a default cutoff for tiles of 1% for GW events (5% for GRBs) and no tiles were above this. To do it manually I had to go down to 0.1%. The 26c skymap was also pretty large, and I lowered the limit to 0.5%.
* The priority was determined to be coverage over repeat visits. The sentinel would only insert a maximum of 50 tiles for GW events, and when we can observe ~130 in a good night we'd prefer to do as many tiles as possible (with a reasonable probability).
* The other thing I considered was visibility. The sentinel is deliberately ambivalent to this, but manually I could use Astroplan to limit the tiles I put in. I don't want GOTO-alert to start doing that, although it might be handy for the sentinel alerts. But that was how I selected the probability limits, which makes it a little harder to do automatically. I suppose the difference there is between altitude and solar limits, the latter will apply wherever you are on Earth so might be reasonable for this module to consider.
Obviously the idea is I shouldn't be needed - the sentinel using GOTO-alert (and GOTO-tile) should insert the pointings it needs.
So, the question is how do we remove the human element? Well #22 is a start, but needs more...
| 1.0 | [WIP] Lessons from S190425z and S190426c - So we've had two actual BNS/NSBH events that we tried to follow up with GOTO. For S190425z and S190426c I ended up doing a fair amount of database insertion myself, from the EMMA workshop in Baltimore. There were a few reasons for that:
* The sentinel failed a couple of times thanks to the way skymaps are stored in the XML VOEvents. See #24 and #25, should be fixed going forward.
* There was a later LVC skymap that also failed due to a problem on LIGO's end (for ivo://gwnet/LVC#S190426c-4-Update). Nothing to do with me, and they did release an update ~45 minutes later. However due to the way the code is ordered the previous tiles were deleted before creating the new ones, so had we been observing we would have had a very annoying gap.
* The first event, 25z, had an incredibly large and wide skymap. The sentinel has a default cutoff for tiles of 1% for GW events (5% for GRBs) and no tiles were above this. To do it manually I had to go down to 0.1%. The 26c skymap was also pretty large, and I lowered the limit to 0.5%.
* The priority was determined to be coverage over repeat visits. The sentinel would only insert a maximum of 50 tiles for GW events, and when we can observe ~130 in a good night we'd prefer to do as many tiles as possible (with a reasonable probability).
* The other thing I considered was visibility. The sentinel is deliberately ambivalent to this, but manually I could use Astroplan to limit the tiles I put in. I don't want GOTO-alert to start doing that, although it might be handy for the sentinel alerts. But that was how I selected the probability limits, which makes it a little harder to do automatically. I suppose the difference there is between altitude and solar limits, the latter will apply wherever you are on Earth so might be reasonable for this module to consider.
Obviously the idea is I shouldn't be needed - the sentinel using GOTO-alert (and GOTO-tile) should insert the pointings it needs.
So, the question is how do we remove the human element? Well #22 is a start, but needs more...
| priority | lessons from and so we ve had two actual bns nsbh events that we tried to follow up with goto for and i ended up doing a fair amount of database insertion myself from the emma workshop in baltimore there were a few reasons for that the sentinel failed a couple of times thanks to the way skymaps are stored in the xml voevents see and should be fixed going forward there was a later lvc skymap that also failed due to a problem on ligo s end for ivo gwnet lvc update nothing to do with me and they did release an update minutes later however due to the way the code is ordered the previous tiles were deleted before creating the new ones so had we been observing we would have had a very annoying gap the first event had an incredibly large and wide skymap the sentinel has a default cutoff for tiles of for gw events for grbs and no tiles were above this to do it manually i had to go down to the skymap was also pretty large and i lowered the limit to the priority was determined to be coverage over repeat visits the sentinel would only insert a maximum of tiles for gw events and when we can observe in a good night we d prefer to do as many tiles as possible with a reasonable probability the other thing i considered was visibility the sentinel is deliberately ambivalent to this but manually i could use astroplan to limit the tiles i put in i don t want goto alert to start doing that although it might be handy for the sentinel alerts but that was how i selected the probability limits which makes it a little harder to do automatically i suppose the difference there is between altitude and solar limits the latter will apply wherever you are on earth so might be reasonable for this module to consider obviously the idea is i shouldn t be needed the sentinel using goto alert and goto tile should insert the pointings it needs so the question is how do we remove the human element well is a start but needs more | 1 |
43,492 | 2,889,811,853 | IssuesEvent | 2015-06-13 19:42:39 | IMAGINARY/imaginary-web | https://api.github.com/repos/IMAGINARY/imaginary-web | opened | Translation of new Texts section | enhancement high priority | The new Texts section has many unstranslated interface strings. | 1.0 | Translation of new Texts section - The new Texts section has many unstranslated interface strings. | priority | translation of new texts section the new texts section has many unstranslated interface strings | 1 |
354,188 | 10,563,508,231 | IssuesEvent | 2019-10-04 21:10:51 | wso2-cellery/sdk | https://api.github.com/repos/wso2-cellery/sdk | closed | Error while executing cellery test command | Priority/Highest Severity/Critical Type/Bug | Traceback (most recent call last):
File "/Library/Cellery/telepresence-0.101/bin/telepresence/telepresence/cli.py", line 136, in crash_reporting
yield
File "/Library/Cellery/telepresence-0.101/bin/telepresence/telepresence/main.py", line 66, in main
env, pod_info = get_remote_env(runner, ssh, remote_info)
File "/Library/Cellery/telepresence-0.101/bin/telepresence/telepresence/remote_env.py", line 34, in get_remote_env
remote_info.container_name, "--", "python3", "podinfo.py"
File "/Library/Cellery/telepresence-0.101/bin/telepresence/telepresence/runner/runner.py", line 445, in get_output
env,
File "/Library/Cellery/telepresence-0.101/bin/telepresence/telepresence/runner/runner.py", line 399, in _run_command_sync
None if stderr_to_stdout else err_logger.get_captured(),
subprocess.CalledProcessError: Command '['kubectl', '--context', 'gke_pzf-cellery-254002_us-east4-a_cellery-cluster568', '--namespace', 'default', 'exec', 'telepresence--telepresence-deployment-d7948c9cb-8khwj', '--container', 'telepresence', '--', 'python3', 'podinfo.py']' returned non-zero exit status 1.
Here are the last few lines of the logfile (see /Users/paul/cdemo/pets/target/telepresence.log for the complete logs):
12.1 TEL |
12.1 TEL | Recent output was:
12.1 TEL | 2019-10-01T16:32:11+0000 [socks.SOCKSv5Factory#info] Starting factory <socks.SOCKSv5Factory object at 0x7f810f233668>
12.1 TEL | 2019-10-01T16:32:11+0000 [-] DNSDatagramProtocol starting on 9053
12.1 TEL | 2019-10-01T16:32:11+0000 [-] Starting protocol <twisted.names.dns.DNSDatagramProtocol object at 0x7f810f233908>
12.1 TEL | 2019-10-01T16:32:11+0000 [-] Loaded.
12.1 TEL | 2019-10-01T16:32:11+0000 [twisted.scripts._twistd_unix.UnixAppLogger#info] twistd 19.2.1 (/usr/bin/python3.6 3.6.8) starting up.
12.1 TEL | 2019-10-01T16:32:11+0000 [twisted.scripts._twistd_unix.UnixAppLogger#info] reactor class: twisted.internet.epollreactor.EPollReactor.
| 1.0 | Error while executing cellery test command - Traceback (most recent call last):
File "/Library/Cellery/telepresence-0.101/bin/telepresence/telepresence/cli.py", line 136, in crash_reporting
yield
File "/Library/Cellery/telepresence-0.101/bin/telepresence/telepresence/main.py", line 66, in main
env, pod_info = get_remote_env(runner, ssh, remote_info)
File "/Library/Cellery/telepresence-0.101/bin/telepresence/telepresence/remote_env.py", line 34, in get_remote_env
remote_info.container_name, "--", "python3", "podinfo.py"
File "/Library/Cellery/telepresence-0.101/bin/telepresence/telepresence/runner/runner.py", line 445, in get_output
env,
File "/Library/Cellery/telepresence-0.101/bin/telepresence/telepresence/runner/runner.py", line 399, in _run_command_sync
None if stderr_to_stdout else err_logger.get_captured(),
subprocess.CalledProcessError: Command '['kubectl', '--context', 'gke_pzf-cellery-254002_us-east4-a_cellery-cluster568', '--namespace', 'default', 'exec', 'telepresence--telepresence-deployment-d7948c9cb-8khwj', '--container', 'telepresence', '--', 'python3', 'podinfo.py']' returned non-zero exit status 1.
Here are the last few lines of the logfile (see /Users/paul/cdemo/pets/target/telepresence.log for the complete logs):
12.1 TEL |
12.1 TEL | Recent output was:
12.1 TEL | 2019-10-01T16:32:11+0000 [socks.SOCKSv5Factory#info] Starting factory <socks.SOCKSv5Factory object at 0x7f810f233668>
12.1 TEL | 2019-10-01T16:32:11+0000 [-] DNSDatagramProtocol starting on 9053
12.1 TEL | 2019-10-01T16:32:11+0000 [-] Starting protocol <twisted.names.dns.DNSDatagramProtocol object at 0x7f810f233908>
12.1 TEL | 2019-10-01T16:32:11+0000 [-] Loaded.
12.1 TEL | 2019-10-01T16:32:11+0000 [twisted.scripts._twistd_unix.UnixAppLogger#info] twistd 19.2.1 (/usr/bin/python3.6 3.6.8) starting up.
12.1 TEL | 2019-10-01T16:32:11+0000 [twisted.scripts._twistd_unix.UnixAppLogger#info] reactor class: twisted.internet.epollreactor.EPollReactor.
| priority | error while executing cellery test command traceback most recent call last file library cellery telepresence bin telepresence telepresence cli py line in crash reporting yield file library cellery telepresence bin telepresence telepresence main py line in main env pod info get remote env runner ssh remote info file library cellery telepresence bin telepresence telepresence remote env py line in get remote env remote info container name podinfo py file library cellery telepresence bin telepresence telepresence runner runner py line in get output env file library cellery telepresence bin telepresence telepresence runner runner py line in run command sync none if stderr to stdout else err logger get captured subprocess calledprocesserror command returned non zero exit status here are the last few lines of the logfile see users paul cdemo pets target telepresence log for the complete logs tel tel recent output was tel starting factory tel dnsdatagramprotocol starting on tel starting protocol tel loaded tel twistd usr bin starting up tel reactor class twisted internet epollreactor epollreactor | 1 |
317,088 | 23,663,509,847 | IssuesEvent | 2022-08-26 18:05:42 | awslabs/amazon-eks-ami | https://api.github.com/repos/awslabs/amazon-eks-ami | closed | Bring back EKS AMI versions in AWS documentation | documentation | I used to be able to visit https://docs.aws.amazon.com/eks/latest/userguide/eks-linux-ami-versions.html to get a list of valid AMI versions for EKS.
Now, the list is gone.
I have to visit https://github.com/awslabs/amazon-eks-ami/blob/master/CHANGELOG.md, do information stitching with trial and error to get the right AMI versions.
Example:
What's the corresponding version for AMI Release v20220811?
To figure out major and minor version of supported k8s, I need to get it from S3 location: s3://amazon-eks/1.22.12/2022-07-27/
Thus, a valid version is: 1.22.12-20220811.
Why this version is important?
CDK EKS Cluster class has an implementation add_nodegroup_capacity that supports release_version.
We need to make it easy to find out what version is available. Above method IS NOT easy.
| 1.0 | Bring back EKS AMI versions in AWS documentation - I used to be able to visit https://docs.aws.amazon.com/eks/latest/userguide/eks-linux-ami-versions.html to get a list of valid AMI versions for EKS.
Now, the list is gone.
I have to visit https://github.com/awslabs/amazon-eks-ami/blob/master/CHANGELOG.md, do information stitching with trial and error to get the right AMI versions.
Example:
What's the corresponding version for AMI Release v20220811?
To figure out major and minor version of supported k8s, I need to get it from S3 location: s3://amazon-eks/1.22.12/2022-07-27/
Thus, a valid version is: 1.22.12-20220811.
Why this version is important?
CDK EKS Cluster class has an implementation add_nodegroup_capacity that supports release_version.
We need to make it easy to find out what version is available. Above method IS NOT easy.
| non_priority | bring back eks ami versions in aws documentation i used to be able to visit to get a list of valid ami versions for eks now the list is gone i have to visit do information stitching with trial and error to get the right ami versions example what s the corresponding version for ami release to figure out major and minor version of supported i need to get it from location amazon eks thus a valid version is why this version is important cdk eks cluster class has an implementation add nodegroup capacity that supports release version we need to make it easy to find out what version is available above method is not easy | 0 |
712,392 | 24,493,870,770 | IssuesEvent | 2022-10-10 06:43:41 | open-telemetry/opentelemetry-js-contrib | https://api.github.com/repos/open-telemetry/opentelemetry-js-contrib | closed | hapi: support direct plugins with prefixed paths | bug up-for-grabs stale priority:p1 | <!--
Please answer these questions before submitting a bug report.
-->
### What version of OpenTelemetry are you using?
`@opentelemetry/instrumentation-hapi 0.27.0`
### What version of Node are you using?
`16.10.0`
### What did you do?
Passed plugin directly to `server.register` along with options:
```javascript
const Hapi = require('@hapi/hapi');
const init = async () => {
const server = Hapi.server({port: 3000, host: 'localhost'});
const plugin = {
name: 'myPlugin',
version: '1.0.0',
register: async function (server, options) {
server.route({
method: 'GET',
path: '/test',
handler: function (request, h) {
return 'hello, world';
}
});
},
};
await server.register(myPlugin, {
routes: {
prefix: '/foo',
}
});
await server.start();
};
init();
```
Did a request to `http://localhost:3000/foo/test`, got the following error:
```
TypeError: Cannot read properties of undefined (reading 'name')
at Object.getPluginName (node_modules/@opentelemetry/instrumentation-hapi/build/src/utils.js:23:16)
at HapiInstrumentation._wrapRegisterHandler (node_modules/@opentelemetry/instrumentation-hapi/build/src/instrumentation.js:184:36)
at internals.Server.register (node_modules/@opentelemetry/instrumentation-hapi/build/src/instrumentation.js:103:33)
at init (index.js:27:18)
```
### What did you expect to see?
Application starting up and a span named `myplugin: route - /foo/test`
### What did you see instead?
Application not booting due to an error in the instrumentation.
| 1.0 | hapi: support direct plugins with prefixed paths - <!--
Please answer these questions before submitting a bug report.
-->
### What version of OpenTelemetry are you using?
`@opentelemetry/instrumentation-hapi 0.27.0`
### What version of Node are you using?
`16.10.0`
### What did you do?
Passed plugin directly to `server.register` along with options:
```javascript
const Hapi = require('@hapi/hapi');
const init = async () => {
const server = Hapi.server({port: 3000, host: 'localhost'});
const plugin = {
name: 'myPlugin',
version: '1.0.0',
register: async function (server, options) {
server.route({
method: 'GET',
path: '/test',
handler: function (request, h) {
return 'hello, world';
}
});
},
};
await server.register(myPlugin, {
routes: {
prefix: '/foo',
}
});
await server.start();
};
init();
```
Did a request to `http://localhost:3000/foo/test`, got the following error:
```
TypeError: Cannot read properties of undefined (reading 'name')
at Object.getPluginName (node_modules/@opentelemetry/instrumentation-hapi/build/src/utils.js:23:16)
at HapiInstrumentation._wrapRegisterHandler (node_modules/@opentelemetry/instrumentation-hapi/build/src/instrumentation.js:184:36)
at internals.Server.register (node_modules/@opentelemetry/instrumentation-hapi/build/src/instrumentation.js:103:33)
at init (index.js:27:18)
```
### What did you expect to see?
Application starting up and a span named `myplugin: route - /foo/test`
### What did you see instead?
Application not booting due to an error in the instrumentation.
| priority | hapi support direct plugins with prefixed paths please answer these questions before submitting a bug report what version of opentelemetry are you using opentelemetry instrumentation hapi what version of node are you using what did you do passed plugin directly to server register along with options javascript const hapi require hapi hapi const init async const server hapi server port host localhost const plugin name myplugin version register async function server options server route method get path test handler function request h return hello world await server register myplugin routes prefix foo await server start init did a request to got the following error typeerror cannot read properties of undefined reading name at object getpluginname node modules opentelemetry instrumentation hapi build src utils js at hapiinstrumentation wrapregisterhandler node modules opentelemetry instrumentation hapi build src instrumentation js at internals server register node modules opentelemetry instrumentation hapi build src instrumentation js at init index js what did you expect to see application starting up and a span named myplugin route foo test what did you see instead application not booting due to an error in the instrumentation | 1 |
17,294 | 6,401,186,191 | IssuesEvent | 2017-08-05 18:28:10 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Building export templates (release & release_debug) X11 failed (core/rid.h) | bug platform:linux topic:buildsystem | *Bugsquad edit:* Current master HEAD.
Impossible to compile export templates for X11
scons platform=x11 -> works
but:
scons platform=x11 tools=no target=release_debug bits=64 -> failed
scons platform=x11 tools=no target=release bits=64 -> failed
```
scons platform=x11 tools=no target=release bits=64
scons: Reading SConscript files ...
i586-mingw32msvc-gcc (GCC) 4.2.1-sjlj (mingw32-2)
Copyright (C) 2007 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
i686-w64-mingw32-gcc (GCC) 4.8.2
Copyright (C) 2013 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
Enabling ALSA
Enabling PulseAudio
scons: done reading SConscript files.
scons: Building targets ...
g++ -o platform/x11/godot_x11.x11.opt.64.o -c -O3 -ffast-math -DNDEBUG -DFREETYPE_ENABLED -DENABLE_DEPRECATED -DOPENGL_ENABLED -DALSA_ENABLED -DJOYDEV_ENABLED -DPULSEAUDIO_ENABLED -DX11_ENABLED -DUNIX_ENABLED -DGLES2_ENABLED -DGLES_OVER_GL -DGDSCRIPT_ENABLED -DMINIZIP_ENABLED -DXML_ENABLED -DGLAD_ENABLED -DGLES_OVER_GL -D_REENTRANT -Icore -Icore/math -Itools -Idrivers -I. -Iplatform/x11 -I/usr/include/freetype2 -I/usr/include/libpng12 -I/usr/include/alsa -Iplatform/x11 platform/x11/godot_x11.cpp
In file included from core/ref_ptr.h:37:0,
from core/variant.h:50,
from core/object.h:33,
from core/os/input.h:32,
from platform/x11/os_x11.h:33,
from platform/x11/godot_x11.cpp:30:
core/rid.h: In member function 'bool RID_OwnerBase::_is_owner(RID&) const':
core/rid.h:118:22: error: 'class RID' has no member named '_owner'
return this==p_rid._owner;
^
core/rid.h: In member function 'void RID_OwnerBase::_remove_owner(RID&)':
core/rid.h:124:16: error: 'class RID' has no member named '_owner'
return p_rid._owner=NULL;
^
In file included from /usr/include/string.h:32:0,
from core/os/copymem.h:40,
from core/os/input_event.h:34,
from core/variant.h:47,
from core/object.h:33,
from core/os/input.h:32,
from platform/x11/os_x11.h:33,
from platform/x11/godot_x11.cpp:30:
core/rid.h:124:23: error: return-statement with a value, in function returning 'void' [-fpermissive]
return p_rid._owner=NULL;
^
scons: *** [platform/x11/godot_x11.x11.opt.64.o] Error 1
scons: building terminated because of errors.
```
```
scons platform=x11 tools=no target=release_debug bits=64
scons: Reading SConscript files ...
i586-mingw32msvc-gcc (GCC) 4.2.1-sjlj (mingw32-2)
Copyright (C) 2007 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
i686-w64-mingw32-gcc (GCC) 4.8.2
Copyright (C) 2013 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
Enabling ALSA
Enabling PulseAudio
scons: done reading SConscript files.
scons: Building targets ...
g++ -o scene/3d/gi_probe.x11.opt.debug.64.o -c -O2 -ffast-math -DDEBUG_ENABLED -DFREETYPE_ENABLED -DENABLE_DEPRECATED -DOPENGL_ENABLED -DALSA_ENABLED -DJOYDEV_ENABLED -DPULSEAUDIO_ENABLED -DX11_ENABLED -DUNIX_ENABLED -DGLES2_ENABLED -DGLES_OVER_GL -DGDSCRIPT_ENABLED -DMINIZIP_ENABLED -DXML_ENABLED -DGLAD_ENABLED -DGLES_OVER_GL -D_REENTRANT -Icore -Icore/math -Itools -Idrivers -I. -Iplatform/x11 -I/usr/include/freetype2 -I/usr/include/libpng12 -I/usr/include/alsa -Iscene/3d scene/3d/gi_probe.cpp
scene/3d/gi_probe.cpp: In member function 'void GIProbe::_create_debug_mesh(GIProbe::Baker*)':
scene/3d/gi_probe.cpp:1317:18: error: 'class SceneTree' has no member named 'get_edited_scene_root'
if (get_tree()->get_edited_scene_root()==this){
^
scons: *** [scene/3d/gi_probe.x11.opt.debug.64.o] Error 1
scons: building terminated because of errors.
``` | 1.0 | Building export templates (release & release_debug) X11 failed (core/rid.h) - *Bugsquad edit:* Current master HEAD.
Impossible to compile export templates for X11
scons platform=x11 -> works
but:
scons platform=x11 tools=no target=release_debug bits=64 -> failed
scons platform=x11 tools=no target=release bits=64 -> failed
```
scons platform=x11 tools=no target=release bits=64
scons: Reading SConscript files ...
i586-mingw32msvc-gcc (GCC) 4.2.1-sjlj (mingw32-2)
Copyright (C) 2007 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
i686-w64-mingw32-gcc (GCC) 4.8.2
Copyright (C) 2013 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
Enabling ALSA
Enabling PulseAudio
scons: done reading SConscript files.
scons: Building targets ...
g++ -o platform/x11/godot_x11.x11.opt.64.o -c -O3 -ffast-math -DNDEBUG -DFREETYPE_ENABLED -DENABLE_DEPRECATED -DOPENGL_ENABLED -DALSA_ENABLED -DJOYDEV_ENABLED -DPULSEAUDIO_ENABLED -DX11_ENABLED -DUNIX_ENABLED -DGLES2_ENABLED -DGLES_OVER_GL -DGDSCRIPT_ENABLED -DMINIZIP_ENABLED -DXML_ENABLED -DGLAD_ENABLED -DGLES_OVER_GL -D_REENTRANT -Icore -Icore/math -Itools -Idrivers -I. -Iplatform/x11 -I/usr/include/freetype2 -I/usr/include/libpng12 -I/usr/include/alsa -Iplatform/x11 platform/x11/godot_x11.cpp
In file included from core/ref_ptr.h:37:0,
from core/variant.h:50,
from core/object.h:33,
from core/os/input.h:32,
from platform/x11/os_x11.h:33,
from platform/x11/godot_x11.cpp:30:
core/rid.h: In member function 'bool RID_OwnerBase::_is_owner(RID&) const':
core/rid.h:118:22: error: 'class RID' has no member named '_owner'
return this==p_rid._owner;
^
core/rid.h: In member function 'void RID_OwnerBase::_remove_owner(RID&)':
core/rid.h:124:16: error: 'class RID' has no member named '_owner'
return p_rid._owner=NULL;
^
In file included from /usr/include/string.h:32:0,
from core/os/copymem.h:40,
from core/os/input_event.h:34,
from core/variant.h:47,
from core/object.h:33,
from core/os/input.h:32,
from platform/x11/os_x11.h:33,
from platform/x11/godot_x11.cpp:30:
core/rid.h:124:23: error: return-statement with a value, in function returning 'void' [-fpermissive]
return p_rid._owner=NULL;
^
scons: *** [platform/x11/godot_x11.x11.opt.64.o] Error 1
scons: building terminated because of errors.
```
```
scons platform=x11 tools=no target=release_debug bits=64
scons: Reading SConscript files ...
i586-mingw32msvc-gcc (GCC) 4.2.1-sjlj (mingw32-2)
Copyright (C) 2007 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
i686-w64-mingw32-gcc (GCC) 4.8.2
Copyright (C) 2013 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
Enabling ALSA
Enabling PulseAudio
scons: done reading SConscript files.
scons: Building targets ...
g++ -o scene/3d/gi_probe.x11.opt.debug.64.o -c -O2 -ffast-math -DDEBUG_ENABLED -DFREETYPE_ENABLED -DENABLE_DEPRECATED -DOPENGL_ENABLED -DALSA_ENABLED -DJOYDEV_ENABLED -DPULSEAUDIO_ENABLED -DX11_ENABLED -DUNIX_ENABLED -DGLES2_ENABLED -DGLES_OVER_GL -DGDSCRIPT_ENABLED -DMINIZIP_ENABLED -DXML_ENABLED -DGLAD_ENABLED -DGLES_OVER_GL -D_REENTRANT -Icore -Icore/math -Itools -Idrivers -I. -Iplatform/x11 -I/usr/include/freetype2 -I/usr/include/libpng12 -I/usr/include/alsa -Iscene/3d scene/3d/gi_probe.cpp
scene/3d/gi_probe.cpp: In member function 'void GIProbe::_create_debug_mesh(GIProbe::Baker*)':
scene/3d/gi_probe.cpp:1317:18: error: 'class SceneTree' has no member named 'get_edited_scene_root'
if (get_tree()->get_edited_scene_root()==this){
^
scons: *** [scene/3d/gi_probe.x11.opt.debug.64.o] Error 1
scons: building terminated because of errors.
``` | non_priority | building export templates release release debug failed core rid h bugsquad edit current master head impossible to compile export templates for scons platform works but scons platform tools no target release debug bits failed scons platform tools no target release bits failed scons platform tools no target release bits scons reading sconscript files gcc gcc sjlj copyright c free software foundation inc this is free software see the source for copying conditions there is no warranty not even for merchantability or fitness for a particular purpose gcc gcc copyright c free software foundation inc this is free software see the source for copying conditions there is no warranty not even for merchantability or fitness for a particular purpose enabling alsa enabling pulseaudio scons done reading sconscript files scons building targets g o platform godot opt o c ffast math dndebug dfreetype enabled denable deprecated dopengl enabled dalsa enabled djoydev enabled dpulseaudio enabled enabled dunix enabled enabled dgles over gl dgdscript enabled dminizip enabled dxml enabled dglad enabled dgles over gl d reentrant icore icore math itools idrivers i iplatform i usr include i usr include i usr include alsa iplatform platform godot cpp in file included from core ref ptr h from core variant h from core object h from core os input h from platform os h from platform godot cpp core rid h in member function bool rid ownerbase is owner rid const core rid h error class rid has no member named owner return this p rid owner core rid h in member function void rid ownerbase remove owner rid core rid h error class rid has no member named owner return p rid owner null in file included from usr include string h from core os copymem h from core os input event h from core variant h from core object h from core os input h from platform os h from platform godot cpp core rid h error return statement with a value in function returning void return p rid owner null scons error scons building terminated because of errors scons platform tools no target release debug bits scons reading sconscript files gcc gcc sjlj copyright c free software foundation inc this is free software see the source for copying conditions there is no warranty not even for merchantability or fitness for a particular purpose gcc gcc copyright c free software foundation inc this is free software see the source for copying conditions there is no warranty not even for merchantability or fitness for a particular purpose enabling alsa enabling pulseaudio scons done reading sconscript files scons building targets g o scene gi probe opt debug o c ffast math ddebug enabled dfreetype enabled denable deprecated dopengl enabled dalsa enabled djoydev enabled dpulseaudio enabled enabled dunix enabled enabled dgles over gl dgdscript enabled dminizip enabled dxml enabled dglad enabled dgles over gl d reentrant icore icore math itools idrivers i iplatform i usr include i usr include i usr include alsa iscene scene gi probe cpp scene gi probe cpp in member function void giprobe create debug mesh giprobe baker scene gi probe cpp error class scenetree has no member named get edited scene root if get tree get edited scene root this scons error scons building terminated because of errors | 0 |
165,085 | 6,262,024,367 | IssuesEvent | 2017-07-15 06:05:33 | HabitRPG/habitica | https://api.github.com/repos/HabitRPG/habitica | closed | Add Reasons for Leaving Group/Kicking from Group | priority: minor status: issue: help welcome now | Need to add an optional text field so people can state why they're leaving a group or party leaders can send a message explaining why they're kicking someone out.
## <bountysource-plugin>
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/1419952-add-reasons-for-leaving-group-kicking-from-group?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github).
</bountysource-plugin>
| 1.0 | Add Reasons for Leaving Group/Kicking from Group - Need to add an optional text field so people can state why they're leaving a group or party leaders can send a message explaining why they're kicking someone out.
## <bountysource-plugin>
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/1419952-add-reasons-for-leaving-group-kicking-from-group?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github).
</bountysource-plugin>
| priority | add reasons for leaving group kicking from group need to add an optional text field so people can state why they re leaving a group or party leaders can send a message explaining why they re kicking someone out want to back this issue we accept bounties via | 1 |
235,533 | 18,051,394,948 | IssuesEvent | 2021-09-19 20:10:25 | orcmid/nfoTools | https://api.github.com/repos/orcmid/nfoTools | opened | Triage nfoWare glossary project | task documentation pattern practice | There is a very old semantic exploration associated with this activity. It has to do with collaborative lexicography and is perhaps tied to the adbib upgrade as well. It certainly deals with networks of agreements.
There is also the matter of having an useful glossary for things nfoTool. | 1.0 | Triage nfoWare glossary project - There is a very old semantic exploration associated with this activity. It has to do with collaborative lexicography and is perhaps tied to the adbib upgrade as well. It certainly deals with networks of agreements.
There is also the matter of having an useful glossary for things nfoTool. | non_priority | triage nfoware glossary project there is a very old semantic exploration associated with this activity it has to do with collaborative lexicography and is perhaps tied to the adbib upgrade as well it certainly deals with networks of agreements there is also the matter of having an useful glossary for things nfotool | 0 |
38,678 | 15,772,613,332 | IssuesEvent | 2021-03-31 22:02:21 | dotnet/fsharp | https://api.github.com/repos/dotnet/fsharp | closed | Differentiate between signature and implementation files in Go to All results | Area-IDE Language Service Feature Request | When looking up `mknull` with Ctrl+T, the results look like this

There's no way to tell which results corresponde to implementations and which to signatures. (And the order of `QuotationPickler.mkNull` results is a bit weird)
What's worse, the first `mkMutableCompGenLocal` result when searching for `mkmut` always takes me to the signature file, but when searching for `mkmuta` the implementation is the first. In other words, where you're navigating is a coin toss and it gets even more annoying when I reopen Go to All and proceed to select the same option again thinking I chose the other one previously...
**Describe the solution you'd like**
Perhaps the bottom line could say `signature for module FSharp.Compiler.TypedTreeOps` instead of `module FSharp.Compiler.TypedTreeOps` or something along those lines.
**Describe alternatives you've considered**
Always order implementation and signature in the same manner, so that I know if I select the first of the 2, I will always be taken to signature for instance. | 1.0 | Differentiate between signature and implementation files in Go to All results - When looking up `mknull` with Ctrl+T, the results look like this

There's no way to tell which results corresponde to implementations and which to signatures. (And the order of `QuotationPickler.mkNull` results is a bit weird)
What's worse, the first `mkMutableCompGenLocal` result when searching for `mkmut` always takes me to the signature file, but when searching for `mkmuta` the implementation is the first. In other words, where you're navigating is a coin toss and it gets even more annoying when I reopen Go to All and proceed to select the same option again thinking I chose the other one previously...
**Describe the solution you'd like**
Perhaps the bottom line could say `signature for module FSharp.Compiler.TypedTreeOps` instead of `module FSharp.Compiler.TypedTreeOps` or something along those lines.
**Describe alternatives you've considered**
Always order implementation and signature in the same manner, so that I know if I select the first of the 2, I will always be taken to signature for instance. | non_priority | differentiate between signature and implementation files in go to all results when looking up mknull with ctrl t the results look like this there s no way to tell which results corresponde to implementations and which to signatures and the order of quotationpickler mknull results is a bit weird what s worse the first mkmutablecompgenlocal result when searching for mkmut always takes me to the signature file but when searching for mkmuta the implementation is the first in other words where you re navigating is a coin toss and it gets even more annoying when i reopen go to all and proceed to select the same option again thinking i chose the other one previously describe the solution you d like perhaps the bottom line could say signature for module fsharp compiler typedtreeops instead of module fsharp compiler typedtreeops or something along those lines describe alternatives you ve considered always order implementation and signature in the same manner so that i know if i select the first of the i will always be taken to signature for instance | 0 |
89,398 | 8,202,632,598 | IssuesEvent | 2018-09-02 12:01:06 | humera987/HumTestData | https://api.github.com/repos/humera987/HumTestData | opened | Humera_Test_Proj : api_v1_dashboard_count-projects_get_auth_invalid | Humera_Test_Proj | Project : Humera_Test_Proj
Job : Stg
Env : Stg
Region : FXLabs/US_WEST_1
Result : fail
Status Code : 500
Headers : {}
Endpoint : http://13.57.51.56/api/v1/dashboard/count-projects
Request :
Response :
I/O error on GET request for "http://13.57.51.56/api/v1/dashboard/count-projects": Connect to 13.57.51.56:80 [/13.57.51.56] failed: connect timed out; nested exception is org.apache.http.conn.ConnectTimeoutException: Connect to 13.57.51.56:80 [/13.57.51.56] failed: connect timed out
Logs :
Assertion [@StatusCode == 401] failed, expected value [401] but found [500]
--- FX Bot --- | 1.0 | Humera_Test_Proj : api_v1_dashboard_count-projects_get_auth_invalid - Project : Humera_Test_Proj
Job : Stg
Env : Stg
Region : FXLabs/US_WEST_1
Result : fail
Status Code : 500
Headers : {}
Endpoint : http://13.57.51.56/api/v1/dashboard/count-projects
Request :
Response :
I/O error on GET request for "http://13.57.51.56/api/v1/dashboard/count-projects": Connect to 13.57.51.56:80 [/13.57.51.56] failed: connect timed out; nested exception is org.apache.http.conn.ConnectTimeoutException: Connect to 13.57.51.56:80 [/13.57.51.56] failed: connect timed out
Logs :
Assertion [@StatusCode == 401] failed, expected value [401] but found [500]
--- FX Bot --- | non_priority | humera test proj api dashboard count projects get auth invalid project humera test proj job stg env stg region fxlabs us west result fail status code headers endpoint request response i o error on get request for connect to failed connect timed out nested exception is org apache http conn connecttimeoutexception connect to failed connect timed out logs assertion failed expected value but found fx bot | 0 |
385,326 | 11,418,648,255 | IssuesEvent | 2020-02-03 05:27:01 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | closed | Incorrect connection url in Analytics configuration | Priority/Low | **Description:**
Readme has incorrect connection url as tcp://localhost:7612/,tcp://localhost:7613/,tcp://localhost:7614/ and due to this it gets Number format exception.
It should be tcp://localhost:7612,tcp://localhost:7613,tcp://localhost:7614. Due to
https://github.com/wso2/carbon-apimgt/blob/v6.1.66/features/apimgt/org.wso2.carbon.apimgt.core.feature/src/main/resources/config/api-manager.xml#L138
**Suggested Labels:**
Minor, Readme
**Suggested Assignees:**
N/A
**Affected Product Version:**
2.1.x
**OS, DB, other environment details and versions:**
N/A
**Steps to reproduce:**
N/A
**Related Issues:**
N/A | 1.0 | Incorrect connection url in Analytics configuration - **Description:**
Readme has incorrect connection url as tcp://localhost:7612/,tcp://localhost:7613/,tcp://localhost:7614/ and due to this it gets Number format exception.
It should be tcp://localhost:7612,tcp://localhost:7613,tcp://localhost:7614. Due to
https://github.com/wso2/carbon-apimgt/blob/v6.1.66/features/apimgt/org.wso2.carbon.apimgt.core.feature/src/main/resources/config/api-manager.xml#L138
**Suggested Labels:**
Minor, Readme
**Suggested Assignees:**
N/A
**Affected Product Version:**
2.1.x
**OS, DB, other environment details and versions:**
N/A
**Steps to reproduce:**
N/A
**Related Issues:**
N/A | priority | incorrect connection url in analytics configuration description readme has incorrect connection url as tcp localhost tcp localhost tcp localhost and due to this it gets number format exception it should be tcp localhost tcp localhost tcp localhost due to suggested labels minor readme suggested assignees n a affected product version x os db other environment details and versions n a steps to reproduce n a related issues n a | 1 |
562,026 | 16,636,625,340 | IssuesEvent | 2021-06-04 00:09:49 | kubernetes/website | https://api.github.com/repos/kubernetes/website | closed | Investigate “Persist Hugo resources Between Builds” plugin for Netlify | area/web-development lifecycle/rotten needs-triage priority/awaiting-more-evidence | **What would you like to be investigated**
Look into whether the [Persist Hugo resources Between Builds](https://github.com/cdeleeuwe/netlify-plugin-hugo-cache-resources) plugin for Netlify would speed up site builds; if so, is it worth employing for the site?
**Why is this needed**
The site build in Netlify takes a while. https://github.com/cdeleeuwe/netlify-plugin-hugo-cache-resources might speed it up.
**Comments**
<!-- Any additional related comments that might help. Drawings/mockups would be extremely helpful (if required). -->
/area web-development
/priority awaiting-more-evidence | 1.0 | Investigate “Persist Hugo resources Between Builds” plugin for Netlify - **What would you like to be investigated**
Look into whether the [Persist Hugo resources Between Builds](https://github.com/cdeleeuwe/netlify-plugin-hugo-cache-resources) plugin for Netlify would speed up site builds; if so, is it worth employing for the site?
**Why is this needed**
The site build in Netlify takes a while. https://github.com/cdeleeuwe/netlify-plugin-hugo-cache-resources might speed it up.
**Comments**
<!-- Any additional related comments that might help. Drawings/mockups would be extremely helpful (if required). -->
/area web-development
/priority awaiting-more-evidence | priority | investigate “persist hugo resources between builds” plugin for netlify what would you like to be investigated look into whether the plugin for netlify would speed up site builds if so is it worth employing for the site why is this needed the site build in netlify takes a while might speed it up comments area web development priority awaiting more evidence | 1 |
315,504 | 9,621,434,651 | IssuesEvent | 2019-05-14 10:37:42 | teambit/bit | https://api.github.com/repos/teambit/bit | opened | Bit overrides my readme file with an autogenerated file during export | area/export priority/critical type/bug | ## Expected Behavior
The readme file should not be changed.
## Actual Behavior
The readme file changed to a link file:
```
/* THIS IS A BIT-AUTO-GENERATED FILE. DO NOT EDIT THIS FILE DIRECTLY. */
module.exports = require('../../../../dist/my-comp/my-comp');
```
## Steps to Reproduce the Problem
```sh
bit init
mkdir my-comp
touch my-comp/my-comp.js
touch my-comp/my-comp.md
echo "some content" > my-comp/my-comp.md
bit add my-comp
bit import bit.envs/compilers/react -c
bit tag -a
bit export my-collection
```
When not using a compiler it didn't happen.
It might be related to the fact that the main file and the MD file are with the same name (didn't test it)
The exported file contains the correct content, so it only changes after the export itself.
## Specifications
- Bit version: 14.1.0
- Node version: 8.10.0
- npm / yarn version: 6.8.0
- Platform: mac os 10.14
- Bit compiler (include version): bit.envs/compilers/react@1.0.2
- Bit tester (include version): NA
| 1.0 | Bit overrides my readme file with an autogenerated file during export - ## Expected Behavior
The readme file should not be changed.
## Actual Behavior
The readme file changed to a link file:
```
/* THIS IS A BIT-AUTO-GENERATED FILE. DO NOT EDIT THIS FILE DIRECTLY. */
module.exports = require('../../../../dist/my-comp/my-comp');
```
## Steps to Reproduce the Problem
```sh
bit init
mkdir my-comp
touch my-comp/my-comp.js
touch my-comp/my-comp.md
echo "some content" > my-comp/my-comp.md
bit add my-comp
bit import bit.envs/compilers/react -c
bit tag -a
bit export my-collection
```
When not using a compiler it didn't happen.
It might be related to the fact that the main file and the MD file are with the same name (didn't test it)
The exported file contains the correct content, so it only changes after the export itself.
## Specifications
- Bit version: 14.1.0
- Node version: 8.10.0
- npm / yarn version: 6.8.0
- Platform: mac os 10.14
- Bit compiler (include version): bit.envs/compilers/react@1.0.2
- Bit tester (include version): NA
| priority | bit overrides my readme file with an autogenerated file during export expected behavior the readme file should not be changed actual behavior the readme file changed to a link file this is a bit auto generated file do not edit this file directly module exports require dist my comp my comp steps to reproduce the problem sh bit init mkdir my comp touch my comp my comp js touch my comp my comp md echo some content my comp my comp md bit add my comp bit import bit envs compilers react c bit tag a bit export my collection when not using a compiler it didn t happen it might be related to the fact that the main file and the md file are with the same name didn t test it the exported file contains the correct content so it only changes after the export itself specifications bit version node version npm yarn version platform mac os bit compiler include version bit envs compilers react bit tester include version na | 1 |
599,252 | 18,268,991,186 | IssuesEvent | 2021-10-04 11:52:36 | Matocolotoe/Skript-1.8 | https://api.github.com/repos/Matocolotoe/Skript-1.8 | closed | Problem of comparing ItemTypes | bug help wanted priority : high completed | For example, using the code below, the `ItemType#isTypeOf` method does not work as expected, with for example nether warts:
```applescript
on click on nether wart:
broadcast "test"
```
Related code: https://github.com/Matocolotoe/Skript-1.8/blob/master/src/main/java/ch/njol/skript/events/EvtClick.java#L197
The ItemStack of `o` equals `ItemStack{NETHER_WARTS x 1}` while the ItemStack of `block` equals `ItemStack{NETHER_STALK x 1}`, which means that the event does not work. | 1.0 | Problem of comparing ItemTypes - For example, using the code below, the `ItemType#isTypeOf` method does not work as expected, with for example nether warts:
```applescript
on click on nether wart:
broadcast "test"
```
Related code: https://github.com/Matocolotoe/Skript-1.8/blob/master/src/main/java/ch/njol/skript/events/EvtClick.java#L197
The ItemStack of `o` equals `ItemStack{NETHER_WARTS x 1}` while the ItemStack of `block` equals `ItemStack{NETHER_STALK x 1}`, which means that the event does not work. | priority | problem of comparing itemtypes for example using the code below the itemtype istypeof method does not work as expected with for example nether warts applescript on click on nether wart broadcast test related code the itemstack of o equals itemstack nether warts x while the itemstack of block equals itemstack nether stalk x which means that the event does not work | 1 |
55,373 | 30,720,313,958 | IssuesEvent | 2023-07-27 15:29:27 | praetorian-inc/noseyparker | https://api.github.com/repos/praetorian-inc/noseyparker | opened | Rework input enumeration to make it possible to enumerate Git repositories in parallel | performance content discovery | Currently, the `scan` command runs in two main phases: input enumeration and content scanning. Each of these phases runs in parallel (but not concurrently; the input enumeration phase completes entirely before the content scanning phase completes).
However, within the input enumeration phase, when a Git repository is discovered on the filesystem, that repository is enumerated sequentially, by a single thread. This becomes noticeable when you are scanning just a single huge repository, such as the Linux kernel, which has over a million commits, several million objects, and can take over a hundred GB of space when uncompressed.
It would be better if Nosey Parker did not have this sequential bottleneck, and was instead able to enumerate a single Git repository in parallel, using all available cores.
The implementation of this will be a bit tricky, requiring rework of the parallelism mechanism in the input enumerator code. That currently uses the [`ignore` crate](https://crates.io/crates/ignore) to do parallel filesystem walking, but that does not seem to expose its thread pool. We would want the proposed parallel Git enumerator to not oversubscribe the system running `scan`; the total number of enumeration threads should be controllable.
Additionally complicated will be figuring out how to build up the Git metadata graph that is being added in #66: the core graph data structure there is not designed for out-of-the-box mutation from many threads. | True | Rework input enumeration to make it possible to enumerate Git repositories in parallel - Currently, the `scan` command runs in two main phases: input enumeration and content scanning. Each of these phases runs in parallel (but not concurrently; the input enumeration phase completes entirely before the content scanning phase completes).
However, within the input enumeration phase, when a Git repository is discovered on the filesystem, that repository is enumerated sequentially, by a single thread. This becomes noticeable when you are scanning just a single huge repository, such as the Linux kernel, which has over a million commits, several million objects, and can take over a hundred GB of space when uncompressed.
It would be better if Nosey Parker did not have this sequential bottleneck, and was instead able to enumerate a single Git repository in parallel, using all available cores.
The implementation of this will be a bit tricky, requiring rework of the parallelism mechanism in the input enumerator code. That currently uses the [`ignore` crate](https://crates.io/crates/ignore) to do parallel filesystem walking, but that does not seem to expose its thread pool. We would want the proposed parallel Git enumerator to not oversubscribe the system running `scan`; the total number of enumeration threads should be controllable.
Additionally complicated will be figuring out how to build up the Git metadata graph that is being added in #66: the core graph data structure there is not designed for out-of-the-box mutation from many threads. | non_priority | rework input enumeration to make it possible to enumerate git repositories in parallel currently the scan command runs in two main phases input enumeration and content scanning each of these phases runs in parallel but not concurrently the input enumeration phase completes entirely before the content scanning phase completes however within the input enumeration phase when a git repository is discovered on the filesystem that repository is enumerated sequentially by a single thread this becomes noticeable when you are scanning just a single huge repository such as the linux kernel which has over a million commits several million objects and can take over a hundred gb of space when uncompressed it would be better if nosey parker did not have this sequential bottleneck and was instead able to enumerate a single git repository in parallel using all available cores the implementation of this will be a bit tricky requiring rework of the parallelism mechanism in the input enumerator code that currently uses the to do parallel filesystem walking but that does not seem to expose its thread pool we would want the proposed parallel git enumerator to not oversubscribe the system running scan the total number of enumeration threads should be controllable additionally complicated will be figuring out how to build up the git metadata graph that is being added in the core graph data structure there is not designed for out of the box mutation from many threads | 0 |
317,012 | 9,659,835,458 | IssuesEvent | 2019-05-20 14:16:58 | carbon-design-system/carbon-website | https://api.github.com/repos/carbon-design-system/carbon-website | closed | Improved component search and filter | inactive priority: low project: website type: dev :robot: wontfix | As a product designer I can discover and know I found the right component without knowing the component's name or clicking into each component's page.
User feedback:
- Feels like form elements should be next to each other
- She had no idea that they are alphabetical!!!!
- Q: How were you imagining that it would be? A: IDK? Grouping elements based on form/function
The only way to find a component is by its name.
- Add tags to components so users can filter by: Page type, platform, use case, language, etc.
| 1.0 | Improved component search and filter - As a product designer I can discover and know I found the right component without knowing the component's name or clicking into each component's page.
User feedback:
- Feels like form elements should be next to each other
- She had no idea that they are alphabetical!!!!
- Q: How were you imagining that it would be? A: IDK? Grouping elements based on form/function
The only way to find a component is by its name.
- Add tags to components so users can filter by: Page type, platform, use case, language, etc.
| priority | improved component search and filter as a product designer i can discover and know i found the right component without knowing the component s name or clicking into each component s page user feedback feels like form elements should be next to each other she had no idea that they are alphabetical q how were you imagining that it would be a idk grouping elements based on form function the only way to find a component is by its name add tags to components so users can filter by page type platform use case language etc | 1 |
56,641 | 14,078,468,698 | IssuesEvent | 2020-11-04 13:36:57 | themagicalmammal/android_kernel_samsung_s5neolte | https://api.github.com/repos/themagicalmammal/android_kernel_samsung_s5neolte | opened | CVE-2019-9003 (High) detected in linuxv3.10 | security vulnerability | ## CVE-2019-9003 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.10</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_s5neolte/commit/f978d7dbb980bbe5267a625da958c4226e1a8ae0">f978d7dbb980bbe5267a625da958c4226e1a8ae0</a></p>
<p>Found in base branch: <b>cosmic-experimental-1.6</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_s5neolte/drivers/char/ipmi/ipmi_msghandler.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_s5neolte/drivers/char/ipmi/ipmi_msghandler.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel before 4.20.5, attackers can trigger a drivers/char/ipmi/ipmi_msghandler.c use-after-free and OOPS by arranging for certain simultaneous execution of the code, as demonstrated by a "service ipmievd restart" loop.
<p>Publish Date: 2019-02-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-9003>CVE-2019-9003</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9003">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9003</a></p>
<p>Release Date: 2019-02-22</p>
<p>Fix Resolution: 4.20.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-9003 (High) detected in linuxv3.10 - ## CVE-2019-9003 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.10</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_s5neolte/commit/f978d7dbb980bbe5267a625da958c4226e1a8ae0">f978d7dbb980bbe5267a625da958c4226e1a8ae0</a></p>
<p>Found in base branch: <b>cosmic-experimental-1.6</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_s5neolte/drivers/char/ipmi/ipmi_msghandler.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_s5neolte/drivers/char/ipmi/ipmi_msghandler.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel before 4.20.5, attackers can trigger a drivers/char/ipmi/ipmi_msghandler.c use-after-free and OOPS by arranging for certain simultaneous execution of the code, as demonstrated by a "service ipmievd restart" loop.
<p>Publish Date: 2019-02-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-9003>CVE-2019-9003</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9003">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9003</a></p>
<p>Release Date: 2019-02-22</p>
<p>Fix Resolution: 4.20.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch cosmic experimental vulnerable source files android kernel samsung drivers char ipmi ipmi msghandler c android kernel samsung drivers char ipmi ipmi msghandler c vulnerability details in the linux kernel before attackers can trigger a drivers char ipmi ipmi msghandler c use after free and oops by arranging for certain simultaneous execution of the code as demonstrated by a service ipmievd restart loop publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
268,799 | 20,361,693,658 | IssuesEvent | 2022-02-20 19:30:05 | sqlalchemy/sqlalchemy | https://api.github.com/repos/sqlalchemy/sqlalchemy | closed | [docs] `make epub` results in error `TopLevelLookupException("Cant locate template for uri 'page.mako'")` | wontfix documentation use case | ### Describe the bug
I try to build the documentation into `epub` format to be able to read the docs on my ebook reader, as I couldn't find the pre-built `epub` anywhere on the https://www.sqlalchemy.org/.
Unfortunately, this fails with the error.
I tried `make html` to see if that's a general issue or only an issue with `epub`, and the `make html` works successfully.
### To Reproduce
```python
$ git clone git@github.com:sqlalchemy/sqlalchemy.git
$ cd sqlalchemy/doc/build/
$ git checkout rel_1_4_31
$ python3.10 -m venv venv
$ . venv/bin/activate
$ python --version
Python 3.10.0
$ pip install -r requirements.txt
$ pip install sphinx
$ pip freeze
alabaster==0.7.12
Babel==2.9.1
certifi==2021.10.8
changelog==0.5.8
charset-normalizer==2.0.12
docutils==0.17.1
idna==3.3
imagesize==1.3.0
Jinja2==3.0.3
libsass==0.21.0
Mako==1.1.6
MarkupSafe==2.1.0
packaging==21.3
Pygments==2.11.2
pyparsing==3.0.7
pytz==2021.3
requests==2.27.1
six==1.16.0
snowballstemmer==2.2.0
Sphinx==4.4.0
sphinx-paramlinks @ git+https://github.com/sqlalchemyorg/sphinx-paramlinks.git@acedb03149e3f87ff599174b033754c2f58f1c95
sphinxcontrib-applehelp==1.0.2
sphinxcontrib-devhelp==1.0.2
sphinxcontrib-htmlhelp==2.0.0
sphinxcontrib-jsmath==1.0.1
sphinxcontrib-qthelp==1.0.3
sphinxcontrib-serializinghtml==1.1.5
urllib3==1.26.8
zzzeeksphinx @ git+https://github.com/sqlalchemyorg/zzzeeksphinx.git@715c91f45c7b46e7c39528cae964f0d0891b421a
$ make epub
```
### Error
After `$ make epub`:
```
sphinx-build -b epub -d output/doctrees -T -j auto . output/epub
Running Sphinx v4.4.0
making output directory... done
building [mo]: targets for 0 po files that are out of date
building [epub]: targets for 168 source files that are out of date
updating environment: [new config] 168 added, 0 changed, 0 removed
reading sources... [100%] tutorial/engine .. venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/module
waiting for workers...
reading changelog note files (version 1.3.25)...[100%] 6392.rst
reading changelog note files (version 1.3.25)...[100%] 6392.rst
/Users/dstasczak/tmp/sqlalchemy/doc/build/tutorial/tutorial_nav_include.rst:9: ERROR: Undefined substitution referenced: "prev".
/Users/dstasczak/tmp/sqlalchemy/doc/build/tutorial/tutorial_nav_include.rst:9: ERROR: Undefined substitution referenced: "next".
/Users/dstasczak/tmp/sqlalchemy/doc/build/tutorial/tutorial_nav_include.rst:13: ERROR: Undefined substitution referenced: "next".
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/base.rst:3: ERROR: Error in "currentmodule" directive:
maximum 1 argument(s) allowed, 3 supplied.
.. currentmodule:: {{ module }}
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/class.rst:3: ERROR: Error in "currentmodule" directive:
maximum 1 argument(s) allowed, 3 supplied.
.. currentmodule:: {{ module }}
WARNING: invalid signature for autoclass ('{{ objname }}')
WARNING: don't know which module to import for autodocumenting '{{ objname }}' (try placing a "module" or "currentmodule" directive in the document, or giving an explicit module name)
WARNING: invalid signature for automodule ('{{ fullname }}')
WARNING: don't know which module to import for autodocumenting '{{ fullname }}' (try placing a "module" or "currentmodule" directive in the document, or giving an explicit module name)
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/module.rst:53: ERROR: Unknown directive type "autosummary".
.. autosummary::
:toctree:
:recursive:
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/module.rst:56: WARNING: Explicit markup ends without a blank line; unexpected unindent.
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/module.rst:58: WARNING: Definition list ends without a blank line; unexpected unindent.
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/ext/asyncio/scoping.py:docstring of sqlalchemy.ext.asyncio.session.connection:12: WARNING: Inline strong start-string without end-string.
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/ext/asyncio/scoping.py:docstring of sqlalchemy.ext.asyncio.session.stream:4: WARNING: Explicit markup ends without a blank line; unexpected unindent.
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/ext/asyncio/session.py:docstring of sqlalchemy.ext.asyncio.session.AsyncSession.connection:8: WARNING: Inline strong start-string without end-string.
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/orm/collections.py:docstring of sqlalchemy.orm.collections.collection:1: WARNING: duplicate object description of sqlalchemy.orm.collections.collection, other instance in orm/collections, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/doc/build/orm/contextual.rst:4: WARNING: Duplicate explicit target name: "threading.local()".
/Users/dstasczak/tmp/sqlalchemy/doc/build/orm/extensions/mypy.rst:4: WARNING: Duplicate explicit target name: "mypy".
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/util/langhelpers.py:docstring of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute:1: WARNING: duplicate object description of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute, other instance in orm/internals, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/util/langhelpers.py:docstring of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute:1: WARNING: duplicate object description of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute, other instance in core/foundation, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/util/langhelpers.py:docstring of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute:1: WARNING: duplicate object description of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute, other instance in core/metadata, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/doc/build/core/reflection.rst:142: ERROR: Unknown interpreted text role "method".
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/util/langhelpers.py:docstring of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute:1: WARNING: duplicate object description of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute, other instance in core/metadata, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/util/langhelpers.py:docstring of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute:1: WARNING: duplicate object description of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute, other instance in core/selectable, use :noindex: for one of them
docstring of sqlalchemy.sql.expression.LABEL_STYLE_TABLENAME_PLUS_COL:6: WARNING: Inline literal start-string without end-string.
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.CHAR:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.CHAR, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.NCHAR:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.NCHAR, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.NVARCHAR:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.NVARCHAR, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.TEXT:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.TEXT, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.VARCHAR:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.VARCHAR, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.BINARY:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.BINARY, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.BLOB:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.BLOB, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.BOOLEAN:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.BOOLEAN, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.DATE:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.DATE, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.VARBINARY:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.VARBINARY, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.REAL:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.REAL, other instance in core/type_basics, use :noindex: for one of them
changelog/changelog_07.rst:5: WARNING: Explicit markup ends without a blank line; unexpected unindent.
/Users/dstasczak/tmp/sqlalchemy/doc/build/changelog/changelog_11.rst:1: WARNING: Inline emphasis start-string without end-string.
looking for now-outdated files... none found
pickling environment... done
checking consistency... /Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/Jinja2-3.0.3.dist-info/LICENSE.rst: WARNING: document isn't included in any toctree
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/MarkupSafe-2.1.0.dist-info/LICENSE.rst: WARNING: document isn't included in any toctree
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/alabaster-0.7.12.dist-info/DESCRIPTION.rst: WARNING: document isn't included in any toctree
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/imagesize-1.3.0.dist-info/LICENSE.rst: WARNING: document isn't included in any toctree
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/base.rst: WARNING: document isn't included in any toctree
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/class.rst: WARNING: document isn't included in any toctree
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/module.rst: WARNING: document isn't included in any toctree
done
preparing documents... done
The name of the builder is: epubThe name of the builder is: epub
Traceback (most recent call last):
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/mako/lookup.py", line 249, in get_template
return self._check(uri, self._collection[uri])
KeyError: 'page.mako'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/html/__init__.py", line 1055, in handle_page
output = self.templates.render(templatename, ctx)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/zzzeeksphinx/mako.py", line 102, in render
return self.lookup.get_template(template).render_unicode(**context)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/mako/lookup.py", line 262, in get_template
raise exceptions.TopLevelLookupException(
mako.exceptions.TopLevelLookupException: Cant locate template for uri 'page.mako'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/cmd/build.py", line 284, in build_main
app.build(args.force_all, filenames)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/application.py", line 337, in build
self.builder.build_update()
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/__init__.py", line 294, in build_update
self.build(to_build,
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/__init__.py", line 358, in build
self.write(docnames, list(updated_docnames), method)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/__init__.py", line 529, in write
self._write_parallel(sorted(docnames),
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/__init__.py", line 556, in _write_parallel
self.write_doc(firstname, doctree)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/_epub_base.py", line 374, in write_doc
super().write_doc(docname, doctree)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/html/__init__.py", line 632, in write_doc
self.handle_page(docname, ctx, event_arg=doctree)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/_epub_base.py", line 467, in handle_page
super().handle_page(pagename, addctx, templatename, outfilename, event_arg)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/html/__init__.py", line 1062, in handle_page
raise ThemeError(__("An error happened in rendering the page %s.\nReason: %r") %
sphinx.errors.ThemeError: An error happened in rendering the page changelog/changelog_01.
Reason: TopLevelLookupException("Cant locate template for uri 'page.mako'")
Theme error:
An error happened in rendering the page changelog/changelog_01.
Reason: TopLevelLookupException("Cant locate template for uri 'page.mako'")
make: *** [epub] Error 2
```
### Versions
- OS: macOS Mojave Version 10.14.6 (18G9323)
- Python: 3.10.0
- SQLAlchemy: `rel_1_4_31`
- Database: None
- DBAPI (eg: psycopg, cx_oracle, mysqlclient): None
### Additional context
_No response_ | 1.0 | [docs] `make epub` results in error `TopLevelLookupException("Cant locate template for uri 'page.mako'")` - ### Describe the bug
I try to build the documentation into `epub` format to be able to read the docs on my ebook reader, as I couldn't find the pre-built `epub` anywhere on the https://www.sqlalchemy.org/.
Unfortunately, this fails with the error.
I tried `make html` to see if that's a general issue or only an issue with `epub`, and the `make html` works successfully.
### To Reproduce
```python
$ git clone git@github.com:sqlalchemy/sqlalchemy.git
$ cd sqlalchemy/doc/build/
$ git checkout rel_1_4_31
$ python3.10 -m venv venv
$ . venv/bin/activate
$ python --version
Python 3.10.0
$ pip install -r requirements.txt
$ pip install sphinx
$ pip freeze
alabaster==0.7.12
Babel==2.9.1
certifi==2021.10.8
changelog==0.5.8
charset-normalizer==2.0.12
docutils==0.17.1
idna==3.3
imagesize==1.3.0
Jinja2==3.0.3
libsass==0.21.0
Mako==1.1.6
MarkupSafe==2.1.0
packaging==21.3
Pygments==2.11.2
pyparsing==3.0.7
pytz==2021.3
requests==2.27.1
six==1.16.0
snowballstemmer==2.2.0
Sphinx==4.4.0
sphinx-paramlinks @ git+https://github.com/sqlalchemyorg/sphinx-paramlinks.git@acedb03149e3f87ff599174b033754c2f58f1c95
sphinxcontrib-applehelp==1.0.2
sphinxcontrib-devhelp==1.0.2
sphinxcontrib-htmlhelp==2.0.0
sphinxcontrib-jsmath==1.0.1
sphinxcontrib-qthelp==1.0.3
sphinxcontrib-serializinghtml==1.1.5
urllib3==1.26.8
zzzeeksphinx @ git+https://github.com/sqlalchemyorg/zzzeeksphinx.git@715c91f45c7b46e7c39528cae964f0d0891b421a
$ make epub
```
### Error
After `$ make epub`:
```
sphinx-build -b epub -d output/doctrees -T -j auto . output/epub
Running Sphinx v4.4.0
making output directory... done
building [mo]: targets for 0 po files that are out of date
building [epub]: targets for 168 source files that are out of date
updating environment: [new config] 168 added, 0 changed, 0 removed
reading sources... [100%] tutorial/engine .. venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/module
waiting for workers...
reading changelog note files (version 1.3.25)...[100%] 6392.rst
reading changelog note files (version 1.3.25)...[100%] 6392.rst
/Users/dstasczak/tmp/sqlalchemy/doc/build/tutorial/tutorial_nav_include.rst:9: ERROR: Undefined substitution referenced: "prev".
/Users/dstasczak/tmp/sqlalchemy/doc/build/tutorial/tutorial_nav_include.rst:9: ERROR: Undefined substitution referenced: "next".
/Users/dstasczak/tmp/sqlalchemy/doc/build/tutorial/tutorial_nav_include.rst:13: ERROR: Undefined substitution referenced: "next".
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/base.rst:3: ERROR: Error in "currentmodule" directive:
maximum 1 argument(s) allowed, 3 supplied.
.. currentmodule:: {{ module }}
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/class.rst:3: ERROR: Error in "currentmodule" directive:
maximum 1 argument(s) allowed, 3 supplied.
.. currentmodule:: {{ module }}
WARNING: invalid signature for autoclass ('{{ objname }}')
WARNING: don't know which module to import for autodocumenting '{{ objname }}' (try placing a "module" or "currentmodule" directive in the document, or giving an explicit module name)
WARNING: invalid signature for automodule ('{{ fullname }}')
WARNING: don't know which module to import for autodocumenting '{{ fullname }}' (try placing a "module" or "currentmodule" directive in the document, or giving an explicit module name)
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/module.rst:53: ERROR: Unknown directive type "autosummary".
.. autosummary::
:toctree:
:recursive:
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/module.rst:56: WARNING: Explicit markup ends without a blank line; unexpected unindent.
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/module.rst:58: WARNING: Definition list ends without a blank line; unexpected unindent.
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/ext/asyncio/scoping.py:docstring of sqlalchemy.ext.asyncio.session.connection:12: WARNING: Inline strong start-string without end-string.
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/ext/asyncio/scoping.py:docstring of sqlalchemy.ext.asyncio.session.stream:4: WARNING: Explicit markup ends without a blank line; unexpected unindent.
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/ext/asyncio/session.py:docstring of sqlalchemy.ext.asyncio.session.AsyncSession.connection:8: WARNING: Inline strong start-string without end-string.
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/orm/collections.py:docstring of sqlalchemy.orm.collections.collection:1: WARNING: duplicate object description of sqlalchemy.orm.collections.collection, other instance in orm/collections, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/doc/build/orm/contextual.rst:4: WARNING: Duplicate explicit target name: "threading.local()".
/Users/dstasczak/tmp/sqlalchemy/doc/build/orm/extensions/mypy.rst:4: WARNING: Duplicate explicit target name: "mypy".
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/util/langhelpers.py:docstring of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute:1: WARNING: duplicate object description of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute, other instance in orm/internals, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/util/langhelpers.py:docstring of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute:1: WARNING: duplicate object description of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute, other instance in core/foundation, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/util/langhelpers.py:docstring of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute:1: WARNING: duplicate object description of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute, other instance in core/metadata, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/doc/build/core/reflection.rst:142: ERROR: Unknown interpreted text role "method".
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/util/langhelpers.py:docstring of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute:1: WARNING: duplicate object description of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute, other instance in core/metadata, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/util/langhelpers.py:docstring of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute:1: WARNING: duplicate object description of sqlalchemy.util.langhelpers.HasMemoized.memoized_attribute, other instance in core/selectable, use :noindex: for one of them
docstring of sqlalchemy.sql.expression.LABEL_STYLE_TABLENAME_PLUS_COL:6: WARNING: Inline literal start-string without end-string.
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.CHAR:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.CHAR, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.NCHAR:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.NCHAR, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.NVARCHAR:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.NVARCHAR, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.TEXT:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.TEXT, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.VARCHAR:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.VARCHAR, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.BINARY:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.BINARY, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.BLOB:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.BLOB, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.BOOLEAN:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.BOOLEAN, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.DATE:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.DATE, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.VARBINARY:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.VARBINARY, other instance in core/type_basics, use :noindex: for one of them
/Users/dstasczak/tmp/sqlalchemy/lib/sqlalchemy/sql/sqltypes.py:docstring of sqlalchemy.sql.sqltypes.REAL:1: WARNING: duplicate object description of sqlalchemy.sql.sqltypes.REAL, other instance in core/type_basics, use :noindex: for one of them
changelog/changelog_07.rst:5: WARNING: Explicit markup ends without a blank line; unexpected unindent.
/Users/dstasczak/tmp/sqlalchemy/doc/build/changelog/changelog_11.rst:1: WARNING: Inline emphasis start-string without end-string.
looking for now-outdated files... none found
pickling environment... done
checking consistency... /Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/Jinja2-3.0.3.dist-info/LICENSE.rst: WARNING: document isn't included in any toctree
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/MarkupSafe-2.1.0.dist-info/LICENSE.rst: WARNING: document isn't included in any toctree
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/alabaster-0.7.12.dist-info/DESCRIPTION.rst: WARNING: document isn't included in any toctree
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/imagesize-1.3.0.dist-info/LICENSE.rst: WARNING: document isn't included in any toctree
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/base.rst: WARNING: document isn't included in any toctree
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/class.rst: WARNING: document isn't included in any toctree
/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/ext/autosummary/templates/autosummary/module.rst: WARNING: document isn't included in any toctree
done
preparing documents... done
The name of the builder is: epubThe name of the builder is: epub
Traceback (most recent call last):
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/mako/lookup.py", line 249, in get_template
return self._check(uri, self._collection[uri])
KeyError: 'page.mako'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/html/__init__.py", line 1055, in handle_page
output = self.templates.render(templatename, ctx)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/zzzeeksphinx/mako.py", line 102, in render
return self.lookup.get_template(template).render_unicode(**context)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/mako/lookup.py", line 262, in get_template
raise exceptions.TopLevelLookupException(
mako.exceptions.TopLevelLookupException: Cant locate template for uri 'page.mako'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/cmd/build.py", line 284, in build_main
app.build(args.force_all, filenames)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/application.py", line 337, in build
self.builder.build_update()
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/__init__.py", line 294, in build_update
self.build(to_build,
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/__init__.py", line 358, in build
self.write(docnames, list(updated_docnames), method)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/__init__.py", line 529, in write
self._write_parallel(sorted(docnames),
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/__init__.py", line 556, in _write_parallel
self.write_doc(firstname, doctree)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/_epub_base.py", line 374, in write_doc
super().write_doc(docname, doctree)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/html/__init__.py", line 632, in write_doc
self.handle_page(docname, ctx, event_arg=doctree)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/_epub_base.py", line 467, in handle_page
super().handle_page(pagename, addctx, templatename, outfilename, event_arg)
File "/Users/dstasczak/tmp/sqlalchemy/doc/build/venv/lib/python3.10/site-packages/sphinx/builders/html/__init__.py", line 1062, in handle_page
raise ThemeError(__("An error happened in rendering the page %s.\nReason: %r") %
sphinx.errors.ThemeError: An error happened in rendering the page changelog/changelog_01.
Reason: TopLevelLookupException("Cant locate template for uri 'page.mako'")
Theme error:
An error happened in rendering the page changelog/changelog_01.
Reason: TopLevelLookupException("Cant locate template for uri 'page.mako'")
make: *** [epub] Error 2
```
### Versions
- OS: macOS Mojave Version 10.14.6 (18G9323)
- Python: 3.10.0
- SQLAlchemy: `rel_1_4_31`
- Database: None
- DBAPI (eg: psycopg, cx_oracle, mysqlclient): None
### Additional context
_No response_ | non_priority | make epub results in error toplevellookupexception cant locate template for uri page mako describe the bug i try to build the documentation into epub format to be able to read the docs on my ebook reader as i couldn t find the pre built epub anywhere on the unfortunately this fails with the error i tried make html to see if that s a general issue or only an issue with epub and the make html works successfully to reproduce python git clone git github com sqlalchemy sqlalchemy git cd sqlalchemy doc build git checkout rel m venv venv venv bin activate python version python pip install r requirements txt pip install sphinx pip freeze alabaster babel certifi changelog charset normalizer docutils idna imagesize libsass mako markupsafe packaging pygments pyparsing pytz requests six snowballstemmer sphinx sphinx paramlinks git sphinxcontrib applehelp sphinxcontrib devhelp sphinxcontrib htmlhelp sphinxcontrib jsmath sphinxcontrib qthelp sphinxcontrib serializinghtml zzzeeksphinx git make epub error after make epub sphinx build b epub d output doctrees t j auto output epub running sphinx making output directory done building targets for po files that are out of date building targets for source files that are out of date updating environment added changed removed reading sources tutorial engine venv lib site packages sphinx ext autosummary templates autosummary module waiting for workers reading changelog note files version rst reading changelog note files version rst users dstasczak tmp sqlalchemy doc build tutorial tutorial nav include rst error undefined substitution referenced prev users dstasczak tmp sqlalchemy doc build tutorial tutorial nav include rst error undefined substitution referenced next users dstasczak tmp sqlalchemy doc build tutorial tutorial nav include rst error undefined substitution referenced next users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx ext autosummary templates autosummary base rst error error in currentmodule directive maximum argument s allowed supplied currentmodule module users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx ext autosummary templates autosummary class rst error error in currentmodule directive maximum argument s allowed supplied currentmodule module warning invalid signature for autoclass objname warning don t know which module to import for autodocumenting objname try placing a module or currentmodule directive in the document or giving an explicit module name warning invalid signature for automodule fullname warning don t know which module to import for autodocumenting fullname try placing a module or currentmodule directive in the document or giving an explicit module name users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx ext autosummary templates autosummary module rst error unknown directive type autosummary autosummary toctree recursive users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx ext autosummary templates autosummary module rst warning explicit markup ends without a blank line unexpected unindent users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx ext autosummary templates autosummary module rst warning definition list ends without a blank line unexpected unindent users dstasczak tmp sqlalchemy lib sqlalchemy ext asyncio scoping py docstring of sqlalchemy ext asyncio session connection warning inline strong start string without end string users dstasczak tmp sqlalchemy lib sqlalchemy ext asyncio scoping py docstring of sqlalchemy ext asyncio session stream warning explicit markup ends without a blank line unexpected unindent users dstasczak tmp sqlalchemy lib sqlalchemy ext asyncio session py docstring of sqlalchemy ext asyncio session asyncsession connection warning inline strong start string without end string users dstasczak tmp sqlalchemy lib sqlalchemy orm collections py docstring of sqlalchemy orm collections collection warning duplicate object description of sqlalchemy orm collections collection other instance in orm collections use noindex for one of them users dstasczak tmp sqlalchemy doc build orm contextual rst warning duplicate explicit target name threading local users dstasczak tmp sqlalchemy doc build orm extensions mypy rst warning duplicate explicit target name mypy users dstasczak tmp sqlalchemy lib sqlalchemy util langhelpers py docstring of sqlalchemy util langhelpers hasmemoized memoized attribute warning duplicate object description of sqlalchemy util langhelpers hasmemoized memoized attribute other instance in orm internals use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy util langhelpers py docstring of sqlalchemy util langhelpers hasmemoized memoized attribute warning duplicate object description of sqlalchemy util langhelpers hasmemoized memoized attribute other instance in core foundation use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy util langhelpers py docstring of sqlalchemy util langhelpers hasmemoized memoized attribute warning duplicate object description of sqlalchemy util langhelpers hasmemoized memoized attribute other instance in core metadata use noindex for one of them users dstasczak tmp sqlalchemy doc build core reflection rst error unknown interpreted text role method users dstasczak tmp sqlalchemy lib sqlalchemy util langhelpers py docstring of sqlalchemy util langhelpers hasmemoized memoized attribute warning duplicate object description of sqlalchemy util langhelpers hasmemoized memoized attribute other instance in core metadata use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy util langhelpers py docstring of sqlalchemy util langhelpers hasmemoized memoized attribute warning duplicate object description of sqlalchemy util langhelpers hasmemoized memoized attribute other instance in core selectable use noindex for one of them docstring of sqlalchemy sql expression label style tablename plus col warning inline literal start string without end string users dstasczak tmp sqlalchemy lib sqlalchemy sql sqltypes py docstring of sqlalchemy sql sqltypes char warning duplicate object description of sqlalchemy sql sqltypes char other instance in core type basics use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy sql sqltypes py docstring of sqlalchemy sql sqltypes nchar warning duplicate object description of sqlalchemy sql sqltypes nchar other instance in core type basics use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy sql sqltypes py docstring of sqlalchemy sql sqltypes nvarchar warning duplicate object description of sqlalchemy sql sqltypes nvarchar other instance in core type basics use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy sql sqltypes py docstring of sqlalchemy sql sqltypes text warning duplicate object description of sqlalchemy sql sqltypes text other instance in core type basics use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy sql sqltypes py docstring of sqlalchemy sql sqltypes varchar warning duplicate object description of sqlalchemy sql sqltypes varchar other instance in core type basics use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy sql sqltypes py docstring of sqlalchemy sql sqltypes binary warning duplicate object description of sqlalchemy sql sqltypes binary other instance in core type basics use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy sql sqltypes py docstring of sqlalchemy sql sqltypes blob warning duplicate object description of sqlalchemy sql sqltypes blob other instance in core type basics use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy sql sqltypes py docstring of sqlalchemy sql sqltypes boolean warning duplicate object description of sqlalchemy sql sqltypes boolean other instance in core type basics use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy sql sqltypes py docstring of sqlalchemy sql sqltypes date warning duplicate object description of sqlalchemy sql sqltypes date other instance in core type basics use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy sql sqltypes py docstring of sqlalchemy sql sqltypes varbinary warning duplicate object description of sqlalchemy sql sqltypes varbinary other instance in core type basics use noindex for one of them users dstasczak tmp sqlalchemy lib sqlalchemy sql sqltypes py docstring of sqlalchemy sql sqltypes real warning duplicate object description of sqlalchemy sql sqltypes real other instance in core type basics use noindex for one of them changelog changelog rst warning explicit markup ends without a blank line unexpected unindent users dstasczak tmp sqlalchemy doc build changelog changelog rst warning inline emphasis start string without end string looking for now outdated files none found pickling environment done checking consistency users dstasczak tmp sqlalchemy doc build venv lib site packages dist info license rst warning document isn t included in any toctree users dstasczak tmp sqlalchemy doc build venv lib site packages markupsafe dist info license rst warning document isn t included in any toctree users dstasczak tmp sqlalchemy doc build venv lib site packages alabaster dist info description rst warning document isn t included in any toctree users dstasczak tmp sqlalchemy doc build venv lib site packages imagesize dist info license rst warning document isn t included in any toctree users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx ext autosummary templates autosummary base rst warning document isn t included in any toctree users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx ext autosummary templates autosummary class rst warning document isn t included in any toctree users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx ext autosummary templates autosummary module rst warning document isn t included in any toctree done preparing documents done the name of the builder is epubthe name of the builder is epub traceback most recent call last file users dstasczak tmp sqlalchemy doc build venv lib site packages mako lookup py line in get template return self check uri self collection keyerror page mako during handling of the above exception another exception occurred traceback most recent call last file users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx builders html init py line in handle page output self templates render templatename ctx file users dstasczak tmp sqlalchemy doc build venv lib site packages zzzeeksphinx mako py line in render return self lookup get template template render unicode context file users dstasczak tmp sqlalchemy doc build venv lib site packages mako lookup py line in get template raise exceptions toplevellookupexception mako exceptions toplevellookupexception cant locate template for uri page mako the above exception was the direct cause of the following exception traceback most recent call last file users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx cmd build py line in build main app build args force all filenames file users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx application py line in build self builder build update file users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx builders init py line in build update self build to build file users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx builders init py line in build self write docnames list updated docnames method file users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx builders init py line in write self write parallel sorted docnames file users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx builders init py line in write parallel self write doc firstname doctree file users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx builders epub base py line in write doc super write doc docname doctree file users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx builders html init py line in write doc self handle page docname ctx event arg doctree file users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx builders epub base py line in handle page super handle page pagename addctx templatename outfilename event arg file users dstasczak tmp sqlalchemy doc build venv lib site packages sphinx builders html init py line in handle page raise themeerror an error happened in rendering the page s nreason r sphinx errors themeerror an error happened in rendering the page changelog changelog reason toplevellookupexception cant locate template for uri page mako theme error an error happened in rendering the page changelog changelog reason toplevellookupexception cant locate template for uri page mako make error versions os macos mojave version python sqlalchemy rel database none dbapi eg psycopg cx oracle mysqlclient none additional context no response | 0 |
54,056 | 13,247,072,056 | IssuesEvent | 2020-08-19 16:39:01 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Implement to parse and store the syntax tree API | Area/BuildTools Points/3 Type/Task | **Description:**
Parse the ballerina files to build and give a syntax tree
| 1.0 | Implement to parse and store the syntax tree API - **Description:**
Parse the ballerina files to build and give a syntax tree
| non_priority | implement to parse and store the syntax tree api description parse the ballerina files to build and give a syntax tree | 0 |
31,247 | 11,893,280,814 | IssuesEvent | 2020-03-29 10:47:31 | nihalmurmu/automata | https://api.github.com/repos/nihalmurmu/automata | closed | WS-2016-0075 (Medium) detected in moment-2.8.4.min.js | security vulnerability | ## WS-2016-0075 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moment-2.8.4.min.js</b></p></summary>
<p>Parse, validate, manipulate, and display dates</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/moment.js/2.8.4/moment.min.js">https://cdnjs.cloudflare.com/ajax/libs/moment.js/2.8.4/moment.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/automata/node_modules/vis/examples/graph2d/08_performance.html</p>
<p>Path to vulnerable library: /automata/node_modules/vis/examples/graph2d/08_performance.html,/automata/node_modules/vis/examples/timeline/other/stressPerformance.html</p>
<p>
Dependency Hierarchy:
- :x: **moment-2.8.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nihalmurmu/automata/commit/9c21bcf8921fe44ec20fe8844090100903ceec5d">9c21bcf8921fe44ec20fe8844090100903ceec5d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Regular expression denial of service vulnerability in the moment package, by using a specific 40 characters long string in the "format" method.
<p>Publish Date: 2016-10-24
<p>URL: <a href=https://github.com/moment/moment/commit/663f33e333212b3800b63592cd8e237ac8fabdb9>WS-2016-0075</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/moment/moment/commit/663f33e333212b3800b63592cd8e237ac8fabdb9">https://github.com/moment/moment/commit/663f33e333212b3800b63592cd8e237ac8fabdb9</a></p>
<p>Release Date: 2016-10-24</p>
<p>Fix Resolution: Replace or update the following files: month.js, lt.js</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2016-0075 (Medium) detected in moment-2.8.4.min.js - ## WS-2016-0075 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moment-2.8.4.min.js</b></p></summary>
<p>Parse, validate, manipulate, and display dates</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/moment.js/2.8.4/moment.min.js">https://cdnjs.cloudflare.com/ajax/libs/moment.js/2.8.4/moment.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/automata/node_modules/vis/examples/graph2d/08_performance.html</p>
<p>Path to vulnerable library: /automata/node_modules/vis/examples/graph2d/08_performance.html,/automata/node_modules/vis/examples/timeline/other/stressPerformance.html</p>
<p>
Dependency Hierarchy:
- :x: **moment-2.8.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nihalmurmu/automata/commit/9c21bcf8921fe44ec20fe8844090100903ceec5d">9c21bcf8921fe44ec20fe8844090100903ceec5d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Regular expression denial of service vulnerability in the moment package, by using a specific 40 characters long string in the "format" method.
<p>Publish Date: 2016-10-24
<p>URL: <a href=https://github.com/moment/moment/commit/663f33e333212b3800b63592cd8e237ac8fabdb9>WS-2016-0075</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/moment/moment/commit/663f33e333212b3800b63592cd8e237ac8fabdb9">https://github.com/moment/moment/commit/663f33e333212b3800b63592cd8e237ac8fabdb9</a></p>
<p>Release Date: 2016-10-24</p>
<p>Fix Resolution: Replace or update the following files: month.js, lt.js</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws medium detected in moment min js ws medium severity vulnerability vulnerable library moment min js parse validate manipulate and display dates library home page a href path to dependency file tmp ws scm automata node modules vis examples performance html path to vulnerable library automata node modules vis examples performance html automata node modules vis examples timeline other stressperformance html dependency hierarchy x moment min js vulnerable library found in head commit a href vulnerability details regular expression denial of service vulnerability in the moment package by using a specific characters long string in the format method publish date url a href cvss score details base score metrics not available suggested fix type change files origin a href release date fix resolution replace or update the following files month js lt js step up your open source security game with whitesource | 0 |
19,603 | 11,254,685,527 | IssuesEvent | 2020-01-12 01:59:41 | tktaofik/airnd-market | https://api.github.com/repos/tktaofik/airnd-market | closed | Bootstrap user service | user-service | - [x] start app with makefile
- [x] postgreSQL db
- [x] basic api routes
- [x] health endpoints
- [x] dockerfile
- [x] docker compose file
- [x] start app with docker-compose | 1.0 | Bootstrap user service - - [x] start app with makefile
- [x] postgreSQL db
- [x] basic api routes
- [x] health endpoints
- [x] dockerfile
- [x] docker compose file
- [x] start app with docker-compose | non_priority | bootstrap user service start app with makefile postgresql db basic api routes health endpoints dockerfile docker compose file start app with docker compose | 0 |
531,484 | 15,499,101,338 | IssuesEvent | 2021-03-11 07:29:02 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.youtube.com - see bug description | browser-firefox engine-gecko priority-critical | <!-- @browser: Firefox 88.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:88.0) Gecko/20100101 Firefox/88.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/68112 -->
**URL**: https://www.youtube.com/results?search_query=hello+darkness+my+old+friend
**Browser / Version**: Firefox 88.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: I can't use "space" in the searchfield. When searching while a video runs, "space" in the searchfield causes the video to stop.
**Steps to Reproduce**:
Go to youtube.com and type in the searchfield. I can't use "space".
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/3/36721e64-7bce-4464-915c-310fb8d0abd7.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210310093927</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/3/ec5b5d9f-a996-415e-8e33-0fb016b27039)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.youtube.com - see bug description - <!-- @browser: Firefox 88.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:88.0) Gecko/20100101 Firefox/88.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/68112 -->
**URL**: https://www.youtube.com/results?search_query=hello+darkness+my+old+friend
**Browser / Version**: Firefox 88.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: I can't use "space" in the searchfield. When searching while a video runs, "space" in the searchfield causes the video to stop.
**Steps to Reproduce**:
Go to youtube.com and type in the searchfield. I can't use "space".
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/3/36721e64-7bce-4464-915c-310fb8d0abd7.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210310093927</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/3/ec5b5d9f-a996-415e-8e33-0fb016b27039)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | see bug description url browser version firefox operating system windows tested another browser yes chrome problem type something else description i can t use space in the searchfield when searching while a video runs space in the searchfield causes the video to stop steps to reproduce go to youtube com and type in the searchfield i can t use space view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
322,635 | 9,820,741,112 | IssuesEvent | 2019-06-14 04:10:42 | oh-my-fish/oh-my-fish | https://api.github.com/repos/oh-my-fish/oh-my-fish | closed | Installer script checksum is out of sync | installer priority: high | You forgot to update the installer checksum in the latest install script update.
The new hash should be
```
bbace7ef16956d87fd40bff91cd1992a90621e7931ac3055f16b7f6d679e8fff install
```
as @foxcpp mentioned too under the [commit](https://github.com/oh-my-fish/oh-my-fish/commit/a4b2f1cfaac12a614c491e3b8c7ea7e0b8842eee).
| 1.0 | Installer script checksum is out of sync - You forgot to update the installer checksum in the latest install script update.
The new hash should be
```
bbace7ef16956d87fd40bff91cd1992a90621e7931ac3055f16b7f6d679e8fff install
```
as @foxcpp mentioned too under the [commit](https://github.com/oh-my-fish/oh-my-fish/commit/a4b2f1cfaac12a614c491e3b8c7ea7e0b8842eee).
| priority | installer script checksum is out of sync you forgot to update the installer checksum in the latest install script update the new hash should be install as foxcpp mentioned too under the | 1 |
491,416 | 14,163,381,158 | IssuesEvent | 2020-11-12 02:14:05 | codidact/qpixel | https://api.github.com/repos/codidact/qpixel | opened | Expose a network profile | area: backend area: frontend complexity: unassessed priority: medium type: change request | You have a network *account*, and a *profile* on each community that you've visited while logged in. But there's nothing that ties them together; to see your posts (or other information) on a community you first have to visit that community and then click on your user link.
Can we add a network profile that, at minimum, lists and links to your community profiles? This network profile would also be a good place to attach network-wide preferences when we get to them. And we have a [request for a unified subscription list](https://meta.codidact.com/questions/276419), which could be made available here. (I'll file a separate issue for that. This current one is about enablement -- we need a network profile at *all* before we can do stuff on it.)
The network profile should be linked from the community profile and publicly visible.
| 1.0 | Expose a network profile - You have a network *account*, and a *profile* on each community that you've visited while logged in. But there's nothing that ties them together; to see your posts (or other information) on a community you first have to visit that community and then click on your user link.
Can we add a network profile that, at minimum, lists and links to your community profiles? This network profile would also be a good place to attach network-wide preferences when we get to them. And we have a [request for a unified subscription list](https://meta.codidact.com/questions/276419), which could be made available here. (I'll file a separate issue for that. This current one is about enablement -- we need a network profile at *all* before we can do stuff on it.)
The network profile should be linked from the community profile and publicly visible.
| priority | expose a network profile you have a network account and a profile on each community that you ve visited while logged in but there s nothing that ties them together to see your posts or other information on a community you first have to visit that community and then click on your user link can we add a network profile that at minimum lists and links to your community profiles this network profile would also be a good place to attach network wide preferences when we get to them and we have a which could be made available here i ll file a separate issue for that this current one is about enablement we need a network profile at all before we can do stuff on it the network profile should be linked from the community profile and publicly visible | 1 |
828,213 | 31,816,718,842 | IssuesEvent | 2023-09-13 21:06:04 | AUBGTheHUB/monolith | https://api.github.com/repos/AUBGTheHUB/monolith | closed | [BUG] HTTPSconnection error when accessing /v2/logs | bug high priority python | ## Brief description:
There is a general problem when doing POST requests:


## How to achieve it:
Fix it
| 1.0 | [BUG] HTTPSconnection error when accessing /v2/logs - ## Brief description:
There is a general problem when doing POST requests:


## How to achieve it:
Fix it
| priority | httpsconnection error when accessing logs brief description there is a general problem when doing post requests how to achieve it fix it | 1 |
53,861 | 13,262,397,723 | IssuesEvent | 2020-08-20 21:42:34 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | lilliput attemtps to use nonexistent project rock_bottom (Trac #2206) | Migrated from Trac defect other | I get this error when I run cmake (py3-v4, combo r166833):
```text
-- + lilliput
-- +-- python [symlinks]
-- + ipdf support
-- + Minuit2 support
-- - multinest support...not found
-- + NLopt support
CMake Error at cmake/project.cmake:94 (message):
Attempt to use nonexistent project 'rock_bottom'
Call Stack (most recent call first):
cmake/project.cmake:249 (use_projects)
lilliput/CMakeLists.txt:126 (i3_add_library)
-- Configuring incomplete, errors occurred!
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2206">https://code.icecube.wisc.edu/projects/icecube/ticket/2206</a>, reported by thomas.kittlerand owned by kkrings</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2018-11-16T15:25:14",
"_ts": "1542381914248311",
"description": "I get this error when I run cmake (py3-v4, combo r166833):\n{{{\n-- + lilliput\n-- +-- python [symlinks] \n-- + ipdf support\n-- + Minuit2 support\n-- - multinest support...not found\n-- + NLopt support\nCMake Error at cmake/project.cmake:94 (message):\n Attempt to use nonexistent project 'rock_bottom'\nCall Stack (most recent call first):\n cmake/project.cmake:249 (use_projects)\n lilliput/CMakeLists.txt:126 (i3_add_library)\n\n\n-- Configuring incomplete, errors occurred!\n}}}",
"reporter": "thomas.kittler",
"cc": "",
"resolution": "fixed",
"time": "2018-11-16T10:56:57",
"component": "other",
"summary": "lilliput attemtps to use nonexistent project rock_bottom",
"priority": "minor",
"keywords": "",
"milestone": "",
"owner": "kkrings",
"type": "defect"
}
```
</p>
</details>
| 1.0 | lilliput attemtps to use nonexistent project rock_bottom (Trac #2206) - I get this error when I run cmake (py3-v4, combo r166833):
```text
-- + lilliput
-- +-- python [symlinks]
-- + ipdf support
-- + Minuit2 support
-- - multinest support...not found
-- + NLopt support
CMake Error at cmake/project.cmake:94 (message):
Attempt to use nonexistent project 'rock_bottom'
Call Stack (most recent call first):
cmake/project.cmake:249 (use_projects)
lilliput/CMakeLists.txt:126 (i3_add_library)
-- Configuring incomplete, errors occurred!
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2206">https://code.icecube.wisc.edu/projects/icecube/ticket/2206</a>, reported by thomas.kittlerand owned by kkrings</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2018-11-16T15:25:14",
"_ts": "1542381914248311",
"description": "I get this error when I run cmake (py3-v4, combo r166833):\n{{{\n-- + lilliput\n-- +-- python [symlinks] \n-- + ipdf support\n-- + Minuit2 support\n-- - multinest support...not found\n-- + NLopt support\nCMake Error at cmake/project.cmake:94 (message):\n Attempt to use nonexistent project 'rock_bottom'\nCall Stack (most recent call first):\n cmake/project.cmake:249 (use_projects)\n lilliput/CMakeLists.txt:126 (i3_add_library)\n\n\n-- Configuring incomplete, errors occurred!\n}}}",
"reporter": "thomas.kittler",
"cc": "",
"resolution": "fixed",
"time": "2018-11-16T10:56:57",
"component": "other",
"summary": "lilliput attemtps to use nonexistent project rock_bottom",
"priority": "minor",
"keywords": "",
"milestone": "",
"owner": "kkrings",
"type": "defect"
}
```
</p>
</details>
| non_priority | lilliput attemtps to use nonexistent project rock bottom trac i get this error when i run cmake combo text lilliput python ipdf support support multinest support not found nlopt support cmake error at cmake project cmake message attempt to use nonexistent project rock bottom call stack most recent call first cmake project cmake use projects lilliput cmakelists txt add library configuring incomplete errors occurred migrated from json status closed changetime ts description i get this error when i run cmake combo n n lilliput n python n ipdf support n support n multinest support not found n nlopt support ncmake error at cmake project cmake message n attempt to use nonexistent project rock bottom ncall stack most recent call first n cmake project cmake use projects n lilliput cmakelists txt add library n n n configuring incomplete errors occurred n reporter thomas kittler cc resolution fixed time component other summary lilliput attemtps to use nonexistent project rock bottom priority minor keywords milestone owner kkrings type defect | 0 |
18,797 | 3,711,364,912 | IssuesEvent | 2016-03-02 10:01:54 | NishantUpadhyay-BTC/BLISS-Issue-Tracking | https://api.github.com/repos/NishantUpadhyay-BTC/BLISS-Issue-Tracking | closed | Check Availability display no lodging even there exist in Office UI | bug Deployed to Test | When I check lodgings for 19 July, for 1 person and 1 night, it should display lodgings but somehow not displaying any lodging. | 1.0 | Check Availability display no lodging even there exist in Office UI - When I check lodgings for 19 July, for 1 person and 1 night, it should display lodgings but somehow not displaying any lodging. | non_priority | check availability display no lodging even there exist in office ui when i check lodgings for july for person and night it should display lodgings but somehow not displaying any lodging | 0 |
331 | 5,436,478,860 | IssuesEvent | 2017-03-06 01:18:12 | jrz22/royalesuite | https://api.github.com/repos/jrz22/royalesuite | closed | Push code from GitHub to AWS S3 bucket | automation todo | Probably a githook.
- [x] After I push to master, code is pushed to bucket
- [x] After I push to any other branch and merge with master, code is pushed to bucket | 1.0 | Push code from GitHub to AWS S3 bucket - Probably a githook.
- [x] After I push to master, code is pushed to bucket
- [x] After I push to any other branch and merge with master, code is pushed to bucket | non_priority | push code from github to aws bucket probably a githook after i push to master code is pushed to bucket after i push to any other branch and merge with master code is pushed to bucket | 0 |
277,306 | 8,628,633,103 | IssuesEvent | 2018-11-21 17:58:03 | robot-lab/judyst-main-web-service | https://api.github.com/repos/robot-lab/judyst-main-web-service | opened | Адаптация текущей функциональности под решение задачи №2 | priority/high type/idea | # Idea proposal
## В чём заключается идея
Адаптировать последние наработки команды по анализу текста и внести соответствующие изменения в БД, модели и другие затрагиваемые модули.
## Возможное решение
Узнать у @mehwhatever0 подробности новых изменений. Он их обещал оформить в виде отчёта в PR. Часть данных по задаче уже отправлялась в конференцию в ВК.
## Дополнительный контекст или ссылки на связанные с данной задачей issues
За дополнительными подробностями обращаться к @mehwhatever0.
| 1.0 | Адаптация текущей функциональности под решение задачи №2 - # Idea proposal
## В чём заключается идея
Адаптировать последние наработки команды по анализу текста и внести соответствующие изменения в БД, модели и другие затрагиваемые модули.
## Возможное решение
Узнать у @mehwhatever0 подробности новых изменений. Он их обещал оформить в виде отчёта в PR. Часть данных по задаче уже отправлялась в конференцию в ВК.
## Дополнительный контекст или ссылки на связанные с данной задачей issues
За дополнительными подробностями обращаться к @mehwhatever0.
| priority | адаптация текущей функциональности под решение задачи № idea proposal в чём заключается идея адаптировать последние наработки команды по анализу текста и внести соответствующие изменения в бд модели и другие затрагиваемые модули возможное решение узнать у подробности новых изменений он их обещал оформить в виде отчёта в pr часть данных по задаче уже отправлялась в конференцию в вк дополнительный контекст или ссылки на связанные с данной задачей issues за дополнительными подробностями обращаться к | 1 |
168,972 | 6,392,862,733 | IssuesEvent | 2017-08-04 04:54:08 | samsung-cnct/k2 | https://api.github.com/repos/samsung-cnct/k2 | closed | fix helm charts to delete EBS volumes when services are removed | feature request K2 priority-p3 | Bringing down a cluster leaves behind ebs volumes with the aws provider. | 1.0 | fix helm charts to delete EBS volumes when services are removed - Bringing down a cluster leaves behind ebs volumes with the aws provider. | priority | fix helm charts to delete ebs volumes when services are removed bringing down a cluster leaves behind ebs volumes with the aws provider | 1 |
684,497 | 23,420,516,611 | IssuesEvent | 2022-08-13 16:16:53 | tusen-ai/naive-ui | https://api.github.com/repos/tusen-ai/naive-ui | closed | 分页器中每页条数的选择器的位置能否调整位置 | feature request priority: low | <!-- generated by issue-helper DO NOT REMOVE __BUG__ -->
### TuSimple/naive-ui version (版本)
2.32.2-beta.0
### Vue version (Vue 版本)
3.2.37
### Browser and its version (浏览器及其版本)
Chrome(103.0.5060.134)
### System and its version (系统及其版本)
window 11
### Node version (Node 版本)
### Reappearance link (重现链接)
https://www.naiveui.com/zh-CN/light/components/pagination
### Reappearance steps (重现步骤)
分页器中每页条数的选择器位置好像只能在分页器的固定位置展示,能否放到分页器中的任意位置
### Expected results (期望的结果)
期望能把分页器中每页条数的选择的位置放在分页前后缀中
### Actual results (实际的结果)
实际分页器中每页条数的选择器的位置只能按固定位置展示
### Remarks (补充说明)
<!-- generated by issue-helper DO NOT REMOVE __BUG__ --> | 1.0 | 分页器中每页条数的选择器的位置能否调整位置 - <!-- generated by issue-helper DO NOT REMOVE __BUG__ -->
### TuSimple/naive-ui version (版本)
2.32.2-beta.0
### Vue version (Vue 版本)
3.2.37
### Browser and its version (浏览器及其版本)
Chrome(103.0.5060.134)
### System and its version (系统及其版本)
window 11
### Node version (Node 版本)
### Reappearance link (重现链接)
https://www.naiveui.com/zh-CN/light/components/pagination
### Reappearance steps (重现步骤)
分页器中每页条数的选择器位置好像只能在分页器的固定位置展示,能否放到分页器中的任意位置
### Expected results (期望的结果)
期望能把分页器中每页条数的选择的位置放在分页前后缀中
### Actual results (实际的结果)
实际分页器中每页条数的选择器的位置只能按固定位置展示
### Remarks (补充说明)
<!-- generated by issue-helper DO NOT REMOVE __BUG__ --> | priority | 分页器中每页条数的选择器的位置能否调整位置 tusimple naive ui version 版本 beta vue version vue 版本 browser and its version 浏览器及其版本 chrome( ) system and its version 系统及其版本 window node version node 版本 reappearance link 重现链接 reappearance steps 重现步骤 分页器中每页条数的选择器位置好像只能在分页器的固定位置展示,能否放到分页器中的任意位置 expected results 期望的结果 期望能把分页器中每页条数的选择的位置放在分页前后缀中 actual results 实际的结果 实际分页器中每页条数的选择器的位置只能按固定位置展示 remarks 补充说明 | 1 |
763,556 | 26,762,145,115 | IssuesEvent | 2023-01-31 07:57:00 | kingwangzzang1234/repo-setup-sample | https://api.github.com/repos/kingwangzzang1234/repo-setup-sample | opened | This is what you have to do.. | For: CLI Priority: Critical Type: Feature Status: In Progress | 팀장: Create Org -> Create repository
1. README.md에 팀장은 템플릿을 구성하고, 팀원은 각자 자기소개 README.md에 업로드하여 PR하기
2. fizzbuzz, montyhall problem research -> fb.md, monty.md
## Requirements
- 각 개인은 2개 이상의 PR을 만들어야 함
- fork -> PR
- issue PR template
- label
- Create issue -> Work -> Open PR -> Code Review -> Merge PR
| 1.0 | This is what you have to do.. - 팀장: Create Org -> Create repository
1. README.md에 팀장은 템플릿을 구성하고, 팀원은 각자 자기소개 README.md에 업로드하여 PR하기
2. fizzbuzz, montyhall problem research -> fb.md, monty.md
## Requirements
- 각 개인은 2개 이상의 PR을 만들어야 함
- fork -> PR
- issue PR template
- label
- Create issue -> Work -> Open PR -> Code Review -> Merge PR
| priority | this is what you have to do 팀장 create org create repository readme md에 팀장은 템플릿을 구성하고 팀원은 각자 자기소개 readme md에 업로드하여 pr하기 fizzbuzz montyhall problem research fb md monty md requirements 각 개인은 이상의 pr을 만들어야 함 fork pr issue pr template label create issue work open pr code review merge pr | 1 |
538,085 | 15,762,174,234 | IssuesEvent | 2021-03-31 10:47:08 | AY2021S2-CS2113-T10-4/tp | https://api.github.com/repos/AY2021S2-CS2113-T10-4/tp | closed | As a user, I want to choose the canteens to look at | priority.High type.Enhancement | So that I can see which canteens are better | 1.0 | As a user, I want to choose the canteens to look at - So that I can see which canteens are better | priority | as a user i want to choose the canteens to look at so that i can see which canteens are better | 1 |
98,103 | 4,017,578,069 | IssuesEvent | 2016-05-16 05:26:50 | gama-platform/gama | https://api.github.com/repos/gama-platform/gama | opened | FIPA and Physics plugins models do not show up in the release | > Bug Concerns Modeling In Models Library OS All Priority High Version 1.7 beta | The models described in the maths plugin show up normally in `Plugin models`, but not the ones of the FIPA and physics plugins. It is probable that they cannot be found within the jar files. | 1.0 | FIPA and Physics plugins models do not show up in the release - The models described in the maths plugin show up normally in `Plugin models`, but not the ones of the FIPA and physics plugins. It is probable that they cannot be found within the jar files. | priority | fipa and physics plugins models do not show up in the release the models described in the maths plugin show up normally in plugin models but not the ones of the fipa and physics plugins it is probable that they cannot be found within the jar files | 1 |
35,782 | 6,497,589,270 | IssuesEvent | 2017-08-22 14:29:38 | arangodb/arangodb | https://api.github.com/repos/arangodb/arangodb | closed | There is a TypeError when using"Graph._create('GraphName'); " in arangosh. | 1 Question 3 Documentation | ## my environment running ArangoDB
I'm using the latest ArangoDB of the respective release series:
- [ ] 2.8
- [x] 3.0
- [ ] self-compiled devel branch
On this operating system:
- [ ] DCOS on
- [ ] AWS
- [ ] Azure
- [ ] own infrastructure
- [ ] Linux
- [ ] Debian .deb
- [ ] Ubuntu .deb
- [x] SUSE .rpm
- [ ] RedHat .rpm
- [ ] Fedora .rpm
- [ ] Gentoo
- [ ] docker - official docker library
- [ ] other:
- [ ] Windows, version:
- [ ] MacOS, version:
### My problem is:
Graph._create('knows-graph');
JavaScript exception in file '/usr/share/arangodb3/js/common/modules/@arangodb/general-graph.js' at 534,12: TypeError: Cannot read property 'forEach' of undefined
! sGEDs.forEach(
! ^
stacktrace: TypeError: Cannot read property 'forEach' of undefined
at /usr/share/arangodb3/js/common/modules/@arangodb/general-graph.js:534:12
at Array.forEach (native)
at Object._create (/usr/share/arangodb3/js/common/modules/@arangodb/general-graph.js:531:17)
at <shell command>:1:7
| 1.0 | There is a TypeError when using"Graph._create('GraphName'); " in arangosh. - ## my environment running ArangoDB
I'm using the latest ArangoDB of the respective release series:
- [ ] 2.8
- [x] 3.0
- [ ] self-compiled devel branch
On this operating system:
- [ ] DCOS on
- [ ] AWS
- [ ] Azure
- [ ] own infrastructure
- [ ] Linux
- [ ] Debian .deb
- [ ] Ubuntu .deb
- [x] SUSE .rpm
- [ ] RedHat .rpm
- [ ] Fedora .rpm
- [ ] Gentoo
- [ ] docker - official docker library
- [ ] other:
- [ ] Windows, version:
- [ ] MacOS, version:
### My problem is:
Graph._create('knows-graph');
JavaScript exception in file '/usr/share/arangodb3/js/common/modules/@arangodb/general-graph.js' at 534,12: TypeError: Cannot read property 'forEach' of undefined
! sGEDs.forEach(
! ^
stacktrace: TypeError: Cannot read property 'forEach' of undefined
at /usr/share/arangodb3/js/common/modules/@arangodb/general-graph.js:534:12
at Array.forEach (native)
at Object._create (/usr/share/arangodb3/js/common/modules/@arangodb/general-graph.js:531:17)
at <shell command>:1:7
| non_priority | there is a typeerror when using graph create graphname in arangosh my environment running arangodb i m using the latest arangodb of the respective release series self compiled devel branch on this operating system dcos on aws azure own infrastructure linux debian deb ubuntu deb suse rpm redhat rpm fedora rpm gentoo docker official docker library other windows version macos version my problem is graph create knows graph javascript exception in file usr share js common modules arangodb general graph js at typeerror cannot read property foreach of undefined sgeds foreach stacktrace typeerror cannot read property foreach of undefined at usr share js common modules arangodb general graph js at array foreach native at object create usr share js common modules arangodb general graph js at | 0 |
10,810 | 4,827,537,238 | IssuesEvent | 2016-11-07 13:56:47 | CartoDB/cartodb | https://api.github.com/repos/CartoDB/cartodb | closed | Points to polygon analysis always runs COUNT aggregation | bug Builder | In the point to polygon analysis, value aggregation always runs as _COUNT_, even if other method is specified.

| 1.0 | Points to polygon analysis always runs COUNT aggregation - In the point to polygon analysis, value aggregation always runs as _COUNT_, even if other method is specified.

| non_priority | points to polygon analysis always runs count aggregation in the point to polygon analysis value aggregation always runs as count even if other method is specified | 0 |
759,323 | 26,589,213,848 | IssuesEvent | 2023-01-23 06:39:53 | LuanRT/YouTube.js | https://api.github.com/repos/LuanRT/YouTube.js | closed | YouTube kids support | enhancement priority: medium | ### Describe your suggestion
Add support for YouTube kids.
YouTube kids uses InnerTube and since YouTube music is already supported I figured it wouldnt hurt to open a request.
This would be useful for searching for Kid Friendly Content as well as getting dedicated feeds for the content.
### Other details
YouTube Kids (workflow to get to content): https://www.youtubekids.com/
- Click `I'm a Parent`
- Click `Next`
- Enter 1989 (or any year, really. Maybe use current year - 30?)
- Unskippable 25 second video plays
- Click Next
- Click Skip (Or sign in)
- Click "I agree"
- Set child age ( < 4, 5-8 or 9-12)
- Enable/Disable search
- See home page
Note:
Some URLS can be written as YouTube urls.
Ex:
https://www.youtubekids.com/watch?v=LRKB6WzT9kc => https://www.youtube.com/watch?v=LRKB6WzT9kc
https://www.youtubekids.com/channel/UCMryAHWeLf3TxDyDxSWs4AQ => https://www.youtube.com/channel/UCMryAHWeLf3TxDyDxSWs4AQ
The video + channel pages do hold different data when viewing with YouTube Kids


Some videos dont have subscribe button and dont link back to the channel:

### Checklist
- [X] I am running the latest version.
- [X] I checked the documentation and found no answer.
- [X] I have searched the existing issues and made sure this is not a duplicate.
- [X] I have provided sufficient information. | 1.0 | YouTube kids support - ### Describe your suggestion
Add support for YouTube kids.
YouTube kids uses InnerTube and since YouTube music is already supported I figured it wouldnt hurt to open a request.
This would be useful for searching for Kid Friendly Content as well as getting dedicated feeds for the content.
### Other details
YouTube Kids (workflow to get to content): https://www.youtubekids.com/
- Click `I'm a Parent`
- Click `Next`
- Enter 1989 (or any year, really. Maybe use current year - 30?)
- Unskippable 25 second video plays
- Click Next
- Click Skip (Or sign in)
- Click "I agree"
- Set child age ( < 4, 5-8 or 9-12)
- Enable/Disable search
- See home page
Note:
Some URLS can be written as YouTube urls.
Ex:
https://www.youtubekids.com/watch?v=LRKB6WzT9kc => https://www.youtube.com/watch?v=LRKB6WzT9kc
https://www.youtubekids.com/channel/UCMryAHWeLf3TxDyDxSWs4AQ => https://www.youtube.com/channel/UCMryAHWeLf3TxDyDxSWs4AQ
The video + channel pages do hold different data when viewing with YouTube Kids


Some videos dont have subscribe button and dont link back to the channel:

### Checklist
- [X] I am running the latest version.
- [X] I checked the documentation and found no answer.
- [X] I have searched the existing issues and made sure this is not a duplicate.
- [X] I have provided sufficient information. | priority | youtube kids support describe your suggestion add support for youtube kids youtube kids uses innertube and since youtube music is already supported i figured it wouldnt hurt to open a request this would be useful for searching for kid friendly content as well as getting dedicated feeds for the content other details youtube kids workflow to get to content click i m a parent click next enter or any year really maybe use current year unskippable second video plays click next click skip or sign in click i agree set child age or enable disable search see home page note some urls can be written as youtube urls ex the video channel pages do hold different data when viewing with youtube kids some videos dont have subscribe button and dont link back to the channel checklist i am running the latest version i checked the documentation and found no answer i have searched the existing issues and made sure this is not a duplicate i have provided sufficient information | 1 |
142,879 | 5,478,285,818 | IssuesEvent | 2017-03-12 16:44:36 | vbmacher/emuStudio | https://api.github.com/repos/vbmacher/emuStudio | closed | Separate emulation control and GUI completely | feature high-priority | Emulation functionality should be implemented in a library, maybe emuLib or some other. The reason is that GUI - currently "main module", should be oriented only to visual features such as text editor, and debugger. The configuration file structure should be also the property of the GUI.
This separation should allow to implement more than 1 monopoly GUI, perhaps being able to integrate emuStudio with some IDE like NetBeans or IntelliJ IDEA. Also, there can be headless (command-line) wrappers of the library.
| 1.0 | Separate emulation control and GUI completely - Emulation functionality should be implemented in a library, maybe emuLib or some other. The reason is that GUI - currently "main module", should be oriented only to visual features such as text editor, and debugger. The configuration file structure should be also the property of the GUI.
This separation should allow to implement more than 1 monopoly GUI, perhaps being able to integrate emuStudio with some IDE like NetBeans or IntelliJ IDEA. Also, there can be headless (command-line) wrappers of the library.
| priority | separate emulation control and gui completely emulation functionality should be implemented in a library maybe emulib or some other the reason is that gui currently main module should be oriented only to visual features such as text editor and debugger the configuration file structure should be also the property of the gui this separation should allow to implement more than monopoly gui perhaps being able to integrate emustudio with some ide like netbeans or intellij idea also there can be headless command line wrappers of the library | 1 |
549,853 | 16,101,189,515 | IssuesEvent | 2021-04-27 09:29:54 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | pay.ebay.com - site is not usable | browser-firefox engine-gecko os-ios priority-critical | <!-- @browser: Firefox iOS 33.0 -->
<!-- @ua_header: Mozilla/5.0 (iPhone; CPU OS 14_4_2 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/33.0 Mobile/15E148 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/72026 -->
**URL**: https://pay.ebay.com/rxo?action=view
**Browser / Version**: Firefox iOS 33.0
**Operating System**: iOS 14.4.2
**Tested Another Browser**: Yes Safari
**Problem type**: Site is not usable
**Description**: Browser unsupported
**Steps to Reproduce**:
Browser unsupported, because it’s out of date, even after updating the app
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | pay.ebay.com - site is not usable - <!-- @browser: Firefox iOS 33.0 -->
<!-- @ua_header: Mozilla/5.0 (iPhone; CPU OS 14_4_2 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/33.0 Mobile/15E148 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/72026 -->
**URL**: https://pay.ebay.com/rxo?action=view
**Browser / Version**: Firefox iOS 33.0
**Operating System**: iOS 14.4.2
**Tested Another Browser**: Yes Safari
**Problem type**: Site is not usable
**Description**: Browser unsupported
**Steps to Reproduce**:
Browser unsupported, because it’s out of date, even after updating the app
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | pay ebay com site is not usable url browser version firefox ios operating system ios tested another browser yes safari problem type site is not usable description browser unsupported steps to reproduce browser unsupported because it’s out of date even after updating the app browser configuration none from with ❤️ | 1 |
8,900 | 7,477,916,510 | IssuesEvent | 2018-04-04 09:49:55 | acelabini/ibinex-option-one | https://api.github.com/repos/acelabini/ibinex-option-one | opened | Background color and disance from other folds | SECURITY | **Platform: Desktop**
Expected result:

Actual result:

| True | Background color and disance from other folds - **Platform: Desktop**
Expected result:

Actual result:

| non_priority | background color and disance from other folds platform desktop expected result actual result | 0 |
31,653 | 7,431,211,992 | IssuesEvent | 2018-03-25 12:33:06 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | [4.0] Request: On Collapsing Backend Mainmenu | J4 Issue No Code Attached Yet | I know that J!4 is still all in development but would like to suggest to consider to make the backend menu more useable that it is now.
Maybe something like this:
http://demos.getbootstrapkit.com/dynamic/portlets.html
Submenus expands down instead of sidebar offset. Especially when the mainmenu is collapsed. | 1.0 | [4.0] Request: On Collapsing Backend Mainmenu - I know that J!4 is still all in development but would like to suggest to consider to make the backend menu more useable that it is now.
Maybe something like this:
http://demos.getbootstrapkit.com/dynamic/portlets.html
Submenus expands down instead of sidebar offset. Especially when the mainmenu is collapsed. | non_priority | request on collapsing backend mainmenu i know that j is still all in development but would like to suggest to consider to make the backend menu more useable that it is now maybe something like this submenus expands down instead of sidebar offset especially when the mainmenu is collapsed | 0 |
20,844 | 6,114,243,063 | IssuesEvent | 2017-06-22 00:21:12 | ganeti/ganeti | https://api.github.com/repos/ganeti/ganeti | closed | http request body programmatically to perform POST /2/instances (i.e. create instance) | imported_from_google_code Status:WontFix | Originally reported of Google Code with ID 111.
```
<b>What steps will reproduce the problem?</b>
1.I have formed http-request in Java https://<primarynodename>:5080/2/instances, method POST
2.I have formed parameter list by folloing name=...,disk-template=... pnode=... and others
3.I have sent request programmatically. So, i already had success stories to perform get and post Ganeti API by JAVA.
<b>What is the expected output? What do you see instead?</b>
I see response:
"code": 400,
"explain": "Bad request syntax or unsupported method",
"message": "Invalid body contents, not a dictionary"
So, I understand, that I formed http-body wrong for that case. But documentation on this API has not tell anything about.
<b>Please use labels and text to provide additional information.</b>
```
Originally added on 2010-08-09 07:52:59 +0000 UTC. | 1.0 | http request body programmatically to perform POST /2/instances (i.e. create instance) - Originally reported of Google Code with ID 111.
```
<b>What steps will reproduce the problem?</b>
1.I have formed http-request in Java https://<primarynodename>:5080/2/instances, method POST
2.I have formed parameter list by folloing name=...,disk-template=... pnode=... and others
3.I have sent request programmatically. So, i already had success stories to perform get and post Ganeti API by JAVA.
<b>What is the expected output? What do you see instead?</b>
I see response:
"code": 400,
"explain": "Bad request syntax or unsupported method",
"message": "Invalid body contents, not a dictionary"
So, I understand, that I formed http-body wrong for that case. But documentation on this API has not tell anything about.
<b>Please use labels and text to provide additional information.</b>
```
Originally added on 2010-08-09 07:52:59 +0000 UTC. | non_priority | http request body programmatically to perform post instances i e create instance originally reported of google code with id what steps will reproduce the problem i have formed http request in java method post i have formed parameter list by folloing name disk template pnode and others i have sent request programmatically so i already had success stories to perform get and post ganeti api by java what is the expected output what do you see instead i see response code explain bad request syntax or unsupported method message invalid body contents not a dictionary so i understand that i formed http body wrong for that case but documentation on this api has not tell anything about please use labels and text to provide additional information originally added on utc | 0 |
28,560 | 4,422,239,852 | IssuesEvent | 2016-08-16 01:21:25 | directus/directus | https://api.github.com/repos/directus/directus | closed | MySQL Strict Mode should prevent installation pre-checks from passing | 4. Minor 7. Testing | I installed Directus on a fresh Ubuntu 16.04x64 box yesterday, on top of PHP 7 and the latest version of MySQL that has - as I learned after a long time - enabled strict mode by default, breaking most legacy queries.
It took me a long time to take the "**Strict mode enabled**" warning on the installation pre-check seriously, and considering the total failure of the installation because of it, I think it should be a check in itself, required enable the install button and to subsequently finish the installation.
Furthermore, I suggest renaming it to **MySQL strict mode** and providing links to the question. For me, [this solution](https://serverpilot.io/community/articles/how-to-disable-strict-mode-in-mysql-5-7.html) is the one that did the trick. | 1.0 | MySQL Strict Mode should prevent installation pre-checks from passing - I installed Directus on a fresh Ubuntu 16.04x64 box yesterday, on top of PHP 7 and the latest version of MySQL that has - as I learned after a long time - enabled strict mode by default, breaking most legacy queries.
It took me a long time to take the "**Strict mode enabled**" warning on the installation pre-check seriously, and considering the total failure of the installation because of it, I think it should be a check in itself, required enable the install button and to subsequently finish the installation.
Furthermore, I suggest renaming it to **MySQL strict mode** and providing links to the question. For me, [this solution](https://serverpilot.io/community/articles/how-to-disable-strict-mode-in-mysql-5-7.html) is the one that did the trick. | non_priority | mysql strict mode should prevent installation pre checks from passing i installed directus on a fresh ubuntu box yesterday on top of php and the latest version of mysql that has as i learned after a long time enabled strict mode by default breaking most legacy queries it took me a long time to take the strict mode enabled warning on the installation pre check seriously and considering the total failure of the installation because of it i think it should be a check in itself required enable the install button and to subsequently finish the installation furthermore i suggest renaming it to mysql strict mode and providing links to the question for me is the one that did the trick | 0 |
84,337 | 7,917,287,603 | IssuesEvent | 2018-07-04 09:22:30 | edenlabllc/ehealth.api | https://api.github.com/repos/edenlabllc/ehealth.api | closed | fix patient registration signup error code | BE epic/cabinet kind/task status/test | endpoint `/private/cabinet/registration`
according to [tests](https://edenlab.atlassian.net/wiki/spaces/EH/pages/506069150/Sign+Up+Step+2+tests) and actual FE implementation in case if user enters invalid OTP, FE expects 422 error. But BE responses with the 403 one.
It should be changed to 422:
[spec updated](https://uaehealthapi.docs.apiary.io/#reference/internal.-cabinet/cabinet/patient-registration)
[service logic](https://edenlab.atlassian.net/wiki/spaces/EH/pages/421429945/Sign+Up+2.3+MPI+User+Create+Update) | 1.0 | fix patient registration signup error code - endpoint `/private/cabinet/registration`
according to [tests](https://edenlab.atlassian.net/wiki/spaces/EH/pages/506069150/Sign+Up+Step+2+tests) and actual FE implementation in case if user enters invalid OTP, FE expects 422 error. But BE responses with the 403 one.
It should be changed to 422:
[spec updated](https://uaehealthapi.docs.apiary.io/#reference/internal.-cabinet/cabinet/patient-registration)
[service logic](https://edenlab.atlassian.net/wiki/spaces/EH/pages/421429945/Sign+Up+2.3+MPI+User+Create+Update) | non_priority | fix patient registration signup error code endpoint private cabinet registration according to and actual fe implementation in case if user enters invalid otp fe expects error but be responses with the one it should be changed to | 0 |
478,712 | 13,783,826,766 | IssuesEvent | 2020-10-08 19:51:39 | ansible/awx | https://api.github.com/repos/ansible/awx | closed | Inventory helper is misaligned on JT form | component:ui_next priority:medium qe:visual state:in_progress type:bug | ##### ISSUE TYPE
- Bug Report
##### SUMMARY
When making a new JT, if you are missing an inventory, the error is to the right of the element, rather than below it.

Make a new JT, and try to click save without filling in any fields to see this error. | 1.0 | Inventory helper is misaligned on JT form - ##### ISSUE TYPE
- Bug Report
##### SUMMARY
When making a new JT, if you are missing an inventory, the error is to the right of the element, rather than below it.

Make a new JT, and try to click save without filling in any fields to see this error. | priority | inventory helper is misaligned on jt form issue type bug report summary when making a new jt if you are missing an inventory the error is to the right of the element rather than below it make a new jt and try to click save without filling in any fields to see this error | 1 |
35,150 | 14,624,391,546 | IssuesEvent | 2020-12-23 06:13:24 | submariner-io/submariner | https://api.github.com/repos/submariner-io/submariner | closed | kubectl get serviceexport with "-o wide" does not show more info | enhancement service-discovery | ```
$ oc get serviceexport "nginx-cl-b"
NAME AGE
nginx-cl-b 9m7s
$ oc get serviceexport "nginx-cl-b" -o wide
NAME AGE
nginx-cl-b 9m13s
```
**What would you like to be added**:
Please show more info with the "-o wide" option enabled.
| 1.0 | kubectl get serviceexport with "-o wide" does not show more info - ```
$ oc get serviceexport "nginx-cl-b"
NAME AGE
nginx-cl-b 9m7s
$ oc get serviceexport "nginx-cl-b" -o wide
NAME AGE
nginx-cl-b 9m13s
```
**What would you like to be added**:
Please show more info with the "-o wide" option enabled.
| non_priority | kubectl get serviceexport with o wide does not show more info oc get serviceexport nginx cl b name age nginx cl b oc get serviceexport nginx cl b o wide name age nginx cl b what would you like to be added please show more info with the o wide option enabled | 0 |
595,164 | 18,060,734,359 | IssuesEvent | 2021-09-20 13:44:34 | Progressive335/Progressive335 | https://api.github.com/repos/Progressive335/Progressive335 | closed | Quest: Recover the Cargo! (Dustwallow Marsh) | quest medium priority | **Please fill in the bellow fields to create a meticulous report so that we can fix the issue properly.**
**ID of the Spell/NPC/Quest: https://www.wowhead.com/quest=27237/recover-the-cargo
__Place an X between the brackets to specify the ID__
[X] Quest
**Description of the bug: I can't loot Salvaged Strongbox when I "salvage" shipwreck debris.
**Current behaviour: Salvaged Strongbox can't be looted.
**Steps to reproduce the problem: Take the quest from Renn McGill. Open the given item "Renn's Supply" which gives you the "Salvage Kit" and the "Repaired Diving Gear". Go to one of the of the shipwreck debris and use the "Salvage Kit". Then go back up the surface to loot the floating debris.
**How it should work: I should be able to loot "Salvage Strongbox from floating debris once I salvaged shipwreck debris.
**Any other documentation that could be helpful:**
| 1.0 | Quest: Recover the Cargo! (Dustwallow Marsh) - **Please fill in the bellow fields to create a meticulous report so that we can fix the issue properly.**
**ID of the Spell/NPC/Quest: https://www.wowhead.com/quest=27237/recover-the-cargo
__Place an X between the brackets to specify the ID__
[X] Quest
**Description of the bug: I can't loot Salvaged Strongbox when I "salvage" shipwreck debris.
**Current behaviour: Salvaged Strongbox can't be looted.
**Steps to reproduce the problem: Take the quest from Renn McGill. Open the given item "Renn's Supply" which gives you the "Salvage Kit" and the "Repaired Diving Gear". Go to one of the of the shipwreck debris and use the "Salvage Kit". Then go back up the surface to loot the floating debris.
**How it should work: I should be able to loot "Salvage Strongbox from floating debris once I salvaged shipwreck debris.
**Any other documentation that could be helpful:**
| priority | quest recover the cargo dustwallow marsh please fill in the bellow fields to create a meticulous report so that we can fix the issue properly id of the spell npc quest place an x between the brackets to specify the id quest description of the bug i can t loot salvaged strongbox when i salvage shipwreck debris current behaviour salvaged strongbox can t be looted steps to reproduce the problem take the quest from renn mcgill open the given item renn s supply which gives you the salvage kit and the repaired diving gear go to one of the of the shipwreck debris and use the salvage kit then go back up the surface to loot the floating debris how it should work i should be able to loot salvage strongbox from floating debris once i salvaged shipwreck debris any other documentation that could be helpful | 1 |
131,992 | 10,727,149,965 | IssuesEvent | 2019-10-28 10:58:56 | ubtue/tuefind | https://api.github.com/repos/ubtue/tuefind | closed | Auswirkungen von Nachlässen (LR) auf "Neuzugänge" + Sortierung "neueste zuerst" | System: IxTheo System: RelBib ready for testing | Eine allgemeine Suche nach neuen Zugängen (egal in welchem Zeitraum) auf ptah liefert immer eine größere Treffermenge als eine ähnliche Suche im Produktivsystem.
Beispiele.
Ptah: https://ptah.ub.uni-tuebingen.de/Search/NewItem?range=7&submit=Suchen&botprotect=
Produktivsystem: https://www.ixtheo.de/Search/NewItem?range=7&submit=Suchen&botprotect=
Sollte die Treffermenge perfekt identisch sein? | 1.0 | Auswirkungen von Nachlässen (LR) auf "Neuzugänge" + Sortierung "neueste zuerst" - Eine allgemeine Suche nach neuen Zugängen (egal in welchem Zeitraum) auf ptah liefert immer eine größere Treffermenge als eine ähnliche Suche im Produktivsystem.
Beispiele.
Ptah: https://ptah.ub.uni-tuebingen.de/Search/NewItem?range=7&submit=Suchen&botprotect=
Produktivsystem: https://www.ixtheo.de/Search/NewItem?range=7&submit=Suchen&botprotect=
Sollte die Treffermenge perfekt identisch sein? | non_priority | auswirkungen von nachlässen lr auf neuzugänge sortierung neueste zuerst eine allgemeine suche nach neuen zugängen egal in welchem zeitraum auf ptah liefert immer eine größere treffermenge als eine ähnliche suche im produktivsystem beispiele ptah produktivsystem sollte die treffermenge perfekt identisch sein | 0 |
791,015 | 27,846,757,350 | IssuesEvent | 2023-03-20 15:57:17 | apcountryman/picolibrary-microchip-megaavr0 | https://api.github.com/repos/apcountryman/picolibrary-microchip-megaavr0 | closed | Add Microchip megaAVR 0-series BOD peripheral | priority-normal status-awaiting_review type-feature | Add Microchip megaAVR 0-series BOD peripheral (`::picolibrary::Microchip::megaAVR0::Peripheral::BOD`).
- [x] The `BOD` class should be defined in the `include/picolibrary/microchip/megaavr0/peripheral/bod.h`/`source/picolibrary/microchip/megaavr0/peripheral/bod.cc` header/source file pair
- [x] The following `BOD` class instances should be defined in the `include/picolibrary/microchip/megaavr0/peripheral.h`/`source/picolibrary/microchip/megaavr0/peripheral.cc` header/source file pair
- [x] `::picolibrary::Microchip::megaAVR0::Peripheral::BOD0`
- [x] Update documentation | 1.0 | Add Microchip megaAVR 0-series BOD peripheral - Add Microchip megaAVR 0-series BOD peripheral (`::picolibrary::Microchip::megaAVR0::Peripheral::BOD`).
- [x] The `BOD` class should be defined in the `include/picolibrary/microchip/megaavr0/peripheral/bod.h`/`source/picolibrary/microchip/megaavr0/peripheral/bod.cc` header/source file pair
- [x] The following `BOD` class instances should be defined in the `include/picolibrary/microchip/megaavr0/peripheral.h`/`source/picolibrary/microchip/megaavr0/peripheral.cc` header/source file pair
- [x] `::picolibrary::Microchip::megaAVR0::Peripheral::BOD0`
- [x] Update documentation | priority | add microchip megaavr series bod peripheral add microchip megaavr series bod peripheral picolibrary microchip peripheral bod the bod class should be defined in the include picolibrary microchip peripheral bod h source picolibrary microchip peripheral bod cc header source file pair the following bod class instances should be defined in the include picolibrary microchip peripheral h source picolibrary microchip peripheral cc header source file pair picolibrary microchip peripheral update documentation | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.