Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
7,449 | 6,962,952,022 | IssuesEvent | 2017-12-08 15:38:53 | hzi-braunschweig/SORMAS-Open | https://api.github.com/repos/hzi-braunschweig/SORMAS-Open | closed | Styles of soft-required, required and error fields are confusing [1+X] | 1023X Infrastructure sormas-ui | Suggestion for UI:
- [x] when commit fails, required fields should receive a red highlight (border)
- [x] and the popup shown should list all missing or invalid fields
- [x] soft-required fields should add custom style instead of using the vaadin error mechanism
- [x] style should give the field a primary blue border
- [x] same for required fields, when they are empty
- [x] fix error style not added to or removed from DateTimeField
Android:
- [x] make a suggestion
We should not change anything about the app behaviour in this regard before the re-design is finished, because using e.g. floating labels could mitigate a large part of this problem. | 1.0 | Styles of soft-required, required and error fields are confusing [1+X] - Suggestion for UI:
- [x] when commit fails, required fields should receive a red highlight (border)
- [x] and the popup shown should list all missing or invalid fields
- [x] soft-required fields should add custom style instead of using the vaadin error mechanism
- [x] style should give the field a primary blue border
- [x] same for required fields, when they are empty
- [x] fix error style not added to or removed from DateTimeField
Android:
- [x] make a suggestion
We should not change anything about the app behaviour in this regard before the re-design is finished, because using e.g. floating labels could mitigate a large part of this problem. | infrastructure | styles of soft required required and error fields are confusing suggestion for ui when commit fails required fields should receive a red highlight border and the popup shown should list all missing or invalid fields soft required fields should add custom style instead of using the vaadin error mechanism style should give the field a primary blue border same for required fields when they are empty fix error style not added to or removed from datetimefield android make a suggestion we should not change anything about the app behaviour in this regard before the re design is finished because using e g floating labels could mitigate a large part of this problem | 1 |
4,799 | 5,282,630,217 | IssuesEvent | 2017-02-07 19:20:33 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Cleanup TargetingPackReferences from dev/eng | area-Infrastructure dev-eng | It looks like we're still using "TargetingPackReference" even though reference seems to be all that's needed.
If I remove them I trigger the error that this was working around:
https://github.com/dotnet/buildtools/blob/7b3991e4bba0b86945a06470b313f9183e744e25/src/Microsoft.DotNet.Build.Tasks/PackageFiles/packageresolve.targets#L154
Let's use this issue to track cleaning everything up.
/cc @mellinoe @chcosta | 1.0 | Cleanup TargetingPackReferences from dev/eng - It looks like we're still using "TargetingPackReference" even though reference seems to be all that's needed.
If I remove them I trigger the error that this was working around:
https://github.com/dotnet/buildtools/blob/7b3991e4bba0b86945a06470b313f9183e744e25/src/Microsoft.DotNet.Build.Tasks/PackageFiles/packageresolve.targets#L154
Let's use this issue to track cleaning everything up.
/cc @mellinoe @chcosta | infrastructure | cleanup targetingpackreferences from dev eng it looks like we re still using targetingpackreference even though reference seems to be all that s needed if i remove them i trigger the error that this was working around let s use this issue to track cleaning everything up cc mellinoe chcosta | 1 |
27,645 | 22,063,029,530 | IssuesEvent | 2022-05-30 21:03:14 | Budibase/budibase | https://api.github.com/repos/Budibase/budibase | opened | Server/Worker Service Restructure | infrastructure backend | **COS**
- Make worker responsible for background jobs only
- Remove total dependency between server and worker
- Consolidate worker API code into app service
- send all /api requests directly to app server | 1.0 | Server/Worker Service Restructure - **COS**
- Make worker responsible for background jobs only
- Remove total dependency between server and worker
- Consolidate worker API code into app service
- send all /api requests directly to app server | infrastructure | server worker service restructure cos make worker responsible for background jobs only remove total dependency between server and worker consolidate worker api code into app service send all api requests directly to app server | 1 |
17,610 | 12,479,367,287 | IssuesEvent | 2020-05-29 18:08:55 | eventespresso/event-espresso-core | https://api.github.com/repos/eventespresso/event-espresso-core | closed | Prices not deleted permanently | category:models-and-data-infrastructure type:bug 🐞 | While working on #2899, I noticed that when prices are deleted permanently, there is some server-side error.
**Steps:**
1. Trash a Ticket that has at least one non-default price
2. In filter bar, set to view trashed tickets
3. Delete the above trashed ticket permanently
**Expected**: Ticket and the related non-default price should permanently be deleted.
**Actual**: There is a server-side error that says: `The price could not be deleted because of the following error(s): A valid price could not be retrieved from the database.` | 1.0 | Prices not deleted permanently - While working on #2899, I noticed that when prices are deleted permanently, there is some server-side error.
**Steps:**
1. Trash a Ticket that has at least one non-default price
2. In filter bar, set to view trashed tickets
3. Delete the above trashed ticket permanently
**Expected**: Ticket and the related non-default price should permanently be deleted.
**Actual**: There is a server-side error that says: `The price could not be deleted because of the following error(s): A valid price could not be retrieved from the database.` | infrastructure | prices not deleted permanently while working on i noticed that when prices are deleted permanently there is some server side error steps trash a ticket that has at least one non default price in filter bar set to view trashed tickets delete the above trashed ticket permanently expected ticket and the related non default price should permanently be deleted actual there is a server side error that says the price could not be deleted because of the following error s a valid price could not be retrieved from the database | 1 |
230,962 | 17,658,354,791 | IssuesEvent | 2021-08-21 02:11:40 | drakearch/kaggle-courses | https://api.github.com/repos/drakearch/kaggle-courses | closed | Microchallenges | documentation Learning | Ultra-short challenges to build and test your skill
- [x] Blackjack Microchallenge
- [x] Airline Price Optimization Micro-Challenge | 1.0 | Microchallenges - Ultra-short challenges to build and test your skill
- [x] Blackjack Microchallenge
- [x] Airline Price Optimization Micro-Challenge | non_infrastructure | microchallenges ultra short challenges to build and test your skill blackjack microchallenge airline price optimization micro challenge | 0 |
11,566 | 9,291,872,599 | IssuesEvent | 2019-03-22 00:18:24 | filecoin-project/go-filecoin | https://api.github.com/repos/filecoin-project/go-filecoin | closed | Instrument Metrics on ProcessBlock time and Number of ProcessBlock calls | A-infrastructure P1 | ### Description
Add instrumentation to process block. Metrics should include:
- Process Block duration
- Process Block call count
### Acceptance Criteria
- Process Block duration metric is accessible to infrastructure via prometheus
- Process Block count metric is accessible to infrastructure via prometheus | 1.0 | Instrument Metrics on ProcessBlock time and Number of ProcessBlock calls - ### Description
Add instrumentation to process block. Metrics should include:
- Process Block duration
- Process Block call count
### Acceptance Criteria
- Process Block duration metric is accessible to infrastructure via prometheus
- Process Block count metric is accessible to infrastructure via prometheus | infrastructure | instrument metrics on processblock time and number of processblock calls description add instrumentation to process block metrics should include process block duration process block call count acceptance criteria process block duration metric is accessible to infrastructure via prometheus process block count metric is accessible to infrastructure via prometheus | 1 |
26,401 | 20,074,141,575 | IssuesEvent | 2022-02-04 10:43:35 | google/site-kit-wp | https://api.github.com/repos/google/site-kit-wp | opened | E2E is broken validating AMP for visitors | P0 Type: Infrastructure | ## Bug Description
E2E started all failing together just recently due to invalid AMP for non-logged in users.
```
462 | it( 'validates homepage AMP for non-logged-in users', async () => {
> 463 | await expect( page ).toHaveValidAMPForVisitor();
| ^
464 | } );
465 | } );
466 | } );
```
After taking a closer look at the `ValidationResult` itself, the following error was shown:
```
The native JavaScript AMPHTML Validator (validator.js) has been turned down. If you are seeing this error, update your tooling to instead load the API compatible WebAssembly AMPHTML Validator (validator_wasm.js) instead.
```
It turns out our `amphtml-validator` package is out-of-date and just needs updating 😄
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* <!-- One or more bullet points for acceptance criteria. -->
## Implementation Brief
* <!-- One or more bullet points for how to technically resolve the issue. Make sure to include changes to Storybook and visual regression tests where relevant. -->
### Test Coverage
* <!-- One or more bullet points for how to implement automated tests to verify the issue is resolved. -->
## QA Brief
* <!-- One or more bullet points for how to test that the issue has been resolved. -->
## Changelog entry
* <!-- One sentence summarizing the PR, to be used in the changelog. -->
| 1.0 | E2E is broken validating AMP for visitors - ## Bug Description
E2E started all failing together just recently due to invalid AMP for non-logged in users.
```
462 | it( 'validates homepage AMP for non-logged-in users', async () => {
> 463 | await expect( page ).toHaveValidAMPForVisitor();
| ^
464 | } );
465 | } );
466 | } );
```
After taking a closer look at the `ValidationResult` itself, the following error was shown:
```
The native JavaScript AMPHTML Validator (validator.js) has been turned down. If you are seeing this error, update your tooling to instead load the API compatible WebAssembly AMPHTML Validator (validator_wasm.js) instead.
```
It turns out our `amphtml-validator` package is out-of-date and just needs updating 😄
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* <!-- One or more bullet points for acceptance criteria. -->
## Implementation Brief
* <!-- One or more bullet points for how to technically resolve the issue. Make sure to include changes to Storybook and visual regression tests where relevant. -->
### Test Coverage
* <!-- One or more bullet points for how to implement automated tests to verify the issue is resolved. -->
## QA Brief
* <!-- One or more bullet points for how to test that the issue has been resolved. -->
## Changelog entry
* <!-- One sentence summarizing the PR, to be used in the changelog. -->
| infrastructure | is broken validating amp for visitors bug description started all failing together just recently due to invalid amp for non logged in users it validates homepage amp for non logged in users async await expect page tohavevalidampforvisitor after taking a closer look at the validationresult itself the following error was shown the native javascript amphtml validator validator js has been turned down if you are seeing this error update your tooling to instead load the api compatible webassembly amphtml validator validator wasm js instead it turns out our amphtml validator package is out of date and just needs updating 😄 do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria implementation brief test coverage qa brief changelog entry | 1 |
11,102 | 8,925,781,109 | IssuesEvent | 2019-01-22 00:41:17 | glen3b/CyPatScoreboardBot | https://api.github.com/repos/glen3b/CyPatScoreboardBot | closed | Refactor score retrieval services | enhancement infrastructure | Refactor the code such that a separate project could be created independent of Discord will full-featured score deserialization and retrieval. | 1.0 | Refactor score retrieval services - Refactor the code such that a separate project could be created independent of Discord will full-featured score deserialization and retrieval. | infrastructure | refactor score retrieval services refactor the code such that a separate project could be created independent of discord will full featured score deserialization and retrieval | 1 |
4,852 | 5,300,464,148 | IssuesEvent | 2017-02-10 05:03:51 | crimethinc/website | https://api.github.com/repos/crimethinc/website | closed | Add robots.txt | infrastructure | Make sure Google et al is doing The Right Thing™ with article pages vs pagination and search results
| 1.0 | Add robots.txt - Make sure Google et al is doing The Right Thing™ with article pages vs pagination and search results
| infrastructure | add robots txt make sure google et al is doing the right thing™ with article pages vs pagination and search results | 1 |
35,153 | 30,795,641,700 | IssuesEvent | 2023-07-31 19:39:03 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | Use of proprietary license in AspNetCoreModule-Setup | enhancement area-infrastructure | The license currently present in `AspNetCoreModule-Setup` restricts redistribution of covered software ([source](https://github.com/dotnet/aspnetcore/blob/main/src/Installers/Windows/AspNetCoreModule-Setup/license/license.rtf#L55)):
```
You may not
...
- publish the software for others to copy
```
This is problematic in VMR context as this package is included there with `aspnetcore`, making the license part of the VMR. Proprietary license itself is a show-stopper for various .NET source-build partners.
From the `git` history it seems that the license was part of the once separate project that was later on merged into `aspnetcore`. Would it be possible to alter / remove the license so the project can be used in the VMR?
Thank you | 1.0 | Use of proprietary license in AspNetCoreModule-Setup - The license currently present in `AspNetCoreModule-Setup` restricts redistribution of covered software ([source](https://github.com/dotnet/aspnetcore/blob/main/src/Installers/Windows/AspNetCoreModule-Setup/license/license.rtf#L55)):
```
You may not
...
- publish the software for others to copy
```
This is problematic in VMR context as this package is included there with `aspnetcore`, making the license part of the VMR. Proprietary license itself is a show-stopper for various .NET source-build partners.
From the `git` history it seems that the license was part of the once separate project that was later on merged into `aspnetcore`. Would it be possible to alter / remove the license so the project can be used in the VMR?
Thank you | infrastructure | use of proprietary license in aspnetcoremodule setup the license currently present in aspnetcoremodule setup restricts redistribution of covered software you may not publish the software for others to copy this is problematic in vmr context as this package is included there with aspnetcore making the license part of the vmr proprietary license itself is a show stopper for various net source build partners from the git history it seems that the license was part of the once separate project that was later on merged into aspnetcore would it be possible to alter remove the license so the project can be used in the vmr thank you | 1 |
155,287 | 13,617,653,635 | IssuesEvent | 2020-09-23 17:19:37 | netlify/build | https://api.github.com/repos/netlify/build | closed | Clarify where settings are loaded from | documentation type: feature | If you set a `base` in `netlify.toml`, you need to create a second `netlify.toml` in that base directory to get things like redirects to work
This is fine, but _super_ unintuitive and [not documented](https://docs.netlify.com/configure-builds/file-based-configuration/).
I think we can rectify this with relatively low effort:
1. Log where settings are loaded from:
```
[config] a `base` has been set. File-based settings will be sourced from
/path/to/base/netlify.toml -- see https://docs.link for more information
```
2. Update our documentation for `base` to explain that it introduces the need for multiple `netlify.toml` files | 1.0 | Clarify where settings are loaded from - If you set a `base` in `netlify.toml`, you need to create a second `netlify.toml` in that base directory to get things like redirects to work
This is fine, but _super_ unintuitive and [not documented](https://docs.netlify.com/configure-builds/file-based-configuration/).
I think we can rectify this with relatively low effort:
1. Log where settings are loaded from:
```
[config] a `base` has been set. File-based settings will be sourced from
/path/to/base/netlify.toml -- see https://docs.link for more information
```
2. Update our documentation for `base` to explain that it introduces the need for multiple `netlify.toml` files | non_infrastructure | clarify where settings are loaded from if you set a base in netlify toml you need to create a second netlify toml in that base directory to get things like redirects to work this is fine but super unintuitive and i think we can rectify this with relatively low effort log where settings are loaded from a base has been set file based settings will be sourced from path to base netlify toml see for more information update our documentation for base to explain that it introduces the need for multiple netlify toml files | 0 |
491,591 | 14,166,727,746 | IssuesEvent | 2020-11-12 09:19:27 | googleapis/java-core | https://api.github.com/repos/googleapis/java-core | closed | com.google.cloud.Identity does not support the new IAM deleted user syntax | priority: p2 status: investigating type: bug | The new IAM deleted user syntax is not supported by com.google.cloud.Identity:
https://cloud.google.com/iam/docs/policies#handle-deleted-members
#### Environment details
1. Affects the Core library and all dependant libraries
2. OS type and version: any
3. Java version: any
4. google-cloud-core version(s): any
#### Steps to reproduce
1. Try to parse a deleted user string with Identity.valueOf
2. This results either to a stacktrace (1.74.0 and lower) or a DELETED identity type (1.75.0 and higher)
#### Code example
```java
Identity identity = Identity.valueOf("deleted:user:donald@example.com?uid=234567890123456789012");
System.out.println("type = " + identity.getType());
System.out.println("value = " + identity.getValue());
```
#### Stack trace
1.74.0 and lower
```
java.lang.IllegalArgumentException: Illegal identity string: "deleted:user:donald@example.com?uid=234567890123456789012"
```
1.75.0 and higher
```
type = DELETED
value = user:donald@example.com?uid=234567890123456789012
```
The "fix" is caused by an unrelated change: https://github.com/googleapis/java-core/commit/24667e5c27ab71bd217d6667c1cb0affefe09b13
#### External references such as API reference guides
https://cloud.google.com/iam/docs/policies#handle-deleted-members
| 1.0 | com.google.cloud.Identity does not support the new IAM deleted user syntax - The new IAM deleted user syntax is not supported by com.google.cloud.Identity:
https://cloud.google.com/iam/docs/policies#handle-deleted-members
#### Environment details
1. Affects the Core library and all dependant libraries
2. OS type and version: any
3. Java version: any
4. google-cloud-core version(s): any
#### Steps to reproduce
1. Try to parse a deleted user string with Identity.valueOf
2. This results either to a stacktrace (1.74.0 and lower) or a DELETED identity type (1.75.0 and higher)
#### Code example
```java
Identity identity = Identity.valueOf("deleted:user:donald@example.com?uid=234567890123456789012");
System.out.println("type = " + identity.getType());
System.out.println("value = " + identity.getValue());
```
#### Stack trace
1.74.0 and lower
```
java.lang.IllegalArgumentException: Illegal identity string: "deleted:user:donald@example.com?uid=234567890123456789012"
```
1.75.0 and higher
```
type = DELETED
value = user:donald@example.com?uid=234567890123456789012
```
The "fix" is caused by an unrelated change: https://github.com/googleapis/java-core/commit/24667e5c27ab71bd217d6667c1cb0affefe09b13
#### External references such as API reference guides
https://cloud.google.com/iam/docs/policies#handle-deleted-members
| non_infrastructure | com google cloud identity does not support the new iam deleted user syntax the new iam deleted user syntax is not supported by com google cloud identity environment details affects the core library and all dependant libraries os type and version any java version any google cloud core version s any steps to reproduce try to parse a deleted user string with identity valueof this results either to a stacktrace and lower or a deleted identity type and higher code example java identity identity identity valueof deleted user donald example com uid system out println type identity gettype system out println value identity getvalue stack trace and lower java lang illegalargumentexception illegal identity string deleted user donald example com uid and higher type deleted value user donald example com uid the fix is caused by an unrelated change external references such as api reference guides | 0 |
148,627 | 19,534,417,010 | IssuesEvent | 2021-12-31 01:37:42 | panasalap/linux-4.1.15 | https://api.github.com/repos/panasalap/linux-4.1.15 | opened | CVE-2018-18710 (Medium) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2018-18710 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/cdrom/cdrom.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/cdrom/cdrom.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel through 4.19. An information leak in cdrom_ioctl_select_disc in drivers/cdrom/cdrom.c could be used by local attackers to read kernel memory because a cast from unsigned long to int interferes with bounds checking. This is similar to CVE-2018-10940 and CVE-2018-16658.
<p>Publish Date: 2018-10-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-18710>CVE-2018-18710</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/torvalds/linux/commit/e4f3aa2e1e67bb48dfbaaf1cad59013d5a5bc276">https://github.com/torvalds/linux/commit/e4f3aa2e1e67bb48dfbaaf1cad59013d5a5bc276</a></p>
<p>Release Date: 2018-10-03</p>
<p>Fix Resolution: Replace or update the following file: cdrom.c</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-18710 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2018-18710 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/cdrom/cdrom.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/cdrom/cdrom.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel through 4.19. An information leak in cdrom_ioctl_select_disc in drivers/cdrom/cdrom.c could be used by local attackers to read kernel memory because a cast from unsigned long to int interferes with bounds checking. This is similar to CVE-2018-10940 and CVE-2018-16658.
<p>Publish Date: 2018-10-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-18710>CVE-2018-18710</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/torvalds/linux/commit/e4f3aa2e1e67bb48dfbaaf1cad59013d5a5bc276">https://github.com/torvalds/linux/commit/e4f3aa2e1e67bb48dfbaaf1cad59013d5a5bc276</a></p>
<p>Release Date: 2018-10-03</p>
<p>Fix Resolution: Replace or update the following file: cdrom.c</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in base branch master vulnerable source files drivers cdrom cdrom c drivers cdrom cdrom c vulnerability details an issue was discovered in the linux kernel through an information leak in cdrom ioctl select disc in drivers cdrom cdrom c could be used by local attackers to read kernel memory because a cast from unsigned long to int interferes with bounds checking this is similar to cve and cve publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following file cdrom c step up your open source security game with whitesource | 0 |
1,275 | 3,114,861,446 | IssuesEvent | 2015-09-03 11:28:03 | dart-lang/fletch | https://api.github.com/repos/dart-lang/fletch | closed | corelib/double_parse_test temporarily duplicated | Infrastructure | MacOS's strtod is less precise than the linux version. That causes Fletch to fail two corner case double conversion tests on Mac. That is OK and we want to ignore those failures. For now we have copied corelib/double_parse_test to unsorted/temp_double_parse_test. We are landing a change in the Dart SDK repo that treats these corner cases as a separate test. Once we update the Dart SDK version with that change we should mark only corelib/double_parse_test/02 as failing on mac and remove our local copy. | 1.0 | corelib/double_parse_test temporarily duplicated - MacOS's strtod is less precise than the linux version. That causes Fletch to fail two corner case double conversion tests on Mac. That is OK and we want to ignore those failures. For now we have copied corelib/double_parse_test to unsorted/temp_double_parse_test. We are landing a change in the Dart SDK repo that treats these corner cases as a separate test. Once we update the Dart SDK version with that change we should mark only corelib/double_parse_test/02 as failing on mac and remove our local copy. | infrastructure | corelib double parse test temporarily duplicated macos s strtod is less precise than the linux version that causes fletch to fail two corner case double conversion tests on mac that is ok and we want to ignore those failures for now we have copied corelib double parse test to unsorted temp double parse test we are landing a change in the dart sdk repo that treats these corner cases as a separate test once we update the dart sdk version with that change we should mark only corelib double parse test as failing on mac and remove our local copy | 1 |
259,459 | 8,198,539,364 | IssuesEvent | 2018-08-31 16:45:35 | javaee/jpa-spec | https://api.github.com/repos/javaee/jpa-spec | closed | @Index.columnList should be an array | Priority: Major Type: Improvement | In the JPA 2.1 specification, @Index has a String columnList property, and the specification presents this as a comma-delimited list. I propose that the specification be modified to make this property of type String[], both as a less surprising syntax and as an easier leap from Hibernate syntax. | 1.0 | @Index.columnList should be an array - In the JPA 2.1 specification, @Index has a String columnList property, and the specification presents this as a comma-delimited list. I propose that the specification be modified to make this property of type String[], both as a less surprising syntax and as an easier leap from Hibernate syntax. | non_infrastructure | index columnlist should be an array in the jpa specification index has a string columnlist property and the specification presents this as a comma delimited list i propose that the specification be modified to make this property of type string both as a less surprising syntax and as an easier leap from hibernate syntax | 0 |
381,274 | 11,275,899,016 | IssuesEvent | 2020-01-14 21:49:35 | googleapis/java-asset | https://api.github.com/repos/googleapis/java-asset | closed | Synthesis failed for java-asset | autosynth failure priority: p1 type: bug | Hello! Autosynth couldn't regenerate java-asset. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to branch 'autosynth'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main
last_synth_commit_hash = get_last_metadata_commit(args.metadata_path)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit
text=True,
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run
with Popen(*popenargs, **kwargs) as process:
TypeError: __init__() got an unexpected keyword argument 'text'
```
Google internal developers can see the full log [here](https://sponge/482a7766-2134-4c8a-887d-f6441089f633).
| 1.0 | Synthesis failed for java-asset - Hello! Autosynth couldn't regenerate java-asset. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to branch 'autosynth'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main
last_synth_commit_hash = get_last_metadata_commit(args.metadata_path)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit
text=True,
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run
with Popen(*popenargs, **kwargs) as process:
TypeError: __init__() got an unexpected keyword argument 'text'
```
Google internal developers can see the full log [here](https://sponge/482a7766-2134-4c8a-887d-f6441089f633).
| non_infrastructure | synthesis failed for java asset hello autosynth couldn t regenerate java asset broken heart here s the output from running synth py cloning into working repo switched to branch autosynth traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main last synth commit hash get last metadata commit args metadata path file tmpfs src git autosynth autosynth synth py line in get last metadata commit text true file home kbuilder pyenv versions lib subprocess py line in run with popen popenargs kwargs as process typeerror init got an unexpected keyword argument text google internal developers can see the full log | 0 |
16,574 | 12,056,854,526 | IssuesEvent | 2020-04-15 15:01:31 | patternfly/patternfly | https://api.github.com/repos/patternfly/patternfly | closed | patternfly README file is out of date for dev env. | bug infrastructure | 'dist' directory not being built when following the steps in readme file.
PatternFly 4 Development requires Node v8.0.0 or greater
To setup the PatternFly 4 development environment:
clone the project
run npm install from the project root
run npm run cli:setup (only needed if doing development)
run npm run dev
Last step shows the failure.
> npm run dev
> @patternfly/patternfly@0.0.0-development dev /Users/dougdonahue/WebstormProjects/patternfly-next
> concurrently "npm run copy-assets" "gulp watchSASS" "gatsby develop"
[0]
[0] > @patternfly/patternfly@0.0.0-development copy-assets /Users/dougdonahue/WebstormProjects/patternfly-next
[0] > gulp copyAssets
[0]
[1] [09:57:30] [09:57:30] Using gulpfile ~/WebstormProjects/patternfly-next/gulpfile.js
[1] Using gulpfile ~/WebstormProjects/patternfly-next/gulpfile.js
[1] [09:57:30] Starting 'watchSASS'...
[0] [09:57:30] Starting 'copyAssets'...
[0] [09:57:30] Finished 'copyAssets' after 452 ms
[0] npm run copy-assets exited with code 0
[2] success open and validate gatsby-configs - 0.087s
[2] success load plugins - 1.399s
[2] success onPreInit - 0.001s
[2] success initialize cache - 0.007s
[2] success copy gatsby files - 0.068s
[2] success onPreBootstrap - 0.015s
[2] success createSchemaCustomization - 0.004s
[2] success source and transform nodes - 0.989s
[2] success building schema - 0.419s
[2] success createPages - 7.855s
[2] success createPagesStatefully - 0.178s
[2] success onPreExtractQueries - 0.001s
[2] success update schema - 1.136s
[2] success extract queries from components - 0.443s
[2] success write out requires - 0.058s
[2] success write out redirect data - 0.008s
[2] success Build manifest and related icons - 0.330s
[2] success onPostBootstrap - 0.360s
[2] ⠀
[2] info bootstrap finished - 18.274 s
[2] ⠀
success run queries - 0.468s - 410/410 876.62/s
develop-html
develop
[2] ERROR #98123 WEBPACK
[2]
[2] Generating development JavaScript bundle failed
[2]
[2] Can't resolve './dist/patternfly-addons.css' in '/Users/dougdonahue/WebstormProjects/patternfly-next'
[2]
[2] File: gatsby-browser.js
[2]
[2] ERROR #98123 WEBPACK
[2]
[2] Generating development JavaScript bundle failed
[2]
[2] Can't resolve './dist/patternfly.css' in '/Users/dougdonahue/WebstormProjects/patternfly-next'
[2]
[2] File: gatsby-browser.js
[2]
failed Building development bundle - 8.200s
[2]
[2] 501 pages Failed @p
[2] atternfly/patternfly
[1] gulp watchSASS exited with code SIGINT
[2] gatsby develop exited with code SIGINT
| 1.0 | patternfly README file is out of date for dev env. - 'dist' directory not being built when following the steps in readme file.
PatternFly 4 Development requires Node v8.0.0 or greater
To setup the PatternFly 4 development environment:
clone the project
run npm install from the project root
run npm run cli:setup (only needed if doing development)
run npm run dev
Last step shows the failure.
> npm run dev
> @patternfly/patternfly@0.0.0-development dev /Users/dougdonahue/WebstormProjects/patternfly-next
> concurrently "npm run copy-assets" "gulp watchSASS" "gatsby develop"
[0]
[0] > @patternfly/patternfly@0.0.0-development copy-assets /Users/dougdonahue/WebstormProjects/patternfly-next
[0] > gulp copyAssets
[0]
[1] [09:57:30] [09:57:30] Using gulpfile ~/WebstormProjects/patternfly-next/gulpfile.js
[1] Using gulpfile ~/WebstormProjects/patternfly-next/gulpfile.js
[1] [09:57:30] Starting 'watchSASS'...
[0] [09:57:30] Starting 'copyAssets'...
[0] [09:57:30] Finished 'copyAssets' after 452 ms
[0] npm run copy-assets exited with code 0
[2] success open and validate gatsby-configs - 0.087s
[2] success load plugins - 1.399s
[2] success onPreInit - 0.001s
[2] success initialize cache - 0.007s
[2] success copy gatsby files - 0.068s
[2] success onPreBootstrap - 0.015s
[2] success createSchemaCustomization - 0.004s
[2] success source and transform nodes - 0.989s
[2] success building schema - 0.419s
[2] success createPages - 7.855s
[2] success createPagesStatefully - 0.178s
[2] success onPreExtractQueries - 0.001s
[2] success update schema - 1.136s
[2] success extract queries from components - 0.443s
[2] success write out requires - 0.058s
[2] success write out redirect data - 0.008s
[2] success Build manifest and related icons - 0.330s
[2] success onPostBootstrap - 0.360s
[2] ⠀
[2] info bootstrap finished - 18.274 s
[2] ⠀
success run queries - 0.468s - 410/410 876.62/s
develop-html
develop
[2] ERROR #98123 WEBPACK
[2]
[2] Generating development JavaScript bundle failed
[2]
[2] Can't resolve './dist/patternfly-addons.css' in '/Users/dougdonahue/WebstormProjects/patternfly-next'
[2]
[2] File: gatsby-browser.js
[2]
[2] ERROR #98123 WEBPACK
[2]
[2] Generating development JavaScript bundle failed
[2]
[2] Can't resolve './dist/patternfly.css' in '/Users/dougdonahue/WebstormProjects/patternfly-next'
[2]
[2] File: gatsby-browser.js
[2]
failed Building development bundle - 8.200s
[2]
[2] 501 pages Failed @p
[2] atternfly/patternfly
[1] gulp watchSASS exited with code SIGINT
[2] gatsby develop exited with code SIGINT
| infrastructure | patternfly readme file is out of date for dev env dist directory not being built when following the steps in readme file patternfly development requires node or greater to setup the patternfly development environment clone the project run npm install from the project root run npm run cli setup only needed if doing development run npm run dev last step shows the failure npm run dev patternfly patternfly development dev users dougdonahue webstormprojects patternfly next concurrently npm run copy assets gulp watchsass gatsby develop patternfly patternfly development copy assets users dougdonahue webstormprojects patternfly next gulp copyassets using gulpfile webstormprojects patternfly next gulpfile js using gulpfile webstormprojects patternfly next gulpfile js starting watchsass starting copyassets finished copyassets after ms npm run copy assets exited with code success open and validate gatsby configs success load plugins success onpreinit success initialize cache success copy gatsby files success onprebootstrap success createschemacustomization success source and transform nodes success building schema success createpages success createpagesstatefully success onpreextractqueries success update schema success extract queries from components success write out requires success write out redirect data success build manifest and related icons success onpostbootstrap ⠀ info bootstrap finished s ⠀ success run queries s develop html develop error webpack generating development javascript bundle failed can t resolve dist patternfly addons css in users dougdonahue webstormprojects patternfly next file gatsby browser js error webpack generating development javascript bundle failed can t resolve dist patternfly css in users dougdonahue webstormprojects patternfly next file gatsby browser js failed building development bundle pages failed p atternfly patternfly gulp watchsass exited with code sigint gatsby develop exited with code sigint | 1 |
180,358 | 13,930,135,427 | IssuesEvent | 2020-10-22 01:36:24 | OpenMined/PySyft | https://api.github.com/repos/OpenMined/PySyft | closed | Add torch.Tensor.floor_ to allowlist and test suite | Priority: 2 - High :cold_sweat: Severity: 3 - Medium :unamused: Status: Available :wave: Type: New Feature :heavy_plus_sign: Type: Testing :test_tube: |
# Description
This issue is a part of Syft 0.3.0 Epic 2: https://github.com/OpenMined/PySyft/issues/3696
In this issue, you will be adding support for remote execution of the torch.Tensor.floor_
method or property. This might be a really small project (literally a one-liner) or
it might require adding significant functionality to PySyft OR to the testing suite
in order to make sure the feature is both functional and tested.
## Step 0: Run tests and ./scripts/pre_commit.sh
Before you get started with this project, let's make sure you have everything building and testing
correctly. Clone the codebase and run:
```pip uninstall syft```
followed by
```pip install -e .```
Then run the pre-commit file (which will also run the tests)
```./scripts/pre_commit.sh```
If all of these tests pass, continue on. If not, make sure you have all the
dependencies in requirements.txt installed, etc.
## Step 1: Uncomment your method in the allowlist.py file
Inside [allowlist.py](https://github.com/OpenMined/PySyft/blob/syft_0.3.0/src/syft/lib/torch/allowlist.py) you will find a huge dictionary of methods. Find your method and uncomment the line its on. At the time
of writing this Issue (WARNING: THIS MAY HAVE CHANGED) the dictionary maps from the
string name of the method (in your case 'torch.Tensor.floor_') to the string representation
of the type the method returns.
## Step 2: Run Unit Tests
Run the following:
```python setup.py test```
And wait to see if some of the tests fail. Why might the tests fail now? I'm so glad you asked!
https://github.com/OpenMined/PySyft/blob/syft_0.3.0/tests/syft/lib/torch/tensor/tensor_remote_method_api_suite_test.py
In this file you'll find the torch method test suite. It AUTOMATICALLY loads all methods
from the allowlist.py file you modified in the previous step. It attempts to test them.
# Step 3: If you get a Failing Test
If you get a failing test, this could be for one of a few reasons:
### Reason 1 - The testing suite passed in non-compatible arguments
The testing suite is pretty dumb. It literally just has a permutation of possible
arguments to pass into every method on torch tensors. So, if one of those permutations
doesn't work for your method (aka... perhaps it tries to call your method without
any arguments but torch.Tensor.floor_ actually requires some) then the test will
fail if the error hasn't been seen before.
If this happens - don't worry! Just look inside the only test in that file and look
for the huge lists of error types to ignore. Add your error to the list and keep
going!!!
*WARNING:* make sure that the testing suite actually tests your method via remote
execution once you've gotten all the tests passing. Aka - if the testing suite
doesn't have ANY matching argument permutations for your method, then you're
literally creating a bunch of unit tests that do absolutely nothing. If this is the
case, then ADD MORE ARGUMENT TYPES TO THE TESTING SUITE so that your argument
gets run via remote execution. DO NOT CLOSE THIS ISSUE until you can verify that
torch.Tensor.floor_ is actually executed remotely inside of a unit tests (and not
skipped). Aka - at least one of the test_all_allowlisted_tensor_methods_work_remotely_on_all_types
unit tests with your method should run ALL THE WAY TO THE END (instead of skipping
the last part.)
*Note:* adding another argument type might require some serialization work if
we don't support arguments of that type yet. If so, this is your job to add it
to the protobuf files in order to close this issue!
### Reason 2 - torch.Tensor.floor_ returns a non-supported type
If this happens, you've got a little bit of work in front of you. We don't have
pointer objects to very many remote object types. So, if your method returns anything
other than a single tensor, you probably need to add support for the type it returns
(Such as a bool, None, int, or other types).
*IMPORTANT:* do NOT return the value itself to the end user!!! Return a pointer object
to that type!
*NOTE:* at the time of writing - there are several core pieces of Syft not yet working
to allow you to return any type other than a torch tensor. If you're not comfortable
investigating what those might be - skip this issue and try again later once
someone else has solved these issues.
### Reason 3 - There's something else broken
Chase those stack traces! Talk to friends in Slack. Look at how other methods are supported.
This is a challenging project in a fast moving codebase!
And don't forget - if this project seems to complex - there are plenty of others that
might be easier.
| 2.0 | Add torch.Tensor.floor_ to allowlist and test suite -
# Description
This issue is a part of Syft 0.3.0 Epic 2: https://github.com/OpenMined/PySyft/issues/3696
In this issue, you will be adding support for remote execution of the torch.Tensor.floor_
method or property. This might be a really small project (literally a one-liner) or
it might require adding significant functionality to PySyft OR to the testing suite
in order to make sure the feature is both functional and tested.
## Step 0: Run tests and ./scripts/pre_commit.sh
Before you get started with this project, let's make sure you have everything building and testing
correctly. Clone the codebase and run:
```pip uninstall syft```
followed by
```pip install -e .```
Then run the pre-commit file (which will also run the tests)
```./scripts/pre_commit.sh```
If all of these tests pass, continue on. If not, make sure you have all the
dependencies in requirements.txt installed, etc.
## Step 1: Uncomment your method in the allowlist.py file
Inside [allowlist.py](https://github.com/OpenMined/PySyft/blob/syft_0.3.0/src/syft/lib/torch/allowlist.py) you will find a huge dictionary of methods. Find your method and uncomment the line its on. At the time
of writing this Issue (WARNING: THIS MAY HAVE CHANGED) the dictionary maps from the
string name of the method (in your case 'torch.Tensor.floor_') to the string representation
of the type the method returns.
## Step 2: Run Unit Tests
Run the following:
```python setup.py test```
And wait to see if some of the tests fail. Why might the tests fail now? I'm so glad you asked!
https://github.com/OpenMined/PySyft/blob/syft_0.3.0/tests/syft/lib/torch/tensor/tensor_remote_method_api_suite_test.py
In this file you'll find the torch method test suite. It AUTOMATICALLY loads all methods
from the allowlist.py file you modified in the previous step. It attempts to test them.
# Step 3: If you get a Failing Test
If you get a failing test, this could be for one of a few reasons:
### Reason 1 - The testing suite passed in non-compatible arguments
The testing suite is pretty dumb. It literally just has a permutation of possible
arguments to pass into every method on torch tensors. So, if one of those permutations
doesn't work for your method (aka... perhaps it tries to call your method without
any arguments but torch.Tensor.floor_ actually requires some) then the test will
fail if the error hasn't been seen before.
If this happens - don't worry! Just look inside the only test in that file and look
for the huge lists of error types to ignore. Add your error to the list and keep
going!!!
*WARNING:* make sure that the testing suite actually tests your method via remote
execution once you've gotten all the tests passing. Aka - if the testing suite
doesn't have ANY matching argument permutations for your method, then you're
literally creating a bunch of unit tests that do absolutely nothing. If this is the
case, then ADD MORE ARGUMENT TYPES TO THE TESTING SUITE so that your argument
gets run via remote execution. DO NOT CLOSE THIS ISSUE until you can verify that
torch.Tensor.floor_ is actually executed remotely inside of a unit tests (and not
skipped). Aka - at least one of the test_all_allowlisted_tensor_methods_work_remotely_on_all_types
unit tests with your method should run ALL THE WAY TO THE END (instead of skipping
the last part.)
*Note:* adding another argument type might require some serialization work if
we don't support arguments of that type yet. If so, this is your job to add it
to the protobuf files in order to close this issue!
### Reason 2 - torch.Tensor.floor_ returns a non-supported type
If this happens, you've got a little bit of work in front of you. We don't have
pointer objects to very many remote object types. So, if your method returns anything
other than a single tensor, you probably need to add support for the type it returns
(Such as a bool, None, int, or other types).
*IMPORTANT:* do NOT return the value itself to the end user!!! Return a pointer object
to that type!
*NOTE:* at the time of writing - there are several core pieces of Syft not yet working
to allow you to return any type other than a torch tensor. If you're not comfortable
investigating what those might be - skip this issue and try again later once
someone else has solved these issues.
### Reason 3 - There's something else broken
Chase those stack traces! Talk to friends in Slack. Look at how other methods are supported.
This is a challenging project in a fast moving codebase!
And don't forget - if this project seems to complex - there are plenty of others that
might be easier.
| non_infrastructure | add torch tensor floor to allowlist and test suite description this issue is a part of syft epic in this issue you will be adding support for remote execution of the torch tensor floor method or property this might be a really small project literally a one liner or it might require adding significant functionality to pysyft or to the testing suite in order to make sure the feature is both functional and tested step run tests and scripts pre commit sh before you get started with this project let s make sure you have everything building and testing correctly clone the codebase and run pip uninstall syft followed by pip install e then run the pre commit file which will also run the tests scripts pre commit sh if all of these tests pass continue on if not make sure you have all the dependencies in requirements txt installed etc step uncomment your method in the allowlist py file inside you will find a huge dictionary of methods find your method and uncomment the line its on at the time of writing this issue warning this may have changed the dictionary maps from the string name of the method in your case torch tensor floor to the string representation of the type the method returns step run unit tests run the following python setup py test and wait to see if some of the tests fail why might the tests fail now i m so glad you asked in this file you ll find the torch method test suite it automatically loads all methods from the allowlist py file you modified in the previous step it attempts to test them step if you get a failing test if you get a failing test this could be for one of a few reasons reason the testing suite passed in non compatible arguments the testing suite is pretty dumb it literally just has a permutation of possible arguments to pass into every method on torch tensors so if one of those permutations doesn t work for your method aka perhaps it tries to call your method without any arguments but torch tensor floor actually requires some then the test will fail if the error hasn t been seen before if this happens don t worry just look inside the only test in that file and look for the huge lists of error types to ignore add your error to the list and keep going warning make sure that the testing suite actually tests your method via remote execution once you ve gotten all the tests passing aka if the testing suite doesn t have any matching argument permutations for your method then you re literally creating a bunch of unit tests that do absolutely nothing if this is the case then add more argument types to the testing suite so that your argument gets run via remote execution do not close this issue until you can verify that torch tensor floor is actually executed remotely inside of a unit tests and not skipped aka at least one of the test all allowlisted tensor methods work remotely on all types unit tests with your method should run all the way to the end instead of skipping the last part note adding another argument type might require some serialization work if we don t support arguments of that type yet if so this is your job to add it to the protobuf files in order to close this issue reason torch tensor floor returns a non supported type if this happens you ve got a little bit of work in front of you we don t have pointer objects to very many remote object types so if your method returns anything other than a single tensor you probably need to add support for the type it returns such as a bool none int or other types important do not return the value itself to the end user return a pointer object to that type note at the time of writing there are several core pieces of syft not yet working to allow you to return any type other than a torch tensor if you re not comfortable investigating what those might be skip this issue and try again later once someone else has solved these issues reason there s something else broken chase those stack traces talk to friends in slack look at how other methods are supported this is a challenging project in a fast moving codebase and don t forget if this project seems to complex there are plenty of others that might be easier | 0 |
7,822 | 7,108,343,854 | IssuesEvent | 2018-01-16 23:34:40 | servo/webrender | https://api.github.com/repos/servo/webrender | closed | Capture: post-built frame layer support | area: infrastructure difficulty: moderate type: enhancement | Current capturing of #2232 fetches all the internal structures used **before** building a frame. In some cases, the bugs we are seeing may occur due to accumulation effects in the internal re-used structures, such as `TextureCache` or `GpuCache`. It would be helpful to save exactly their state in a capture, and it doesn't seem to difficult.
This serialization would touch way more internal structures, and I suppose needs to go as a separate cargo feature. | 1.0 | Capture: post-built frame layer support - Current capturing of #2232 fetches all the internal structures used **before** building a frame. In some cases, the bugs we are seeing may occur due to accumulation effects in the internal re-used structures, such as `TextureCache` or `GpuCache`. It would be helpful to save exactly their state in a capture, and it doesn't seem to difficult.
This serialization would touch way more internal structures, and I suppose needs to go as a separate cargo feature. | infrastructure | capture post built frame layer support current capturing of fetches all the internal structures used before building a frame in some cases the bugs we are seeing may occur due to accumulation effects in the internal re used structures such as texturecache or gpucache it would be helpful to save exactly their state in a capture and it doesn t seem to difficult this serialization would touch way more internal structures and i suppose needs to go as a separate cargo feature | 1 |
282,762 | 21,315,948,495 | IssuesEvent | 2022-04-16 09:20:03 | allyfern72/pe | https://api.github.com/repos/allyfern72/pe | opened | [UG] Typo in "Change the price of a dish" | severity.VeryLow type.DocumentationBug | Typo, `Delete` under "Change the price of a dish":

<!--session: 1650096024879-64fa4b51-68b4-4157-8052-f142c2406b7b-->
<!--Version: Web v3.4.2--> | 1.0 | [UG] Typo in "Change the price of a dish" - Typo, `Delete` under "Change the price of a dish":

<!--session: 1650096024879-64fa4b51-68b4-4157-8052-f142c2406b7b-->
<!--Version: Web v3.4.2--> | non_infrastructure | typo in change the price of a dish typo delete under change the price of a dish | 0 |
129,796 | 10,586,949,344 | IssuesEvent | 2019-10-08 20:53:25 | golang/go | https://api.github.com/repos/golang/go | closed | x/net/webdav: TestDir fails on nacl | OS-NaCl Testing | See http://build.golang.org/log/db23eb1244d8fd6bd544d395413cd0c700a7f6be.
```
--- FAIL: TestDir (0.01s)
file_test.go:503: test case #61 "move__ o=F /d /d/n/z want err": got "ok" (<nil>), want "err"
FAIL
FAIL golang.org/x/net/webdav 0.971s
```
| 1.0 | x/net/webdav: TestDir fails on nacl - See http://build.golang.org/log/db23eb1244d8fd6bd544d395413cd0c700a7f6be.
```
--- FAIL: TestDir (0.01s)
file_test.go:503: test case #61 "move__ o=F /d /d/n/z want err": got "ok" (<nil>), want "err"
FAIL
FAIL golang.org/x/net/webdav 0.971s
```
| non_infrastructure | x net webdav testdir fails on nacl see fail testdir file test go test case move o f d d n z want err got ok want err fail fail golang org x net webdav | 0 |
118,375 | 15,284,753,782 | IssuesEvent | 2021-02-23 12:40:27 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | opened | Site Editor: Apply block outline treatment to document | Needs Design Feedback [Feature] Site Editor | Selecting the _document_ in the Site Editor feels a little contradictory when compared to selecting blocks.
Hovering over blocks reveals an outline which helps indicate which block will be selected upon click. There is no such indication when one hovers over the _document_, so there is never 100% confidence in when attempting to select it. I'm thinking it may be worth applying the same treatment to the document itself. Please excuse the very crude mockup and red outline:

Displaying an outline on hover and upon selection aligns the document selection experience with the block selection experience in the Site Editor, increasing visual clarity. Selecting the document should also switch to the "Template" tab when the Inspector is open, this will align with how the post editor behaves currently:

I figured this might be worth a try PR, to see how it feels. | 1.0 | Site Editor: Apply block outline treatment to document - Selecting the _document_ in the Site Editor feels a little contradictory when compared to selecting blocks.
Hovering over blocks reveals an outline which helps indicate which block will be selected upon click. There is no such indication when one hovers over the _document_, so there is never 100% confidence in when attempting to select it. I'm thinking it may be worth applying the same treatment to the document itself. Please excuse the very crude mockup and red outline:

Displaying an outline on hover and upon selection aligns the document selection experience with the block selection experience in the Site Editor, increasing visual clarity. Selecting the document should also switch to the "Template" tab when the Inspector is open, this will align with how the post editor behaves currently:

I figured this might be worth a try PR, to see how it feels. | non_infrastructure | site editor apply block outline treatment to document selecting the document in the site editor feels a little contradictory when compared to selecting blocks hovering over blocks reveals an outline which helps indicate which block will be selected upon click there is no such indication when one hovers over the document so there is never confidence in when attempting to select it i m thinking it may be worth applying the same treatment to the document itself please excuse the very crude mockup and red outline displaying an outline on hover and upon selection aligns the document selection experience with the block selection experience in the site editor increasing visual clarity selecting the document should also switch to the template tab when the inspector is open this will align with how the post editor behaves currently i figured this might be worth a try pr to see how it feels | 0 |
11,601 | 9,309,421,372 | IssuesEvent | 2019-03-25 16:26:28 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | Support more recent MySQL versions in filebeat and metricbeat | :infrastructure Metricbeat enhancement module | Add test cases for MySQL/Percona 8.0 and MariaDB 10.3 in filebeat and metricbeat mysql modules. | 1.0 | Support more recent MySQL versions in filebeat and metricbeat - Add test cases for MySQL/Percona 8.0 and MariaDB 10.3 in filebeat and metricbeat mysql modules. | infrastructure | support more recent mysql versions in filebeat and metricbeat add test cases for mysql percona and mariadb in filebeat and metricbeat mysql modules | 1 |
34,688 | 30,279,870,114 | IssuesEvent | 2023-07-08 01:11:40 | nilearn/nilearn | https://api.github.com/repos/nilearn/nilearn | closed | Integrate citation.cff validation workflow to doc build workflow | Infrastructure | The doc build uses needs the citation.cff to update contributors
https://github.com/nilearn/nilearn/blob/bce754beea677d4d4a51191e21c64d09002f1429/.github/workflows/build-docs.yml#L264
Would be better to validate the citation.cff before running the doc build and only run the build if the file is valid. | 1.0 | Integrate citation.cff validation workflow to doc build workflow - The doc build uses needs the citation.cff to update contributors
https://github.com/nilearn/nilearn/blob/bce754beea677d4d4a51191e21c64d09002f1429/.github/workflows/build-docs.yml#L264
Would be better to validate the citation.cff before running the doc build and only run the build if the file is valid. | infrastructure | integrate citation cff validation workflow to doc build workflow the doc build uses needs the citation cff to update contributors would be better to validate the citation cff before running the doc build and only run the build if the file is valid | 1 |
237,065 | 19,592,057,032 | IssuesEvent | 2022-01-05 14:02:34 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Failing test: Jest Tests.packages/kbn-securitysolution-autocomplete/src/hooks/use_field_value_autocomplete - use_field_value_autocomplete returns suggestions | failed-test Team:SIEM Team: SecuritySolution | A test failed on a tracked branch
```
Error: Timed out in waitForNextUpdate after 1000ms.
at waitForNextUpdate (/opt/local-ssd/buildkite/builds/kb-c2-16-ad2e5133cf38efba/elastic/kibana-hourly/kibana/node_modules/@testing-library/react-hooks/lib/core/asyncUtils.js:102:13)
at runNextTicks (node:internal/process/task_queues:61:5)
at processTimers (node:internal/timers:497:9)
at /opt/local-ssd/buildkite/builds/kb-c2-16-ad2e5133cf38efba/elastic/kibana-hourly/kibana/packages/kbn-securitysolution-autocomplete/src/hooks/use_field_value_autocomplete/index.test.ts:285:7
```
First failure: [CI Build - master](https://buildkite.com/elastic/kibana-hourly/builds/1848#c41ad624-6c9f-4342-bc1c-4fc75257be21)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.packages/kbn-securitysolution-autocomplete/src/hooks/use_field_value_autocomplete","test.name":"use_field_value_autocomplete returns suggestions","test.failCount":1}} --> | 1.0 | Failing test: Jest Tests.packages/kbn-securitysolution-autocomplete/src/hooks/use_field_value_autocomplete - use_field_value_autocomplete returns suggestions - A test failed on a tracked branch
```
Error: Timed out in waitForNextUpdate after 1000ms.
at waitForNextUpdate (/opt/local-ssd/buildkite/builds/kb-c2-16-ad2e5133cf38efba/elastic/kibana-hourly/kibana/node_modules/@testing-library/react-hooks/lib/core/asyncUtils.js:102:13)
at runNextTicks (node:internal/process/task_queues:61:5)
at processTimers (node:internal/timers:497:9)
at /opt/local-ssd/buildkite/builds/kb-c2-16-ad2e5133cf38efba/elastic/kibana-hourly/kibana/packages/kbn-securitysolution-autocomplete/src/hooks/use_field_value_autocomplete/index.test.ts:285:7
```
First failure: [CI Build - master](https://buildkite.com/elastic/kibana-hourly/builds/1848#c41ad624-6c9f-4342-bc1c-4fc75257be21)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.packages/kbn-securitysolution-autocomplete/src/hooks/use_field_value_autocomplete","test.name":"use_field_value_autocomplete returns suggestions","test.failCount":1}} --> | non_infrastructure | failing test jest tests packages kbn securitysolution autocomplete src hooks use field value autocomplete use field value autocomplete returns suggestions a test failed on a tracked branch error timed out in waitfornextupdate after at waitfornextupdate opt local ssd buildkite builds kb elastic kibana hourly kibana node modules testing library react hooks lib core asyncutils js at runnextticks node internal process task queues at processtimers node internal timers at opt local ssd buildkite builds kb elastic kibana hourly kibana packages kbn securitysolution autocomplete src hooks use field value autocomplete index test ts first failure | 0 |
5,069 | 5,415,817,484 | IssuesEvent | 2017-03-01 22:33:34 | vmware/docker-volume-vsphere | https://api.github.com/repos/vmware/docker-volume-vsphere | closed | Improve test coverage by addressing test gaps for vSphere docker volume plugin | component/test-infrastructure kind/test | This issue is focusing on adding more tests for plugin module.
- [ ] #743 (exercise volume operations from multiple endpoint)
- [x] #794 (create vSphere volume on newly created vm)
- [ ] #820 (update access option for the volume) | 1.0 | Improve test coverage by addressing test gaps for vSphere docker volume plugin - This issue is focusing on adding more tests for plugin module.
- [ ] #743 (exercise volume operations from multiple endpoint)
- [x] #794 (create vSphere volume on newly created vm)
- [ ] #820 (update access option for the volume) | infrastructure | improve test coverage by addressing test gaps for vsphere docker volume plugin this issue is focusing on adding more tests for plugin module exercise volume operations from multiple endpoint create vsphere volume on newly created vm update access option for the volume | 1 |
5,637 | 5,833,646,952 | IssuesEvent | 2017-05-09 02:37:10 | servo/servo | https://api.github.com/repos/servo/servo | closed | stylo builds aren't working on TravisCI | A-infrastructure | ```
$ ./mach test-stylo
Compiling stylo_tests v0.0.1 (file:///home/travis/build/servo/servo/tests/unit/stylo)
warning: value assigned to `saw_before` is never read, #[warn(unused_assignments)] on by default
--> /home/travis/build/servo/servo/tests/unit/stylo/sanity_checks.rs:50:9
|
50 | let mut saw_before = false;
| ^^^^^^^^^^^^^^
warning: value assigned to `saw_after` is never read, #[warn(unused_assignments)] on by default
--> /home/travis/build/servo/servo/tests/unit/stylo/sanity_checks.rs:51:9
|
51 | let mut saw_after = false;
| ^^^^^^^^^^^^^
warning: struct is never used: `StylesheetLoader`, #[warn(dead_code)] on by default
--> /home/travis/build/servo/servo/tests/unit/stylo/../../../ports/geckolib/stylesheet_loader.rs:9:1
|
9 | pub struct StylesheetLoader;
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
warning: method is never used: `new`, #[warn(dead_code)] on by default
--> /home/travis/build/servo/servo/tests/unit/stylo/../../../ports/geckolib/stylesheet_loader.rs:12:5
|
12 | pub fn new() -> Self {
| ^
Finished debug [unoptimized + debuginfo] target(s) in 20.29 secs
Running /home/travis/build/servo/servo/target/geckolib/debug/deps/stylo_tests-019e27a7cd146829
running 2 tests
test sanity_checks::assert_basic_pseudo_elements ... ok
test sanity_checks::assert_restyle_hints_match ... ok
test result: ok. 2 passed; 0 failed; 0 ignored; 0 measured
Compiling style v0.0.1 (file:///home/travis/build/servo/servo/components/style)
LLVM ERROR: IO failure on output stream.
error: Could not compile `style`.
To learn more, run the command again with --verbose.
The command "./mach test-stylo" exited with 101.
``` | 1.0 | stylo builds aren't working on TravisCI - ```
$ ./mach test-stylo
Compiling stylo_tests v0.0.1 (file:///home/travis/build/servo/servo/tests/unit/stylo)
warning: value assigned to `saw_before` is never read, #[warn(unused_assignments)] on by default
--> /home/travis/build/servo/servo/tests/unit/stylo/sanity_checks.rs:50:9
|
50 | let mut saw_before = false;
| ^^^^^^^^^^^^^^
warning: value assigned to `saw_after` is never read, #[warn(unused_assignments)] on by default
--> /home/travis/build/servo/servo/tests/unit/stylo/sanity_checks.rs:51:9
|
51 | let mut saw_after = false;
| ^^^^^^^^^^^^^
warning: struct is never used: `StylesheetLoader`, #[warn(dead_code)] on by default
--> /home/travis/build/servo/servo/tests/unit/stylo/../../../ports/geckolib/stylesheet_loader.rs:9:1
|
9 | pub struct StylesheetLoader;
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
warning: method is never used: `new`, #[warn(dead_code)] on by default
--> /home/travis/build/servo/servo/tests/unit/stylo/../../../ports/geckolib/stylesheet_loader.rs:12:5
|
12 | pub fn new() -> Self {
| ^
Finished debug [unoptimized + debuginfo] target(s) in 20.29 secs
Running /home/travis/build/servo/servo/target/geckolib/debug/deps/stylo_tests-019e27a7cd146829
running 2 tests
test sanity_checks::assert_basic_pseudo_elements ... ok
test sanity_checks::assert_restyle_hints_match ... ok
test result: ok. 2 passed; 0 failed; 0 ignored; 0 measured
Compiling style v0.0.1 (file:///home/travis/build/servo/servo/components/style)
LLVM ERROR: IO failure on output stream.
error: Could not compile `style`.
To learn more, run the command again with --verbose.
The command "./mach test-stylo" exited with 101.
``` | infrastructure | stylo builds aren t working on travisci mach test stylo compiling stylo tests file home travis build servo servo tests unit stylo warning value assigned to saw before is never read on by default home travis build servo servo tests unit stylo sanity checks rs let mut saw before false warning value assigned to saw after is never read on by default home travis build servo servo tests unit stylo sanity checks rs let mut saw after false warning struct is never used stylesheetloader on by default home travis build servo servo tests unit stylo ports geckolib stylesheet loader rs pub struct stylesheetloader warning method is never used new on by default home travis build servo servo tests unit stylo ports geckolib stylesheet loader rs pub fn new self finished debug target s in secs running home travis build servo servo target geckolib debug deps stylo tests running tests test sanity checks assert basic pseudo elements ok test sanity checks assert restyle hints match ok test result ok passed failed ignored measured compiling style file home travis build servo servo components style llvm error io failure on output stream error could not compile style to learn more run the command again with verbose the command mach test stylo exited with | 1 |
35,259 | 30,873,326,679 | IssuesEvent | 2023-08-03 12:50:55 | microsoftgraph/microsoft-graph-explorer-v4 | https://api.github.com/repos/microsoftgraph/microsoft-graph-explorer-v4 | closed | [BUG] Incompatible dependencies | Needs: Attention 👋 type:infrastructure | **Describe the bug**
The project has dependencies with conflicting peer dependencies.
**To Reproduce**
Steps to reproduce the behavior:
1. Using node v16.20.0 and npm 8.19.4
1. Run `npm i`
1. package-lock.json is updated to version 2 format
1. Run `npm i`
1. Observe that npm fails to complete the install
this also occurs in CI builds where the v2 package-lock.json is committed
See these CI runs:
- https://microsoftgraph.visualstudio.com/Graph%20Developer%20Experiences/_build/results?buildId=119352&view=logs&j=cb4db6dd-e658-55b7-b337-8761517bc1a3&t=5f69cb92-a4fc-59e7-06e5-11f91b5ae56f
- https://github.com/microsoftgraph/microsoft-graph-explorer-v4/actions/runs/5469927279/jobs/9959414580
**Expected behavior**
npm i works without error when run multiple times.
**Desktop (please complete the following information):**
- OS: Ubuntu 20.04
- Browser n/a
- node: 16.20.0
- npm: 8.19.4
| 1.0 | [BUG] Incompatible dependencies - **Describe the bug**
The project has dependencies with conflicting peer dependencies.
**To Reproduce**
Steps to reproduce the behavior:
1. Using node v16.20.0 and npm 8.19.4
1. Run `npm i`
1. package-lock.json is updated to version 2 format
1. Run `npm i`
1. Observe that npm fails to complete the install
this also occurs in CI builds where the v2 package-lock.json is committed
See these CI runs:
- https://microsoftgraph.visualstudio.com/Graph%20Developer%20Experiences/_build/results?buildId=119352&view=logs&j=cb4db6dd-e658-55b7-b337-8761517bc1a3&t=5f69cb92-a4fc-59e7-06e5-11f91b5ae56f
- https://github.com/microsoftgraph/microsoft-graph-explorer-v4/actions/runs/5469927279/jobs/9959414580
**Expected behavior**
npm i works without error when run multiple times.
**Desktop (please complete the following information):**
- OS: Ubuntu 20.04
- Browser n/a
- node: 16.20.0
- npm: 8.19.4
| infrastructure | incompatible dependencies describe the bug the project has dependencies with conflicting peer dependencies to reproduce steps to reproduce the behavior using node and npm run npm i package lock json is updated to version format run npm i observe that npm fails to complete the install this also occurs in ci builds where the package lock json is committed see these ci runs expected behavior npm i works without error when run multiple times desktop please complete the following information os ubuntu browser n a node npm | 1 |
446,538 | 31,494,512,653 | IssuesEvent | 2023-08-31 00:25:33 | PocketRelay/Website | https://api.github.com/repos/PocketRelay/Website | closed | Remove crates.io references | documentation | ## Description
As crates.io builds are no longer supported (See https://github.com/PocketRelay/Server/issues/40) references to building through it should be removed
## Checklist
- [ ] Crates.io references are removed | 1.0 | Remove crates.io references - ## Description
As crates.io builds are no longer supported (See https://github.com/PocketRelay/Server/issues/40) references to building through it should be removed
## Checklist
- [ ] Crates.io references are removed | non_infrastructure | remove crates io references description as crates io builds are no longer supported see references to building through it should be removed checklist crates io references are removed | 0 |
21,334 | 14,528,664,690 | IssuesEvent | 2020-12-14 16:49:00 | pythonitalia/pycon | https://api.github.com/repos/pythonitalia/pycon | closed | Fix "Migrate DB" action | infrastructure | ubuntu-18 (latest) on GitHub actions has AWS CLI v1 which doesn't support payload-type
We can switch to ubuntu-20 which has CLI v2 or install the latest CLI manually | 1.0 | Fix "Migrate DB" action - ubuntu-18 (latest) on GitHub actions has AWS CLI v1 which doesn't support payload-type
We can switch to ubuntu-20 which has CLI v2 or install the latest CLI manually | infrastructure | fix migrate db action ubuntu latest on github actions has aws cli which doesn t support payload type we can switch to ubuntu which has cli or install the latest cli manually | 1 |
14,077 | 10,596,576,608 | IssuesEvent | 2019-10-09 21:36:12 | celo-org/celo-monorepo | https://api.github.com/repos/celo-org/celo-monorepo | closed | celotool commands that use terraform should have easy to understand logs | infrastructure | ### Expected Behavior
Easy to understand output
### Current Behavior
Some logs are double printed, or unnecessarily printed (like when running `terraform init` before another terraform command that performs behind the scenes)
| 1.0 | celotool commands that use terraform should have easy to understand logs - ### Expected Behavior
Easy to understand output
### Current Behavior
Some logs are double printed, or unnecessarily printed (like when running `terraform init` before another terraform command that performs behind the scenes)
| infrastructure | celotool commands that use terraform should have easy to understand logs expected behavior easy to understand output current behavior some logs are double printed or unnecessarily printed like when running terraform init before another terraform command that performs behind the scenes | 1 |
308,586 | 26,615,352,908 | IssuesEvent | 2023-01-24 06:21:08 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | sql/tests: TestRandomSyntaxFunctions failed | C-test-failure O-robot branch-master | sql/tests.TestRandomSyntaxFunctions [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/8425837?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/8425837?buildTab=artifacts#/) on master @ [dc974cd698364a4db5acb630162dd8b1856cfad6](https://github.com/cockroachdb/cockroach/commits/dc974cd698364a4db5acb630162dd8b1856cfad6):
Fatal error:
```
panic: test timed out after 14m55s
```
Stack:
```
goroutine 6669292 [running]:
testing.(*M).startAlarm.func1()
GOROOT/src/testing/testing.go:2036 +0x8e
created by time.goFunc
GOROOT/src/time/sleep.go:176 +0x32
```
<details><summary>Log preceding fatal error</summary>
<p>
```
SELECT crdb_internal.num_inverted_index_entries(NULL::tsvector, (SELECT NULL)::int) ;
Stack trace:
rsg_test.go:855: Crash detected: server panic: pq: internal error: unexpected error from the vectorized engine: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries((SELECT NULL)::tsvector, (-1606389283102320563):::INT8) ;
Stack trace:
rsg_test.go:828: 4m0s of 5m0s: 650218 executions, 401541 successful
rsg_test.go:855: Crash detected: server panic: pq: internal error: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries(NULL::tsvector, 7481765445855844981:::INT8) ;
Stack trace:
rsg_test.go:828: 4m5s of 5m0s: 665226 executions, 410750 successful
rsg_test.go:828: 4m10s of 5m0s: 677207 executions, 418230 successful
rsg_test.go:828: 4m15s of 5m0s: 690441 executions, 426418 successful
rsg_test.go:828: 4m20s of 5m0s: 705400 executions, 435705 successful
rsg_test.go:855: Crash detected: server panic: pq: internal error: unexpected error from the vectorized engine: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries((SELECT NULL)::tsvector, 5455364076106125598:::INT8) ;
Stack trace:
rsg_test.go:828: 4m25s of 5m0s: 717064 executions, 442888 successful
rsg_test.go:855: Crash detected: server panic: pq: internal error: unexpected error from the vectorized engine: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries((SELECT NULL)::tsvector, NULL::int) ;
Stack trace:
rsg_test.go:828: 4m30s of 5m0s: 730078 executions, 451024 successful
rsg_test.go:855: Crash detected: server panic: pq: internal error: unexpected error from the vectorized engine: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries((SELECT NULL)::tsvector, (-32768):::INT8) ;
Stack trace:
rsg_test.go:855: Crash detected: server panic: pq: internal error: unexpected error from the vectorized engine: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries((SELECT NULL)::tsvector, 611836342379566618:::INT8) ;
Stack trace:
rsg_test.go:855: Crash detected: server panic: pq: internal error: unexpected error from the vectorized engine: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries((SELECT NULL)::tsvector, 3003255369204358124:::INT8) ;
Stack trace:
rsg_test.go:828: 4m35s of 5m0s: 742535 executions, 458659 successful
rsg_test.go:828: 4m40s of 5m0s: 753955 executions, 465738 successful
rsg_test.go:828: 4m45s of 5m0s: 765568 executions, 472992 successful
rsg_test.go:855: Crash detected: server panic: pq: internal error: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries(NULL::tsvector, 940542034895810029:::INT8) ;
Stack trace:
rsg_test.go:855: Crash detected: server panic: pq: internal error: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries(NULL::tsvector, (-7970048270221468361):::INT8) ;
Stack trace:
rsg_test.go:828: 4m50s of 5m0s: 777690 executions, 480381 successful
```
</p>
</details>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #95619 sql/tests: TestRandomSyntaxFunctions failed [C-test-failure O-robot T-sql-sessions branch-release-22.2]
- #95617 sql/tests: TestRandomSyntaxFunctions failed [C-test-failure O-robot T-sql-sessions branch-release-22.2]
- #95616 sql/tests: TestRandomSyntaxFunctions failed [C-test-failure O-robot T-sql-sessions branch-release-22.2]
- #87571 sql/tests: TestRandomSyntaxFunctions failed [C-test-failure O-robot branch-release-22.2]
</p>
</details>
/cc @cockroachdb/sql-sessions
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRandomSyntaxFunctions.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 1.0 | sql/tests: TestRandomSyntaxFunctions failed - sql/tests.TestRandomSyntaxFunctions [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/8425837?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RandomSyntaxTestsBazel/8425837?buildTab=artifacts#/) on master @ [dc974cd698364a4db5acb630162dd8b1856cfad6](https://github.com/cockroachdb/cockroach/commits/dc974cd698364a4db5acb630162dd8b1856cfad6):
Fatal error:
```
panic: test timed out after 14m55s
```
Stack:
```
goroutine 6669292 [running]:
testing.(*M).startAlarm.func1()
GOROOT/src/testing/testing.go:2036 +0x8e
created by time.goFunc
GOROOT/src/time/sleep.go:176 +0x32
```
<details><summary>Log preceding fatal error</summary>
<p>
```
SELECT crdb_internal.num_inverted_index_entries(NULL::tsvector, (SELECT NULL)::int) ;
Stack trace:
rsg_test.go:855: Crash detected: server panic: pq: internal error: unexpected error from the vectorized engine: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries((SELECT NULL)::tsvector, (-1606389283102320563):::INT8) ;
Stack trace:
rsg_test.go:828: 4m0s of 5m0s: 650218 executions, 401541 successful
rsg_test.go:855: Crash detected: server panic: pq: internal error: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries(NULL::tsvector, 7481765445855844981:::INT8) ;
Stack trace:
rsg_test.go:828: 4m5s of 5m0s: 665226 executions, 410750 successful
rsg_test.go:828: 4m10s of 5m0s: 677207 executions, 418230 successful
rsg_test.go:828: 4m15s of 5m0s: 690441 executions, 426418 successful
rsg_test.go:828: 4m20s of 5m0s: 705400 executions, 435705 successful
rsg_test.go:855: Crash detected: server panic: pq: internal error: unexpected error from the vectorized engine: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries((SELECT NULL)::tsvector, 5455364076106125598:::INT8) ;
Stack trace:
rsg_test.go:828: 4m25s of 5m0s: 717064 executions, 442888 successful
rsg_test.go:855: Crash detected: server panic: pq: internal error: unexpected error from the vectorized engine: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries((SELECT NULL)::tsvector, NULL::int) ;
Stack trace:
rsg_test.go:828: 4m30s of 5m0s: 730078 executions, 451024 successful
rsg_test.go:855: Crash detected: server panic: pq: internal error: unexpected error from the vectorized engine: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries((SELECT NULL)::tsvector, (-32768):::INT8) ;
Stack trace:
rsg_test.go:855: Crash detected: server panic: pq: internal error: unexpected error from the vectorized engine: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries((SELECT NULL)::tsvector, 611836342379566618:::INT8) ;
Stack trace:
rsg_test.go:855: Crash detected: server panic: pq: internal error: unexpected error from the vectorized engine: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries((SELECT NULL)::tsvector, 3003255369204358124:::INT8) ;
Stack trace:
rsg_test.go:828: 4m35s of 5m0s: 742535 executions, 458659 successful
rsg_test.go:828: 4m40s of 5m0s: 753955 executions, 465738 successful
rsg_test.go:828: 4m45s of 5m0s: 765568 executions, 472992 successful
rsg_test.go:855: Crash detected: server panic: pq: internal error: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries(NULL::tsvector, 940542034895810029:::INT8) ;
Stack trace:
rsg_test.go:855: Crash detected: server panic: pq: internal error: interface conversion: tree.Datum is tree.dNull, not *tree.DTSVector
SELECT crdb_internal.num_inverted_index_entries(NULL::tsvector, (-7970048270221468361):::INT8) ;
Stack trace:
rsg_test.go:828: 4m50s of 5m0s: 777690 executions, 480381 successful
```
</p>
</details>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #95619 sql/tests: TestRandomSyntaxFunctions failed [C-test-failure O-robot T-sql-sessions branch-release-22.2]
- #95617 sql/tests: TestRandomSyntaxFunctions failed [C-test-failure O-robot T-sql-sessions branch-release-22.2]
- #95616 sql/tests: TestRandomSyntaxFunctions failed [C-test-failure O-robot T-sql-sessions branch-release-22.2]
- #87571 sql/tests: TestRandomSyntaxFunctions failed [C-test-failure O-robot branch-release-22.2]
</p>
</details>
/cc @cockroachdb/sql-sessions
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRandomSyntaxFunctions.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| non_infrastructure | sql tests testrandomsyntaxfunctions failed sql tests testrandomsyntaxfunctions with on master fatal error panic test timed out after stack goroutine testing m startalarm goroot src testing testing go created by time gofunc goroot src time sleep go log preceding fatal error select crdb internal num inverted index entries null tsvector select null int stack trace rsg test go crash detected server panic pq internal error unexpected error from the vectorized engine interface conversion tree datum is tree dnull not tree dtsvector select crdb internal num inverted index entries select null tsvector stack trace rsg test go of executions successful rsg test go crash detected server panic pq internal error interface conversion tree datum is tree dnull not tree dtsvector select crdb internal num inverted index entries null tsvector stack trace rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go crash detected server panic pq internal error unexpected error from the vectorized engine interface conversion tree datum is tree dnull not tree dtsvector select crdb internal num inverted index entries select null tsvector stack trace rsg test go of executions successful rsg test go crash detected server panic pq internal error unexpected error from the vectorized engine interface conversion tree datum is tree dnull not tree dtsvector select crdb internal num inverted index entries select null tsvector null int stack trace rsg test go of executions successful rsg test go crash detected server panic pq internal error unexpected error from the vectorized engine interface conversion tree datum is tree dnull not tree dtsvector select crdb internal num inverted index entries select null tsvector stack trace rsg test go crash detected server panic pq internal error unexpected error from the vectorized engine interface conversion tree datum is tree dnull not tree dtsvector select crdb internal num inverted index entries select null tsvector stack trace rsg test go crash detected server panic pq internal error unexpected error from the vectorized engine interface conversion tree datum is tree dnull not tree dtsvector select crdb internal num inverted index entries select null tsvector stack trace rsg test go of executions successful rsg test go of executions successful rsg test go of executions successful rsg test go crash detected server panic pq internal error interface conversion tree datum is tree dnull not tree dtsvector select crdb internal num inverted index entries null tsvector stack trace rsg test go crash detected server panic pq internal error interface conversion tree datum is tree dnull not tree dtsvector select crdb internal num inverted index entries null tsvector stack trace rsg test go of executions successful help see also same failure on other branches sql tests testrandomsyntaxfunctions failed sql tests testrandomsyntaxfunctions failed sql tests testrandomsyntaxfunctions failed sql tests testrandomsyntaxfunctions failed cc cockroachdb sql sessions | 0 |
414,509 | 27,990,200,371 | IssuesEvent | 2023-03-27 02:32:41 | rizkytegar/express-api | https://api.github.com/repos/rizkytegar/express-api | opened | Feature Overview And Summary of Project Progress | documentation good first issue | [x] Fixed Routes
[x] Fixed Controller
[x] Fixed Model
[x] Add Environment Variable (.env)
[x] Sequelize ORM for SQL Database
[x] Containerized With Docker
[ ] Using MongoDB
[ ] Unit Testing
[ ] Documentation
[ ] Add Authentication for Client Services
[ ] Using JWT token
[ ] Deploy | 1.0 | Feature Overview And Summary of Project Progress - [x] Fixed Routes
[x] Fixed Controller
[x] Fixed Model
[x] Add Environment Variable (.env)
[x] Sequelize ORM for SQL Database
[x] Containerized With Docker
[ ] Using MongoDB
[ ] Unit Testing
[ ] Documentation
[ ] Add Authentication for Client Services
[ ] Using JWT token
[ ] Deploy | non_infrastructure | feature overview and summary of project progress fixed routes fixed controller fixed model add environment variable env sequelize orm for sql database containerized with docker using mongodb unit testing documentation add authentication for client services using jwt token deploy | 0 |
14,855 | 11,202,880,576 | IssuesEvent | 2020-01-04 15:51:36 | PrivacyLx/privacylx-issue-tracker | https://api.github.com/repos/PrivacyLx/privacylx-issue-tracker | closed | Bridge PrivacyLx internal channels | infrastructure irc matrix | This issue is about creating a bridge to PrivacyLx internal channels.
Currently it seems that most people are using Matrix and IRC.
We can later add support for bridges as we see fit.
I added a WIP branch of an Ansible role that setups [Matterbridge](https://github.com/42wim/matterbridge) in order to bridge PrivacyLx channels. The branch can be found here: https://github.com/PrivacyLx/devops/tree/ansible/matterbridge-role | 1.0 | Bridge PrivacyLx internal channels - This issue is about creating a bridge to PrivacyLx internal channels.
Currently it seems that most people are using Matrix and IRC.
We can later add support for bridges as we see fit.
I added a WIP branch of an Ansible role that setups [Matterbridge](https://github.com/42wim/matterbridge) in order to bridge PrivacyLx channels. The branch can be found here: https://github.com/PrivacyLx/devops/tree/ansible/matterbridge-role | infrastructure | bridge privacylx internal channels this issue is about creating a bridge to privacylx internal channels currently it seems that most people are using matrix and irc we can later add support for bridges as we see fit i added a wip branch of an ansible role that setups in order to bridge privacylx channels the branch can be found here | 1 |
273,510 | 29,831,010,620 | IssuesEvent | 2023-06-18 09:17:35 | RG4421/ampere-centos-kernel | https://api.github.com/repos/RG4421/ampere-centos-kernel | closed | CVE-2022-1998 (High) detected in linuxv5.2 - autoclosed | Mend: dependency security vulnerability | ## CVE-2022-1998 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/notify/fanotify/fanotify_user.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/notify/fanotify/fanotify_user.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A use after free in the Linux kernel File System notify functionality was found in the way user triggers copy_info_records_to_user() call to fail in copy_event_to_user(). A local user could use this flaw to crash the system or potentially escalate their privileges on the system.
<p>Publish Date: 2022-06-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1998>CVE-2022-1998</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1998">https://www.linuxkernelcves.com/cves/CVE-2022-1998</a></p>
<p>Release Date: 2022-06-09</p>
<p>Fix Resolution: v5.10.97,v5.15.20,v5.16.6,v5.17-rc3</p>
</p>
</details>
<p></p>
| True | CVE-2022-1998 (High) detected in linuxv5.2 - autoclosed - ## CVE-2022-1998 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/notify/fanotify/fanotify_user.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/notify/fanotify/fanotify_user.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A use after free in the Linux kernel File System notify functionality was found in the way user triggers copy_info_records_to_user() call to fail in copy_event_to_user(). A local user could use this flaw to crash the system or potentially escalate their privileges on the system.
<p>Publish Date: 2022-06-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1998>CVE-2022-1998</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1998">https://www.linuxkernelcves.com/cves/CVE-2022-1998</a></p>
<p>Release Date: 2022-06-09</p>
<p>Fix Resolution: v5.10.97,v5.15.20,v5.16.6,v5.17-rc3</p>
</p>
</details>
<p></p>
| non_infrastructure | cve high detected in autoclosed cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files fs notify fanotify fanotify user c fs notify fanotify fanotify user c vulnerability details a use after free in the linux kernel file system notify functionality was found in the way user triggers copy info records to user call to fail in copy event to user a local user could use this flaw to crash the system or potentially escalate their privileges on the system publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
369,225 | 25,833,620,789 | IssuesEvent | 2022-12-12 17:51:00 | seanmturley/natureddit | https://api.github.com/repos/seanmturley/natureddit | closed | Add initial README | documentation | # Description
Add a _README_ documenting the project, including the following:
- Aim (including learning goals for myself)
- Wireframes
- Technologies used
- Features
- Future work
| 1.0 | Add initial README - # Description
Add a _README_ documenting the project, including the following:
- Aim (including learning goals for myself)
- Wireframes
- Technologies used
- Features
- Future work
| non_infrastructure | add initial readme description add a readme documenting the project including the following aim including learning goals for myself wireframes technologies used features future work | 0 |
8,338 | 7,346,088,550 | IssuesEvent | 2018-03-07 19:32:05 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | Jenkins unit tests generating Gtk warnings | bug interface/infrastructure | The Unit Tests stage on Jenkins shows a lot of Gtk warnings. It probably shouldn't be pulling in Gtk at all, as the Gtk side of things will be covered in the UI tests. | 1.0 | Jenkins unit tests generating Gtk warnings - The Unit Tests stage on Jenkins shows a lot of Gtk warnings. It probably shouldn't be pulling in Gtk at all, as the Gtk side of things will be covered in the UI tests. | infrastructure | jenkins unit tests generating gtk warnings the unit tests stage on jenkins shows a lot of gtk warnings it probably shouldn t be pulling in gtk at all as the gtk side of things will be covered in the ui tests | 1 |
223,111 | 17,103,994,275 | IssuesEvent | 2021-07-09 15:01:26 | PaloAltoNetworks/terraform-provider-prismacloudcompute | https://api.github.com/repos/PaloAltoNetworks/terraform-provider-prismacloudcompute | opened | Documentation links no longer valid | documentation | ## Documentation link
https://github.com/PaloAltoNetworks/terraform-provider-prismacloudcompute
## Describe the problem
The links to the documentation point to https://www.terraform.io/docs/providers/prismacloudcompute/index.html which is no longer valid and generates an error.
I searched in the terraform registry and did not find prisma cloud compute listed there.
https://registry.terraform.io/search/providers?q=prisma
## Suggested fix
Update documentation to put it into the terraform registry then update links to point there
| 1.0 | Documentation links no longer valid - ## Documentation link
https://github.com/PaloAltoNetworks/terraform-provider-prismacloudcompute
## Describe the problem
The links to the documentation point to https://www.terraform.io/docs/providers/prismacloudcompute/index.html which is no longer valid and generates an error.
I searched in the terraform registry and did not find prisma cloud compute listed there.
https://registry.terraform.io/search/providers?q=prisma
## Suggested fix
Update documentation to put it into the terraform registry then update links to point there
| non_infrastructure | documentation links no longer valid documentation link describe the problem the links to the documentation point to which is no longer valid and generates an error i searched in the terraform registry and did not find prisma cloud compute listed there suggested fix update documentation to put it into the terraform registry then update links to point there | 0 |
144,123 | 22,281,346,824 | IssuesEvent | 2022-06-11 00:22:29 | aws/aws-toolkit-vscode | https://api.github.com/repos/aws/aws-toolkit-vscode | closed | Provide link of Learn AWS in the Welcome Page | feature-request needs-design | The Toolkit is a great place to help users develop an understanding about AWS. Determine how and where to guide those curious about learning more.
Collaborate with the [JetBrains Toolkit](https://github.com/aws/aws-toolkit-jetbrains/issues/1020) on design.
This would possibly go into the Welcome Page (#610) as an initial implementation | 1.0 | Provide link of Learn AWS in the Welcome Page - The Toolkit is a great place to help users develop an understanding about AWS. Determine how and where to guide those curious about learning more.
Collaborate with the [JetBrains Toolkit](https://github.com/aws/aws-toolkit-jetbrains/issues/1020) on design.
This would possibly go into the Welcome Page (#610) as an initial implementation | non_infrastructure | provide link of learn aws in the welcome page the toolkit is a great place to help users develop an understanding about aws determine how and where to guide those curious about learning more collaborate with the on design this would possibly go into the welcome page as an initial implementation | 0 |
4,808 | 5,283,472,390 | IssuesEvent | 2017-02-07 21:29:14 | typelevel/general | https://api.github.com/repos/typelevel/general | closed | New Gitter chatrooms for alternate FP discussions | Infrastructure | Yesterday in the Cats Gitter channel some of us started talking about topics not strictly related to the Cats project, but still about functional programming in general (e.g. algebraic effects, Free, derivative parsers). There were [some concerns](https://gitter.im/typelevel/cats?at=587eaef0300f220a66fdff2b) around this perhaps being too off-putting for a channel like typelevel/cats and after brief discussion [typelevel/oleg](https://gitter.im/typelevel/oleg) was created.
That created another concern which is that the naming of the channel might seem too clique-y and in general can come off as unwelcoming - a sort of typelevel/cool-kids.
This issue is to discuss if such an alternate room needs to be created at all, and if so what to name it. Potential names that have come up:
* general-fp
* Has naming overlap with existing typelevel/general room
* theory
* Has unfortunate implication of the (arguably false) theory/practicum dichotomy
* fp | 1.0 | New Gitter chatrooms for alternate FP discussions - Yesterday in the Cats Gitter channel some of us started talking about topics not strictly related to the Cats project, but still about functional programming in general (e.g. algebraic effects, Free, derivative parsers). There were [some concerns](https://gitter.im/typelevel/cats?at=587eaef0300f220a66fdff2b) around this perhaps being too off-putting for a channel like typelevel/cats and after brief discussion [typelevel/oleg](https://gitter.im/typelevel/oleg) was created.
That created another concern which is that the naming of the channel might seem too clique-y and in general can come off as unwelcoming - a sort of typelevel/cool-kids.
This issue is to discuss if such an alternate room needs to be created at all, and if so what to name it. Potential names that have come up:
* general-fp
* Has naming overlap with existing typelevel/general room
* theory
* Has unfortunate implication of the (arguably false) theory/practicum dichotomy
* fp | infrastructure | new gitter chatrooms for alternate fp discussions yesterday in the cats gitter channel some of us started talking about topics not strictly related to the cats project but still about functional programming in general e g algebraic effects free derivative parsers there were around this perhaps being too off putting for a channel like typelevel cats and after brief discussion was created that created another concern which is that the naming of the channel might seem too clique y and in general can come off as unwelcoming a sort of typelevel cool kids this issue is to discuss if such an alternate room needs to be created at all and if so what to name it potential names that have come up general fp has naming overlap with existing typelevel general room theory has unfortunate implication of the arguably false theory practicum dichotomy fp | 1 |
3,345 | 2,757,934,242 | IssuesEvent | 2015-04-27 17:25:37 | CORE-POS/IS4C | https://api.github.com/repos/CORE-POS/IS4C | closed | Agenda Idea for 2015-04-17 - Everything Inventory | Agendas and Minutes Contains Documentation | With the recent March 31 end of the fiscal year for many co-ops, inventory has been on our minds and maybe kept us up late. The topic has also come up recently as an extension to the transactions-and-reports base of POS. There are many facets to it and constraints and points of view on how it should be done.
Inventory support is a work in progress in CORE. Andy Theuninck's [*CORE-POS Whitepaper on Inventory*] (https://drive.google.com/file/d/0B24ZJQ7zS83qakkxSXhYMHFUMjA) (PDF) is the best general starting point, but today we can go into detail on some of the issues below, or others that are on our minds.
The whitepaper begins:
>Inventory is a very simple concept: how much product do we have on hand, right this moment? The devil is in the details as the number constantly fluctuates.
>
>Inventory can also be understood as a means to an end: items on hand is merely one component of measuring whether or not a store is making margin, overall or on an item or group of items.
>
>So how do co-ops approach this problem? There are two somewhat distinct methods:
Inventory Count vs/and Perpetual Inventory
+ Why one or the other, or some of each
+ tradeoff: Cost vs Complete, Current Knowledge
#### Inventory Count Tools
+ Shelf Audit Plugin
The interface is formatted for use on a portable device with a handheld scanner.

+ Other tools, services used
#### Perpetual Inventory
+ What it involves: counting everything all the time
+ What CORE currently offers
+ The essential `vendorItems` `<=>` `products` link is [described in detail] (https://github.com/CORE-POS/IS4C/wiki/Products-and-Vendors)
+ `PurchaseOrder` and `PurchaseOrderItems`
+ `[Ordered]quantity`
+ `receivedQty`
+ For Inventory, the crucial number is `receivedQty`
+ Purchase Orders:
+ created
+ loaded
+ from external systems
+ from invoices/manifests
+ data capture issues
+ What remains to be developed
+ Reconciling purchase orders against invoices/manifests
+ Comparing purchase orders to movement
+ Receiving use cases
+ floor
+ backoffice
#### Shrink recording at Cash
+ `SK`

`trans_status` = `Z` for all items on the screen
+ Returns and shrink: `RF###`

`description` `RF: Spoiled` applies to next item, but for it `trans_status` = `R`, not `Z`
+ StoreCharge accounts for intra-co-op transfers
+ A Shrink member account
#### Backend Shrink Recording
+ Since the whitepaper was written tools have have been added with finer-grain notation of reasons:

This (`fannie/item/shrink/ShrinkTool.php`) adds `trans_status` = `Z` items.
+ There is also a tool for editing current-day shrink items.
+ Co-ops can have their own set of Reasons for shrink.
#### Monitoring margin
+ Reports
+ Movement Reports
+ Margin Movement
+ Monitoring loss
#### Other ways to achieve the objective:
+ Monitoring cost
+ Spot-checking received orders
#### What to implement next?
+ What is important to your co-op? | 1.0 | Agenda Idea for 2015-04-17 - Everything Inventory - With the recent March 31 end of the fiscal year for many co-ops, inventory has been on our minds and maybe kept us up late. The topic has also come up recently as an extension to the transactions-and-reports base of POS. There are many facets to it and constraints and points of view on how it should be done.
Inventory support is a work in progress in CORE. Andy Theuninck's [*CORE-POS Whitepaper on Inventory*] (https://drive.google.com/file/d/0B24ZJQ7zS83qakkxSXhYMHFUMjA) (PDF) is the best general starting point, but today we can go into detail on some of the issues below, or others that are on our minds.
The whitepaper begins:
>Inventory is a very simple concept: how much product do we have on hand, right this moment? The devil is in the details as the number constantly fluctuates.
>
>Inventory can also be understood as a means to an end: items on hand is merely one component of measuring whether or not a store is making margin, overall or on an item or group of items.
>
>So how do co-ops approach this problem? There are two somewhat distinct methods:
Inventory Count vs/and Perpetual Inventory
+ Why one or the other, or some of each
+ tradeoff: Cost vs Complete, Current Knowledge
#### Inventory Count Tools
+ Shelf Audit Plugin
The interface is formatted for use on a portable device with a handheld scanner.

+ Other tools, services used
#### Perpetual Inventory
+ What it involves: counting everything all the time
+ What CORE currently offers
+ The essential `vendorItems` `<=>` `products` link is [described in detail] (https://github.com/CORE-POS/IS4C/wiki/Products-and-Vendors)
+ `PurchaseOrder` and `PurchaseOrderItems`
+ `[Ordered]quantity`
+ `receivedQty`
+ For Inventory, the crucial number is `receivedQty`
+ Purchase Orders:
+ created
+ loaded
+ from external systems
+ from invoices/manifests
+ data capture issues
+ What remains to be developed
+ Reconciling purchase orders against invoices/manifests
+ Comparing purchase orders to movement
+ Receiving use cases
+ floor
+ backoffice
#### Shrink recording at Cash
+ `SK`

`trans_status` = `Z` for all items on the screen
+ Returns and shrink: `RF###`

`description` `RF: Spoiled` applies to next item, but for it `trans_status` = `R`, not `Z`
+ StoreCharge accounts for intra-co-op transfers
+ A Shrink member account
#### Backend Shrink Recording
+ Since the whitepaper was written tools have have been added with finer-grain notation of reasons:

This (`fannie/item/shrink/ShrinkTool.php`) adds `trans_status` = `Z` items.
+ There is also a tool for editing current-day shrink items.
+ Co-ops can have their own set of Reasons for shrink.
#### Monitoring margin
+ Reports
+ Movement Reports
+ Margin Movement
+ Monitoring loss
#### Other ways to achieve the objective:
+ Monitoring cost
+ Spot-checking received orders
#### What to implement next?
+ What is important to your co-op? | non_infrastructure | agenda idea for everything inventory with the recent march end of the fiscal year for many co ops inventory has been on our minds and maybe kept us up late the topic has also come up recently as an extension to the transactions and reports base of pos there are many facets to it and constraints and points of view on how it should be done inventory support is a work in progress in core andy theuninck s pdf is the best general starting point but today we can go into detail on some of the issues below or others that are on our minds the whitepaper begins inventory is a very simple concept how much product do we have on hand right this moment the devil is in the details as the number constantly fluctuates inventory can also be understood as a means to an end items on hand is merely one component of measuring whether or not a store is making margin overall or on an item or group of items so how do co ops approach this problem there are two somewhat distinct methods inventory count vs and perpetual inventory why one or the other or some of each tradeoff cost vs complete current knowledge inventory count tools shelf audit plugin the interface is formatted for use on a portable device with a handheld scanner other tools services used perpetual inventory what it involves counting everything all the time what core currently offers the essential vendoritems products link is purchaseorder and purchaseorderitems quantity receivedqty for inventory the crucial number is receivedqty purchase orders created loaded from external systems from invoices manifests data capture issues what remains to be developed reconciling purchase orders against invoices manifests comparing purchase orders to movement receiving use cases floor backoffice shrink recording at cash sk trans status z for all items on the screen returns and shrink rf description rf spoiled applies to next item but for it trans status r not z storecharge accounts for intra co op transfers a shrink member account backend shrink recording since the whitepaper was written tools have have been added with finer grain notation of reasons this fannie item shrink shrinktool php adds trans status z items there is also a tool for editing current day shrink items co ops can have their own set of reasons for shrink monitoring margin reports movement reports margin movement monitoring loss other ways to achieve the objective monitoring cost spot checking received orders what to implement next what is important to your co op | 0 |
60,984 | 6,720,156,507 | IssuesEvent | 2017-10-16 06:22:44 | tcpd/surf | https://api.github.com/repos/tcpd/surf | closed | Some Merges suggested by Gilles in Gujarat AE are not being covered in Compatible Names. | bug P1 - High Priority Tested. Waiting for Release | Compatible Names is not covering few of the cases in Gujarat AE.
Attaching the correction File.
[1967 Gujarat corrections.xlsx](https://github.com/tcpd/surf/files/1287904/1967.Gujarat.corrections.xlsx)
Eg:
AC_name Position Cand1
BABRA 1 J. D. KANSAGRA
BABRA 3 JINABHAI D KANASAGARA
BABRA 1 KANSAGRA JINA DEVRAJ
BARODA CITY 1 C.K. PARIKH
BARODA CITY 1 CHANDRAKANT M PARIKH | 1.0 | Some Merges suggested by Gilles in Gujarat AE are not being covered in Compatible Names. - Compatible Names is not covering few of the cases in Gujarat AE.
Attaching the correction File.
[1967 Gujarat corrections.xlsx](https://github.com/tcpd/surf/files/1287904/1967.Gujarat.corrections.xlsx)
Eg:
AC_name Position Cand1
BABRA 1 J. D. KANSAGRA
BABRA 3 JINABHAI D KANASAGARA
BABRA 1 KANSAGRA JINA DEVRAJ
BARODA CITY 1 C.K. PARIKH
BARODA CITY 1 CHANDRAKANT M PARIKH | non_infrastructure | some merges suggested by gilles in gujarat ae are not being covered in compatible names compatible names is not covering few of the cases in gujarat ae attaching the correction file eg ac name position babra j d kansagra babra jinabhai d kanasagara babra kansagra jina devraj baroda city c k parikh baroda city chandrakant m parikh | 0 |
30,831 | 25,105,840,283 | IssuesEvent | 2022-11-08 16:33:12 | ceydaduzgec/SWE573 | https://api.github.com/repos/ceydaduzgec/SWE573 | opened | Create tests | Component: Infrastructure Type: Test | - Install coverage.
- Run coverage to see missing parts.
- Write tests for views and models. | 1.0 | Create tests - - Install coverage.
- Run coverage to see missing parts.
- Write tests for views and models. | infrastructure | create tests install coverage run coverage to see missing parts write tests for views and models | 1 |
332,107 | 24,336,547,871 | IssuesEvent | 2022-10-01 06:10:18 | Dun-sin/Whisper | https://api.github.com/repos/Dun-sin/Whisper | closed | [DOCS] wrong link in README | documentation good first issue 🟩 priority: low EddieHub:good-first-issue Easy | ### Description
`View Demo` has a wrong link, replace it with the proper link to the site, which can be found in about section of this repo
### Screenshots
 | 1.0 | [DOCS] wrong link in README - ### Description
`View Demo` has a wrong link, replace it with the proper link to the site, which can be found in about section of this repo
### Screenshots
 | non_infrastructure | wrong link in readme description view demo has a wrong link replace it with the proper link to the site which can be found in about section of this repo screenshots | 0 |
725,175 | 24,953,552,403 | IssuesEvent | 2022-11-01 09:40:48 | metabase/metabase | https://api.github.com/repos/metabase/metabase | closed | Modal filtering doesn't work on joined table when Question has a filter after aggregation | Type:Bug Priority:P2 Querying/Parameters & Variables .Frontend Querying/GUI .Reproduced .Regression | **Describe the bug**
I am not able to select any option from the Modal filtering off a joined table within a Question that has a filter after aggregation.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to New Question -> Sample Database -> Orders -> Join with People -> Summarise Count by Created At: Month -> Filter Count Greater than 1

2. Visualize Result -> Click on Modal Filtering -> Try to Filter on the Joined Table -> Name is Tony -> Apply Filters

3. No filtering is applied
**Expected behavior**
If I have the option to Filter then I should be able to Filter
**Information about your Metabase Installation:**
Metabase 1.44.4
**Workaround**
Don't use the Modal Filtering but go back to the GUI Editor and setup required Filtering from there

| 1.0 | Modal filtering doesn't work on joined table when Question has a filter after aggregation - **Describe the bug**
I am not able to select any option from the Modal filtering off a joined table within a Question that has a filter after aggregation.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to New Question -> Sample Database -> Orders -> Join with People -> Summarise Count by Created At: Month -> Filter Count Greater than 1

2. Visualize Result -> Click on Modal Filtering -> Try to Filter on the Joined Table -> Name is Tony -> Apply Filters

3. No filtering is applied
**Expected behavior**
If I have the option to Filter then I should be able to Filter
**Information about your Metabase Installation:**
Metabase 1.44.4
**Workaround**
Don't use the Modal Filtering but go back to the GUI Editor and setup required Filtering from there

| non_infrastructure | modal filtering doesn t work on joined table when question has a filter after aggregation describe the bug i am not able to select any option from the modal filtering off a joined table within a question that has a filter after aggregation to reproduce steps to reproduce the behavior go to new question sample database orders join with people summarise count by created at month filter count greater than visualize result click on modal filtering try to filter on the joined table name is tony apply filters no filtering is applied expected behavior if i have the option to filter then i should be able to filter information about your metabase installation metabase workaround don t use the modal filtering but go back to the gui editor and setup required filtering from there | 0 |
176,653 | 14,593,159,331 | IssuesEvent | 2020-12-19 21:15:11 | DerwenAI/kglab | https://api.github.com/repos/DerwenAI/kglab | opened | Develop courseware for KGs | documentation | Develop courseware for `kglab` based on Manning [*liveProject*](https://www.manning.com/liveproject-catalog):
* reuse the example notebooks, adding better structure for *progressive exercises*
* learners progress toward a capstone project
* compete on a leaderboard to achieve coverage of semantic tagging for the full recipe dataset
* points for correctness, coverage, and performance
Questions:
* Do we need to approach [Food.com](https://www.food.com/) to partner in promotion? | 1.0 | Develop courseware for KGs - Develop courseware for `kglab` based on Manning [*liveProject*](https://www.manning.com/liveproject-catalog):
* reuse the example notebooks, adding better structure for *progressive exercises*
* learners progress toward a capstone project
* compete on a leaderboard to achieve coverage of semantic tagging for the full recipe dataset
* points for correctness, coverage, and performance
Questions:
* Do we need to approach [Food.com](https://www.food.com/) to partner in promotion? | non_infrastructure | develop courseware for kgs develop courseware for kglab based on manning reuse the example notebooks adding better structure for progressive exercises learners progress toward a capstone project compete on a leaderboard to achieve coverage of semantic tagging for the full recipe dataset points for correctness coverage and performance questions do we need to approach to partner in promotion | 0 |
33,690 | 27,719,417,179 | IssuesEvent | 2023-03-14 19:20:37 | sunpy/sunpy | https://api.github.com/repos/sunpy/sunpy | closed | Parallelise code to update cdaweb datasets | Feature Request Priority Low Infrastructure Performance | ### Describe the feature
I'm currently running the code that scrapes CDAWeb for currently supported datasets. The code can be found here: https://github.com/sunpy/sunpy/blob/8c0b9054dc25918ee5b6286a93edbbca09cde903/sunpy/net/cdaweb/helpers.py#L118
The code only fires off one request at a time, waiting for it to come back before putting in the next request. This means it takes a long time (>~5 mins) to scrape all the datasets. I think it *should* be possible to send multiple requests at once to speed this up.
### Proposed solution
_No response_ | 1.0 | Parallelise code to update cdaweb datasets - ### Describe the feature
I'm currently running the code that scrapes CDAWeb for currently supported datasets. The code can be found here: https://github.com/sunpy/sunpy/blob/8c0b9054dc25918ee5b6286a93edbbca09cde903/sunpy/net/cdaweb/helpers.py#L118
The code only fires off one request at a time, waiting for it to come back before putting in the next request. This means it takes a long time (>~5 mins) to scrape all the datasets. I think it *should* be possible to send multiple requests at once to speed this up.
### Proposed solution
_No response_ | infrastructure | parallelise code to update cdaweb datasets describe the feature i m currently running the code that scrapes cdaweb for currently supported datasets the code can be found here the code only fires off one request at a time waiting for it to come back before putting in the next request this means it takes a long time mins to scrape all the datasets i think it should be possible to send multiple requests at once to speed this up proposed solution no response | 1 |
6,515 | 6,492,337,993 | IssuesEvent | 2017-08-21 12:44:08 | djavorszky/ddn | https://api.github.com/repos/djavorszky/ddn | closed | Containerize DDN | enhancement help wanted infrastructure new feature question | As we discused with @djavorszky, it would be awesome, if we could create a DDN Docker container that includes all the necessary thing for running the DDN server+web.
Possible considerations:
- Use more lightweight DB for the backend (I suggest SQLite)
| 1.0 | Containerize DDN - As we discused with @djavorszky, it would be awesome, if we could create a DDN Docker container that includes all the necessary thing for running the DDN server+web.
Possible considerations:
- Use more lightweight DB for the backend (I suggest SQLite)
| infrastructure | containerize ddn as we discused with djavorszky it would be awesome if we could create a ddn docker container that includes all the necessary thing for running the ddn server web possible considerations use more lightweight db for the backend i suggest sqlite | 1 |
327,375 | 9,974,793,687 | IssuesEvent | 2019-07-09 11:32:18 | upframe/connect | https://api.github.com/repos/upframe/connect | reopened | Free slots are defaulting to mobile size when they shouldn't | bug priority | <img width="1244" alt="image" src="https://user-images.githubusercontent.com/6923650/60706625-29209000-9f0a-11e9-9b09-f548e06b5058.png">
As you can see on the screen the free slot containers are occupying the whole section for some unexplained reason. My screen size is 1337x1155. | 1.0 | Free slots are defaulting to mobile size when they shouldn't - <img width="1244" alt="image" src="https://user-images.githubusercontent.com/6923650/60706625-29209000-9f0a-11e9-9b09-f548e06b5058.png">
As you can see on the screen the free slot containers are occupying the whole section for some unexplained reason. My screen size is 1337x1155. | non_infrastructure | free slots are defaulting to mobile size when they shouldn t img width alt image src as you can see on the screen the free slot containers are occupying the whole section for some unexplained reason my screen size is | 0 |
22,627 | 15,338,893,239 | IssuesEvent | 2021-02-27 00:07:55 | tmahlburg/jugendstadtplan | https://api.github.com/repos/tmahlburg/jugendstadtplan | closed | Deployment verbessern | hosting infrastructure | - periodisches Erneuern des SSL-Zertifikats | ERLEDIGT
- periodisches Datenbank-Backup | 1.0 | Deployment verbessern - - periodisches Erneuern des SSL-Zertifikats | ERLEDIGT
- periodisches Datenbank-Backup | infrastructure | deployment verbessern periodisches erneuern des ssl zertifikats erledigt periodisches datenbank backup | 1 |
20,744 | 14,138,666,722 | IssuesEvent | 2020-11-10 08:46:28 | fluencelabs/fluence | https://api.github.com/repos/fluencelabs/fluence | closed | Cumulative health for all containers on the Node | critical ~infrastructure | It should be possible to detect if any of the containers supervised by Node (i.e., worker and tendermint) isn't running or healthy.
This would enable us to monitor stability of all apps running on the devnet. Currently it's only the basic ones (tictactoe and such).
Related #671 | 1.0 | Cumulative health for all containers on the Node - It should be possible to detect if any of the containers supervised by Node (i.e., worker and tendermint) isn't running or healthy.
This would enable us to monitor stability of all apps running on the devnet. Currently it's only the basic ones (tictactoe and such).
Related #671 | infrastructure | cumulative health for all containers on the node it should be possible to detect if any of the containers supervised by node i e worker and tendermint isn t running or healthy this would enable us to monitor stability of all apps running on the devnet currently it s only the basic ones tictactoe and such related | 1 |
35,049 | 30,718,843,207 | IssuesEvent | 2023-07-27 14:38:49 | ProjectPythia/intake-cookbook | https://api.github.com/repos/ProjectPythia/intake-cookbook | opened | Link checker is failing on NOAA homepage | infrastructure | Recently the nightly link check started failing with
```
403 Client Error: Forbidden for url: https://www.noaa.gov/
```
This *looks to me* like the same issue we encountered in Foundations with certain websites rejecting the sphinx link-checker, see here:
https://github.com/ProjectPythia/pythia-foundations/issues/260
https://github.com/sphinx-doc/sphinx/issues/10343
https://github.com/ProjectPythia/pythia-foundations/pull/262
Two possible workarounds:
- skip the link check for this site only, by adding "https://www.noaa.gov" to the `linkcheck_ignore:` line in the config file
- Run a valid link check for this site by "spoofing" a user agent with `linkcheck_request_headers:` as we did in Foundations https://github.com/ProjectPythia/pythia-foundations/pull/262
| 1.0 | Link checker is failing on NOAA homepage - Recently the nightly link check started failing with
```
403 Client Error: Forbidden for url: https://www.noaa.gov/
```
This *looks to me* like the same issue we encountered in Foundations with certain websites rejecting the sphinx link-checker, see here:
https://github.com/ProjectPythia/pythia-foundations/issues/260
https://github.com/sphinx-doc/sphinx/issues/10343
https://github.com/ProjectPythia/pythia-foundations/pull/262
Two possible workarounds:
- skip the link check for this site only, by adding "https://www.noaa.gov" to the `linkcheck_ignore:` line in the config file
- Run a valid link check for this site by "spoofing" a user agent with `linkcheck_request_headers:` as we did in Foundations https://github.com/ProjectPythia/pythia-foundations/pull/262
| infrastructure | link checker is failing on noaa homepage recently the nightly link check started failing with client error forbidden for url this looks to me like the same issue we encountered in foundations with certain websites rejecting the sphinx link checker see here two possible workarounds skip the link check for this site only by adding to the linkcheck ignore line in the config file run a valid link check for this site by spoofing a user agent with linkcheck request headers as we did in foundations | 1 |
24,945 | 17,934,087,887 | IssuesEvent | 2021-09-10 13:17:41 | google/web-stories-wp | https://api.github.com/repos/google/web-stories-wp | opened | Add e2e tests for archive page customization | P2 Type: Infrastructure Needs Tests Pod: WP & Infra Package: E2E Tests | <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Feature Description
<!-- A clear and concise description of what the problem is and what you want to happen. -->
It would be nice to have some e2e test coverage for #8811 and #8812, since this new feature will affect many areas in the plugin throughout whole WordPress admin and of course also the frontend.
* [ ] Test that disabling the archive completely hides the archive link option for blocks/widgets/classic editor
* [ ] Test that choosing a custom page for the archive will use that page as the archive whenever linked to it
## Alternatives Considered
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional Context
<!-- Add any other context or screenshots about the feature request. -->
| 1.0 | Add e2e tests for archive page customization - <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Feature Description
<!-- A clear and concise description of what the problem is and what you want to happen. -->
It would be nice to have some e2e test coverage for #8811 and #8812, since this new feature will affect many areas in the plugin throughout whole WordPress admin and of course also the frontend.
* [ ] Test that disabling the archive completely hides the archive link option for blocks/widgets/classic editor
* [ ] Test that choosing a custom page for the archive will use that page as the archive whenever linked to it
## Alternatives Considered
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional Context
<!-- Add any other context or screenshots about the feature request. -->
| infrastructure | add tests for archive page customization feature description it would be nice to have some test coverage for and since this new feature will affect many areas in the plugin throughout whole wordpress admin and of course also the frontend test that disabling the archive completely hides the archive link option for blocks widgets classic editor test that choosing a custom page for the archive will use that page as the archive whenever linked to it alternatives considered additional context | 1 |
31,626 | 25,946,119,082 | IssuesEvent | 2022-12-17 01:26:11 | bootstrapworld/curriculum | https://api.github.com/repos/bootstrapworld/curriculum | closed | Data cycle formatting is wacky for measures of center | Infrastructure | Text that should be beneath the blank lines and text that should land on blank lines is overlapping.
Fall2023 branch
`measures-of-center/data-cycle-practice.adoc`

| 1.0 | Data cycle formatting is wacky for measures of center - Text that should be beneath the blank lines and text that should land on blank lines is overlapping.
Fall2023 branch
`measures-of-center/data-cycle-practice.adoc`

| infrastructure | data cycle formatting is wacky for measures of center text that should be beneath the blank lines and text that should land on blank lines is overlapping branch measures of center data cycle practice adoc | 1 |
35,168 | 30,812,964,951 | IssuesEvent | 2023-08-01 11:42:57 | grafana/agent | https://api.github.com/repos/grafana/agent | reopened | Cannot set the allowlist field in node_exporter's diskstat collector | bug type/infrastructure | The logic around the include/exclude fields for the diskstat collector in the node_exporter integration, that is also inherited by the prometheus.exporter.unix component was that a) the `device_exclude` field had a default value and b) the `device_exclude` and `device_include` fields were mutually exclusive.
This meant that the first field was always being set implicitly, and if the second was defined then node_exporter could not start. The following two config files failed to start the Agent, and there is no way to make use of `device_include`.
```yaml
integrations:
node_exporter:
enabled: true
scrape_integration: true
set_collectors:
- "diskstats"
diskstats_device_include: "/dev/vd.*"
```
```river
prometheus.exporter.unix {
disk {
device_include = "foo"
}
}
```
| 1.0 | Cannot set the allowlist field in node_exporter's diskstat collector - The logic around the include/exclude fields for the diskstat collector in the node_exporter integration, that is also inherited by the prometheus.exporter.unix component was that a) the `device_exclude` field had a default value and b) the `device_exclude` and `device_include` fields were mutually exclusive.
This meant that the first field was always being set implicitly, and if the second was defined then node_exporter could not start. The following two config files failed to start the Agent, and there is no way to make use of `device_include`.
```yaml
integrations:
node_exporter:
enabled: true
scrape_integration: true
set_collectors:
- "diskstats"
diskstats_device_include: "/dev/vd.*"
```
```river
prometheus.exporter.unix {
disk {
device_include = "foo"
}
}
```
| infrastructure | cannot set the allowlist field in node exporter s diskstat collector the logic around the include exclude fields for the diskstat collector in the node exporter integration that is also inherited by the prometheus exporter unix component was that a the device exclude field had a default value and b the device exclude and device include fields were mutually exclusive this meant that the first field was always being set implicitly and if the second was defined then node exporter could not start the following two config files failed to start the agent and there is no way to make use of device include yaml integrations node exporter enabled true scrape integration true set collectors diskstats diskstats device include dev vd river prometheus exporter unix disk device include foo | 1 |
8,680 | 7,558,797,665 | IssuesEvent | 2018-04-20 00:04:51 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | closed | azure-cli client issues a duplicate command after 2 minutes | Infrastructure Question | I'm seeing in some cases that azure-cli client duplicates already issued command after 2 minutes.
I'm attaching deployment-related HTTP requests (5 fields per request) for the resource group.
The first and the last requests are identical:
PreciseTimeStamp
operationName
httpStatusCode
targetUri
userAgent
## 1 ##
2017-12-05 06:48:42.7163123
PUT/SUBSCRIPTIONS/RESOURCEGROUPS/PROVIDERS/MICROSOFT.RESOURCES/DEPLOYMENTS/
201
https://management.azure.com/subscriptions/<sub>/resourcegroups/<rg>/providers/Microsoft.Resources/deployments/azurecli1512456518.6712350?api-version=2017-05-10
python/2.7.12 (Linux-4.11.0-1015-azure-x86_64-with-Ubuntu-16.04-xenial) requests/2.18.4 msrest/0.4.19 msrest_azure/0.4.17 resourcemanagementclient/1.2.1 Azure-SDK-For-Python AZURECLI/2.0.21
## 2 ##
2017-12-05 06:48:52.8234114
POST/SUBSCRIPTIONS/RESOURCEGROUPS/PROVIDERS/MICROSOFT.RESOURCES/DEPLOYMENTS/VALIDATE
200
https://management.azure.com/subscriptions/<sub>/resourcegroups/<rg>/providers/Microsoft.Resources/deployments/195d0e90-fed2-4d83-8684-f309bec5e0b0/validate?api-version=2016-09-01
Microsoft.Azure.Management.ResourceManager.ResourceManagementClient/1.0.0.0
## 3 ##
2017-12-05 06:48:56.1942056
POST/SUBSCRIPTIONS/RESOURCEGROUPS/PROVIDERS/MICROSOFT.RESOURCES/DEPLOYMENTS/VALIDATE
200
https://management.azure.com/subscriptions/<sub>/resourcegroups/<rg>/providers/Microsoft.Resources/deployments/b87da223-eeca-488b-9267-b6cbc5cb42b9/validate?api-version=2016-09-01
Microsoft.Azure.Management.ResourceManager.ResourceManagementClient/1.0.0.0
## 4 ##
2017-12-05 06:48:59.4633091
PUT/SUBSCRIPTIONS/RESOURCEGROUPS/DEPLOYMENTS/
201
https://management.azure.com/subscriptions/<sub>/resourcegroups/<rg>/deployments/b87da223-eeca-488b-9267-b6cbc5cb42b9?api-version=2014-04-01-preview
Microsoft.Azure.Management.Resources.ResourceManagementClient/2.0.0.0
## 5 ##
2017-12-05 06:50:21.5758500
PUT/SUBSCRIPTIONS/RESOURCEGROUPS/PROVIDERS/MICROSOFT.RESOURCES/DEPLOYMENTS/
409
https://management.azure.com/subscriptions/<sub>/resourcegroups/<rg>/providers/Microsoft.Resources/deployments/azurecli1512456518.6712350?api-version=2017-05-10
python/2.7.12 (Linux-4.11.0-1015-azure-x86_64-with-Ubuntu-16.04-xenial) requests/2.18.4 msrest/0.4.19 msrest_azure/0.4.17 resourcemanagementclient/1.2.1 Azure-SDK-For-Python AZURECLI/2.0.21
| 1.0 | azure-cli client issues a duplicate command after 2 minutes - I'm seeing in some cases that azure-cli client duplicates already issued command after 2 minutes.
I'm attaching deployment-related HTTP requests (5 fields per request) for the resource group.
The first and the last requests are identical:
PreciseTimeStamp
operationName
httpStatusCode
targetUri
userAgent
## 1 ##
2017-12-05 06:48:42.7163123
PUT/SUBSCRIPTIONS/RESOURCEGROUPS/PROVIDERS/MICROSOFT.RESOURCES/DEPLOYMENTS/
201
https://management.azure.com/subscriptions/<sub>/resourcegroups/<rg>/providers/Microsoft.Resources/deployments/azurecli1512456518.6712350?api-version=2017-05-10
python/2.7.12 (Linux-4.11.0-1015-azure-x86_64-with-Ubuntu-16.04-xenial) requests/2.18.4 msrest/0.4.19 msrest_azure/0.4.17 resourcemanagementclient/1.2.1 Azure-SDK-For-Python AZURECLI/2.0.21
## 2 ##
2017-12-05 06:48:52.8234114
POST/SUBSCRIPTIONS/RESOURCEGROUPS/PROVIDERS/MICROSOFT.RESOURCES/DEPLOYMENTS/VALIDATE
200
https://management.azure.com/subscriptions/<sub>/resourcegroups/<rg>/providers/Microsoft.Resources/deployments/195d0e90-fed2-4d83-8684-f309bec5e0b0/validate?api-version=2016-09-01
Microsoft.Azure.Management.ResourceManager.ResourceManagementClient/1.0.0.0
## 3 ##
2017-12-05 06:48:56.1942056
POST/SUBSCRIPTIONS/RESOURCEGROUPS/PROVIDERS/MICROSOFT.RESOURCES/DEPLOYMENTS/VALIDATE
200
https://management.azure.com/subscriptions/<sub>/resourcegroups/<rg>/providers/Microsoft.Resources/deployments/b87da223-eeca-488b-9267-b6cbc5cb42b9/validate?api-version=2016-09-01
Microsoft.Azure.Management.ResourceManager.ResourceManagementClient/1.0.0.0
## 4 ##
2017-12-05 06:48:59.4633091
PUT/SUBSCRIPTIONS/RESOURCEGROUPS/DEPLOYMENTS/
201
https://management.azure.com/subscriptions/<sub>/resourcegroups/<rg>/deployments/b87da223-eeca-488b-9267-b6cbc5cb42b9?api-version=2014-04-01-preview
Microsoft.Azure.Management.Resources.ResourceManagementClient/2.0.0.0
## 5 ##
2017-12-05 06:50:21.5758500
PUT/SUBSCRIPTIONS/RESOURCEGROUPS/PROVIDERS/MICROSOFT.RESOURCES/DEPLOYMENTS/
409
https://management.azure.com/subscriptions/<sub>/resourcegroups/<rg>/providers/Microsoft.Resources/deployments/azurecli1512456518.6712350?api-version=2017-05-10
python/2.7.12 (Linux-4.11.0-1015-azure-x86_64-with-Ubuntu-16.04-xenial) requests/2.18.4 msrest/0.4.19 msrest_azure/0.4.17 resourcemanagementclient/1.2.1 Azure-SDK-For-Python AZURECLI/2.0.21
| infrastructure | azure cli client issues a duplicate command after minutes i m seeing in some cases that azure cli client duplicates already issued command after minutes i m attaching deployment related http requests fields per request for the resource group the first and the last requests are identical precisetimestamp operationname httpstatuscode targeturi useragent put subscriptions resourcegroups providers microsoft resources deployments python linux azure with ubuntu xenial requests msrest msrest azure resourcemanagementclient azure sdk for python azurecli post subscriptions resourcegroups providers microsoft resources deployments validate microsoft azure management resourcemanager resourcemanagementclient post subscriptions resourcegroups providers microsoft resources deployments validate microsoft azure management resourcemanager resourcemanagementclient put subscriptions resourcegroups deployments microsoft azure management resources resourcemanagementclient put subscriptions resourcegroups providers microsoft resources deployments python linux azure with ubuntu xenial requests msrest msrest azure resourcemanagementclient azure sdk for python azurecli | 1 |
241,310 | 20,115,244,173 | IssuesEvent | 2022-02-07 18:48:12 | mennaelkashef/eShop | https://api.github.com/repos/mennaelkashef/eShop | opened | No description entered by the user. | Hello! RULE-GOT-APPLIED DOES-NOT-CONTAIN-STRING Rule-works-on-convert-to-bug test instabug | # :clipboard: Bug Details
>No description entered by the user.
key | value
--|--
Reported At | 2022-02-07 18:47:07 UTC
Email | imohamady@instabug.com
Categories | Report a bug
Tags | test, Hello!, RULE-GOT-APPLIED, DOES-NOT-CONTAIN-STRING, Rule-works-on-convert-to-bug, instabug
App Version | 1.1 (1)
Session Duration | 7
Device | Google AOSP on IA Emulator, OS Level 28
Display | 1080x2160 (xhdpi)
Location | Giza, Egypt (en)
## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8393?utm_source=github&utm_medium=integrations) :point_left:
___
# :iphone: View Hierarchy
This bug was reported from **com.example.app.main.MainFragment**
Find its interactive view hierarchy with all its subviews here: :point_right: **[Check View Hierarchy](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8393?show-hierarchy-view=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :chart_with_downwards_trend: Session Profiler
Here is what the app was doing right before the bug was reported:
Key | Value
--|--
Used Memory | 53.5% - 0.78/1.46 GB
Used Storage | 4.0% - 0.23/5.81 GB
Connectivity | LTE - Android
Battery | 100% - unplugged
Orientation | portrait
Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8393?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :bust_in_silhouette: User Info
### User Attributes
```
key_name -2055613585: key value bla bla bla la
key_name 101701368: key value bla bla bla la
key_name 1348114218: key value bla bla bla la
```
___
# :mag_right: Logs
### User Steps
Here are the last 10 steps done by the user right before the bug was reported:
```
18:47:04 In activity com.example.app.main.MainActivity: fragment com.example.app.core.CoreFragment was detached.
18:47:04 In activity com.example.app.main.MainActivity: fragment com.example.app.main.MainFragment was resumed.
18:47:05 Tap in "Sdk version: 10..." of type "androidx.appcompat.widget.AppCompatTextView" in "com.example.app.main.MainActivity"
18:47:05 com.example.app.main.MainActivity was paused.
18:47:05 In activity com.example.app.main.MainActivity: fragment com.example.app.main.MainFragment was paused.
18:47:05 In activity com.example.app.main.MainActivity: fragment com.example.app.main.MainFragment was resumed.
18:47:05 com.example.app.main.MainActivity was resumed.
18:47:05 com.example.app.main.MainActivity was paused.
18:47:05 In activity com.example.app.main.MainActivity: fragment com.example.app.main.MainFragment was paused.
18:47:07 Tap in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.app.main.MainActivity"
```
Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8393?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left:
### Console Log
Here are the last 10 console logs logged right before the bug was reported:
```
18:47:11 D/EGL_emulation( 6912): eglMakeCurrent: 0xe1587700: ver 3 0 (tinfo 0xc9fff0a0)
18:47:11 I/chatty ( 6912): uid=10086(com.example.app) RenderThread identical 8 lines
18:47:11 D/EGL_emulation( 6912): eglMakeCurrent: 0xe1587700: ver 3 0 (tinfo 0xc9fff0a0)
18:47:11 D/IB-InstabugFeaturesManager( 6912): start saving app_features
18:47:11 D/IB-IBActivityLifecycleListener( 6912): MainActivity SaveInstanceState
18:47:11 D/IB-InstabugFeaturesManager( 6912): finish saving app_features
18:47:12 D/LeakCanary( 6912): Setting up flushing for Thread[PixelCopy,5,main]
18:47:12 D/IB-BaseReportingPresenter( 6912): checkUserEmailValid :non-empty-email
18:47:12 D/IB-ActionsOrchestrator( 6912): runAction
18:47:13 D/IB-AttachmentsUtility( 6912): encryptAttachments
```
Find all the logged console logs throughout the session here: :point_right: **[View All Console Log](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8393?show-logs=console_log&utm_source=github&utm_medium=integrations)** :point_left:
___
# :camera: Images
[](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/17819549/0ea2129d1633fe5e076d7704e6cd3fa3_original/24944167/bug_1644259625525_.jpg?Expires=4799933291&Signature=Ue6ruioum6FgWGVxaYBjKayahJFtM0npI7gtuhNlCcYFHuAnuCvN-PiNYi3jWkH1X52G4wKbVz6a7Ly4mw9LQ0n~WmbatZur9kw-GnLl1Qt1umKuBz4N2bu6Rxq8lPHf3NWEQ3OhMAjQTkNfaYm9oU1FcQ6gI0T61GsDaxtELQrwaN3cogR8onQ5FfUCeoTF7O2XfXwnMYmwxPGIhek2RbEJBPwSZJ~UZYK5VCfhwBBxRMszeTfSQplmwM3V6U6VUxC6wUW9BoyIESFyuCbgNWonw85pGCSeCQa1gXH5aFEtXn4CneTt8p4AjKC2FDtKP207iZVv~rIiUyVMpvPqdw__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)
[](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/17819549/7ff0e87c6278b1846e5a0ed9086f8c9b_original/24944171/bug_1644259630806_.jpg?Expires=4799933291&Signature=lb-eS2NVZw5Lp7l-vWCNJpRhA-eqUAP9wIRhZEXcoevEJqJA1zsbRgDi6AvJyBASeLCIjP736Gxnl67nw0qIqx9NEQHYqgiTYI6-8ZJTZjQWSRfg0DaDdBMGU2tmw0w0rJrBx0-uPtOua0euxocPQUE7oG2523KuZqOS95TQdo4iHCxZ35T4svCPKn-Npp5aKnG8KTvz0A9a1P46aBA6W0tDwtCKXGuLVuhst~WKgIIwonHganP600N4-q2FkmOVArmXZvEU~9vojnVZvCKhKoTn8SVLPcwTrNlVYs8IlvPjG01j9yfSPwRGjnHCh4Zk8Uo9JiUnngKs2FaswdMGjA__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)
___
# :warning: Looking for More Details?
1. **Network Log**: we are unable to capture your network requests automatically. If you are using HttpUrlConnection or Okhttp requests, [**check the details mentioned here**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations#section-network-logs).
2. **User Events**: start capturing custom User Events to send them along with each report. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations).
3. **Instabug Log**: start adding Instabug logs to see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations). | 1.0 | No description entered by the user. - # :clipboard: Bug Details
>No description entered by the user.
key | value
--|--
Reported At | 2022-02-07 18:47:07 UTC
Email | imohamady@instabug.com
Categories | Report a bug
Tags | test, Hello!, RULE-GOT-APPLIED, DOES-NOT-CONTAIN-STRING, Rule-works-on-convert-to-bug, instabug
App Version | 1.1 (1)
Session Duration | 7
Device | Google AOSP on IA Emulator, OS Level 28
Display | 1080x2160 (xhdpi)
Location | Giza, Egypt (en)
## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8393?utm_source=github&utm_medium=integrations) :point_left:
___
# :iphone: View Hierarchy
This bug was reported from **com.example.app.main.MainFragment**
Find its interactive view hierarchy with all its subviews here: :point_right: **[Check View Hierarchy](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8393?show-hierarchy-view=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :chart_with_downwards_trend: Session Profiler
Here is what the app was doing right before the bug was reported:
Key | Value
--|--
Used Memory | 53.5% - 0.78/1.46 GB
Used Storage | 4.0% - 0.23/5.81 GB
Connectivity | LTE - Android
Battery | 100% - unplugged
Orientation | portrait
Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8393?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :bust_in_silhouette: User Info
### User Attributes
```
key_name -2055613585: key value bla bla bla la
key_name 101701368: key value bla bla bla la
key_name 1348114218: key value bla bla bla la
```
___
# :mag_right: Logs
### User Steps
Here are the last 10 steps done by the user right before the bug was reported:
```
18:47:04 In activity com.example.app.main.MainActivity: fragment com.example.app.core.CoreFragment was detached.
18:47:04 In activity com.example.app.main.MainActivity: fragment com.example.app.main.MainFragment was resumed.
18:47:05 Tap in "Sdk version: 10..." of type "androidx.appcompat.widget.AppCompatTextView" in "com.example.app.main.MainActivity"
18:47:05 com.example.app.main.MainActivity was paused.
18:47:05 In activity com.example.app.main.MainActivity: fragment com.example.app.main.MainFragment was paused.
18:47:05 In activity com.example.app.main.MainActivity: fragment com.example.app.main.MainFragment was resumed.
18:47:05 com.example.app.main.MainActivity was resumed.
18:47:05 com.example.app.main.MainActivity was paused.
18:47:05 In activity com.example.app.main.MainActivity: fragment com.example.app.main.MainFragment was paused.
18:47:07 Tap in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.app.main.MainActivity"
```
Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8393?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left:
### Console Log
Here are the last 10 console logs logged right before the bug was reported:
```
18:47:11 D/EGL_emulation( 6912): eglMakeCurrent: 0xe1587700: ver 3 0 (tinfo 0xc9fff0a0)
18:47:11 I/chatty ( 6912): uid=10086(com.example.app) RenderThread identical 8 lines
18:47:11 D/EGL_emulation( 6912): eglMakeCurrent: 0xe1587700: ver 3 0 (tinfo 0xc9fff0a0)
18:47:11 D/IB-InstabugFeaturesManager( 6912): start saving app_features
18:47:11 D/IB-IBActivityLifecycleListener( 6912): MainActivity SaveInstanceState
18:47:11 D/IB-InstabugFeaturesManager( 6912): finish saving app_features
18:47:12 D/LeakCanary( 6912): Setting up flushing for Thread[PixelCopy,5,main]
18:47:12 D/IB-BaseReportingPresenter( 6912): checkUserEmailValid :non-empty-email
18:47:12 D/IB-ActionsOrchestrator( 6912): runAction
18:47:13 D/IB-AttachmentsUtility( 6912): encryptAttachments
```
Find all the logged console logs throughout the session here: :point_right: **[View All Console Log](https://dashboard.instabug.com/applications/android-sample/beta/bugs/8393?show-logs=console_log&utm_source=github&utm_medium=integrations)** :point_left:
___
# :camera: Images
[](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/17819549/0ea2129d1633fe5e076d7704e6cd3fa3_original/24944167/bug_1644259625525_.jpg?Expires=4799933291&Signature=Ue6ruioum6FgWGVxaYBjKayahJFtM0npI7gtuhNlCcYFHuAnuCvN-PiNYi3jWkH1X52G4wKbVz6a7Ly4mw9LQ0n~WmbatZur9kw-GnLl1Qt1umKuBz4N2bu6Rxq8lPHf3NWEQ3OhMAjQTkNfaYm9oU1FcQ6gI0T61GsDaxtELQrwaN3cogR8onQ5FfUCeoTF7O2XfXwnMYmwxPGIhek2RbEJBPwSZJ~UZYK5VCfhwBBxRMszeTfSQplmwM3V6U6VUxC6wUW9BoyIESFyuCbgNWonw85pGCSeCQa1gXH5aFEtXn4CneTt8p4AjKC2FDtKP207iZVv~rIiUyVMpvPqdw__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)
[](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/17819549/7ff0e87c6278b1846e5a0ed9086f8c9b_original/24944171/bug_1644259630806_.jpg?Expires=4799933291&Signature=lb-eS2NVZw5Lp7l-vWCNJpRhA-eqUAP9wIRhZEXcoevEJqJA1zsbRgDi6AvJyBASeLCIjP736Gxnl67nw0qIqx9NEQHYqgiTYI6-8ZJTZjQWSRfg0DaDdBMGU2tmw0w0rJrBx0-uPtOua0euxocPQUE7oG2523KuZqOS95TQdo4iHCxZ35T4svCPKn-Npp5aKnG8KTvz0A9a1P46aBA6W0tDwtCKXGuLVuhst~WKgIIwonHganP600N4-q2FkmOVArmXZvEU~9vojnVZvCKhKoTn8SVLPcwTrNlVYs8IlvPjG01j9yfSPwRGjnHCh4Zk8Uo9JiUnngKs2FaswdMGjA__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)
___
# :warning: Looking for More Details?
1. **Network Log**: we are unable to capture your network requests automatically. If you are using HttpUrlConnection or Okhttp requests, [**check the details mentioned here**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations#section-network-logs).
2. **User Events**: start capturing custom User Events to send them along with each report. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations).
3. **Instabug Log**: start adding Instabug logs to see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/android-logging?utm_source=github&utm_medium=integrations). | non_infrastructure | no description entered by the user clipboard bug details no description entered by the user key value reported at utc email imohamady instabug com categories report a bug tags test hello rule got applied does not contain string rule works on convert to bug instabug app version session duration device google aosp on ia emulator os level display xhdpi location giza egypt en point right point left iphone view hierarchy this bug was reported from com example app main mainfragment find its interactive view hierarchy with all its subviews here point right point left chart with downwards trend session profiler here is what the app was doing right before the bug was reported key value used memory gb used storage gb connectivity lte android battery unplugged orientation portrait find all the changes that happened in the parameters mentioned above during the last seconds before the bug was reported here point right point left bust in silhouette user info user attributes key name key value bla bla bla la key name key value bla bla bla la key name key value bla bla bla la mag right logs user steps here are the last steps done by the user right before the bug was reported in activity com example app main mainactivity fragment com example app core corefragment was detached in activity com example app main mainactivity fragment com example app main mainfragment was resumed tap in sdk version of type androidx appcompat widget appcompattextview in com example app main mainactivity com example app main mainactivity was paused in activity com example app main mainactivity fragment com example app main mainfragment was paused in activity com example app main mainactivity fragment com example app main mainfragment was resumed com example app main mainactivity was resumed com example app main mainactivity was paused in activity com example app main mainactivity fragment com example app main mainfragment was paused tap in androidx constraintlayout widget constraintlayout in com example app main mainactivity find all the user steps done by the user throughout the session here point right point left console log here are the last console logs logged right before the bug was reported d egl emulation eglmakecurrent ver tinfo i chatty uid com example app renderthread identical lines d egl emulation eglmakecurrent ver tinfo d ib instabugfeaturesmanager start saving app features d ib ibactivitylifecyclelistener mainactivity saveinstancestate d ib instabugfeaturesmanager finish saving app features d leakcanary setting up flushing for thread d ib basereportingpresenter checkuseremailvalid non empty email d ib actionsorchestrator runaction d ib attachmentsutility encryptattachments find all the logged console logs throughout the session here point right point left camera images warning looking for more details network log we are unable to capture your network requests automatically if you are using httpurlconnection or okhttp requests user events start capturing custom user events to send them along with each report instabug log start adding instabug logs to see them right inside each report you receive | 0 |
181,199 | 14,008,325,081 | IssuesEvent | 2020-10-28 23:24:14 | rancher/dashboard | https://api.github.com/repos/rancher/dashboard | closed | info bubbles don't work | [zube]: To Test area/general | deployments info bubbles do not work:

service example of working:

Deployments has the preferred placement of the `i` by the label, but they do not work | 1.0 | info bubbles don't work - deployments info bubbles do not work:

service example of working:

Deployments has the preferred placement of the `i` by the label, but they do not work | non_infrastructure | info bubbles don t work deployments info bubbles do not work service example of working deployments has the preferred placement of the i by the label but they do not work | 0 |
31,315 | 25,547,401,431 | IssuesEvent | 2022-11-29 20:05:30 | CDCgov/data-exchange-hl7 | https://api.github.com/repos/CDCgov/data-exchange-hl7 | closed | Resilience/Redundant script for virtual machine setup | infrastructure | The goal of this ticket is to properly look into a scalable setup of github self-hosted runner within virtual machine(s).
Right now, I manually installed everything into the VM and it works!
This setup is fragile, and should at a minimum be scripted and documented.
- [ ] provide installation scripts or an image to replace the VM should this VM ever go down.
- [ ] provide documentation on the VM setup (portions of it already documented). | 1.0 | Resilience/Redundant script for virtual machine setup - The goal of this ticket is to properly look into a scalable setup of github self-hosted runner within virtual machine(s).
Right now, I manually installed everything into the VM and it works!
This setup is fragile, and should at a minimum be scripted and documented.
- [ ] provide installation scripts or an image to replace the VM should this VM ever go down.
- [ ] provide documentation on the VM setup (portions of it already documented). | infrastructure | resilience redundant script for virtual machine setup the goal of this ticket is to properly look into a scalable setup of github self hosted runner within virtual machine s right now i manually installed everything into the vm and it works this setup is fragile and should at a minimum be scripted and documented provide installation scripts or an image to replace the vm should this vm ever go down provide documentation on the vm setup portions of it already documented | 1 |
362,051 | 10,722,391,514 | IssuesEvent | 2019-10-27 11:41:01 | webiny/webiny-js | https://api.github.com/repos/webiny/webiny-js | closed | CMS editor history actions. | priority: low tag: bug fix | Sometimes when undoing/redoing, elements get messed up especially when manipulating row columns. A possible cause is the fact that when manipulating columns multiple actions are dispatched during a single "manipulation", thus making history middleware create states that are incomplete.
Things to inspect:
- see which element plugins dispatch multiple actions during one transaction
- **try refactoring code to make row/column elements more dependent on each other but handle the entire logic in the Row plugin to avoid dispatching multiple actions.**
- ~if all else fails, try adding a `commit changes` action that will trigger `history` middleware (maybe this is the best way to begin with?)~ (this doesn't work)
## Implementation
Delete/Drop logic must work with a single action dispatch, meaning, once an element is dropped or deleted - plugins must modify the `parent` element and return the new data which will then be set using `update element` action at once. That way we avoid multiple history triggers. | 1.0 | CMS editor history actions. - Sometimes when undoing/redoing, elements get messed up especially when manipulating row columns. A possible cause is the fact that when manipulating columns multiple actions are dispatched during a single "manipulation", thus making history middleware create states that are incomplete.
Things to inspect:
- see which element plugins dispatch multiple actions during one transaction
- **try refactoring code to make row/column elements more dependent on each other but handle the entire logic in the Row plugin to avoid dispatching multiple actions.**
- ~if all else fails, try adding a `commit changes` action that will trigger `history` middleware (maybe this is the best way to begin with?)~ (this doesn't work)
## Implementation
Delete/Drop logic must work with a single action dispatch, meaning, once an element is dropped or deleted - plugins must modify the `parent` element and return the new data which will then be set using `update element` action at once. That way we avoid multiple history triggers. | non_infrastructure | cms editor history actions sometimes when undoing redoing elements get messed up especially when manipulating row columns a possible cause is the fact that when manipulating columns multiple actions are dispatched during a single manipulation thus making history middleware create states that are incomplete things to inspect see which element plugins dispatch multiple actions during one transaction try refactoring code to make row column elements more dependent on each other but handle the entire logic in the row plugin to avoid dispatching multiple actions if all else fails try adding a commit changes action that will trigger history middleware maybe this is the best way to begin with this doesn t work implementation delete drop logic must work with a single action dispatch meaning once an element is dropped or deleted plugins must modify the parent element and return the new data which will then be set using update element action at once that way we avoid multiple history triggers | 0 |
53,875 | 13,220,211,007 | IssuesEvent | 2020-08-17 11:59:42 | zalando/restful-api-guidelines | https://api.github.com/repos/zalando/restful-api-guidelines | closed | Travis build is broken | bug build | The last master build one month ago did still work, now the [same build is broken](https://travis-ci.org/github/zalando/restful-api-guidelines/builds/703500390).
```
docker run -v $TRAVIS_BUILD_DIR:/documents/ --name asciidoc-to-pdf asciidoctor/docker-asciidoctor asciidoctor-pdf -D /documents/output index.adoc
no implicit conversion of nil into String
Use --trace for backtrace
```
Running with `--trace` we can find the root cause to be:
```
docker run -v $TRAVIS_BUILD_DIR:/documents/ --name asciidoc-to-pdf asciidoctor/docker-asciidoctor asciidoctor-pdf --trace -D /documents/output index.adoc
/usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:108:in `basename': no implicit conversion of nil into String (TypeError)
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:108:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:75:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:84:in `block in content'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:84:in `map'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:84:in `content'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:454:in `convert_section'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:106:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:75:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:84:in `block in content'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:84:in `map'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:84:in `content'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/document.rb:1020:in `content'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:359:in `add_chapter'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:250:in `convert_document'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:106:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/document.rb:951:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/convert.rb:118:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/convert.rb:183:in `block in convert_file'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/convert.rb:183:in `open'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/convert.rb:183:in `convert_file'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/cli/invoker.rb:128:in `block in invoke!'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/cli/invoker.rb:111:in `each'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/cli/invoker.rb:111:in `invoke!'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/bin/asciidoctor-epub3:28:in `<top (required)>'
from /usr/bin/asciidoctor-epub3:23:in `load'
from /usr/bin/asciidoctor-epub3:23:in `<main>'
```
This also is blocking the merge of #596. | 1.0 | Travis build is broken - The last master build one month ago did still work, now the [same build is broken](https://travis-ci.org/github/zalando/restful-api-guidelines/builds/703500390).
```
docker run -v $TRAVIS_BUILD_DIR:/documents/ --name asciidoc-to-pdf asciidoctor/docker-asciidoctor asciidoctor-pdf -D /documents/output index.adoc
no implicit conversion of nil into String
Use --trace for backtrace
```
Running with `--trace` we can find the root cause to be:
```
docker run -v $TRAVIS_BUILD_DIR:/documents/ --name asciidoc-to-pdf asciidoctor/docker-asciidoctor asciidoctor-pdf --trace -D /documents/output index.adoc
/usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:108:in `basename': no implicit conversion of nil into String (TypeError)
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:108:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:75:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:84:in `block in content'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:84:in `map'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:84:in `content'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:454:in `convert_section'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:106:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:75:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:84:in `block in content'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:84:in `map'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/abstract_block.rb:84:in `content'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/document.rb:1020:in `content'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:359:in `add_chapter'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:250:in `convert_document'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/lib/asciidoctor-epub3/converter.rb:106:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/document.rb:951:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/convert.rb:118:in `convert'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/convert.rb:183:in `block in convert_file'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/convert.rb:183:in `open'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/convert.rb:183:in `convert_file'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/cli/invoker.rb:128:in `block in invoke!'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/cli/invoker.rb:111:in `each'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-2.0.10/lib/asciidoctor/cli/invoker.rb:111:in `invoke!'
from /usr/lib/ruby/gems/2.7.0/gems/asciidoctor-epub3-1.5.0.alpha.18/bin/asciidoctor-epub3:28:in `<top (required)>'
from /usr/bin/asciidoctor-epub3:23:in `load'
from /usr/bin/asciidoctor-epub3:23:in `<main>'
```
This also is blocking the merge of #596. | non_infrastructure | travis build is broken the last master build one month ago did still work now the docker run v travis build dir documents name asciidoc to pdf asciidoctor docker asciidoctor asciidoctor pdf d documents output index adoc no implicit conversion of nil into string use trace for backtrace running with trace we can find the root cause to be docker run v travis build dir documents name asciidoc to pdf asciidoctor docker asciidoctor asciidoctor pdf trace d documents output index adoc usr lib ruby gems gems asciidoctor alpha lib asciidoctor converter rb in basename no implicit conversion of nil into string typeerror from usr lib ruby gems gems asciidoctor alpha lib asciidoctor converter rb in convert from usr lib ruby gems gems asciidoctor lib asciidoctor abstract block rb in convert from usr lib ruby gems gems asciidoctor lib asciidoctor abstract block rb in block in content from usr lib ruby gems gems asciidoctor lib asciidoctor abstract block rb in map from usr lib ruby gems gems asciidoctor lib asciidoctor abstract block rb in content from usr lib ruby gems gems asciidoctor alpha lib asciidoctor converter rb in convert section from usr lib ruby gems gems asciidoctor alpha lib asciidoctor converter rb in convert from usr lib ruby gems gems asciidoctor lib asciidoctor abstract block rb in convert from usr lib ruby gems gems asciidoctor lib asciidoctor abstract block rb in block in content from usr lib ruby gems gems asciidoctor lib asciidoctor abstract block rb in map from usr lib ruby gems gems asciidoctor lib asciidoctor abstract block rb in content from usr lib ruby gems gems asciidoctor lib asciidoctor document rb in content from usr lib ruby gems gems asciidoctor alpha lib asciidoctor converter rb in add chapter from usr lib ruby gems gems asciidoctor alpha lib asciidoctor converter rb in convert document from usr lib ruby gems gems asciidoctor alpha lib asciidoctor converter rb in convert from usr lib ruby gems gems asciidoctor lib asciidoctor document rb in convert from usr lib ruby gems gems asciidoctor lib asciidoctor convert rb in convert from usr lib ruby gems gems asciidoctor lib asciidoctor convert rb in block in convert file from usr lib ruby gems gems asciidoctor lib asciidoctor convert rb in open from usr lib ruby gems gems asciidoctor lib asciidoctor convert rb in convert file from usr lib ruby gems gems asciidoctor lib asciidoctor cli invoker rb in block in invoke from usr lib ruby gems gems asciidoctor lib asciidoctor cli invoker rb in each from usr lib ruby gems gems asciidoctor lib asciidoctor cli invoker rb in invoke from usr lib ruby gems gems asciidoctor alpha bin asciidoctor in from usr bin asciidoctor in load from usr bin asciidoctor in this also is blocking the merge of | 0 |
12,704 | 9,923,066,812 | IssuesEvent | 2019-07-01 05:58:39 | astrolabsoftware/fink-broker | https://api.github.com/repos/astrolabsoftware/fink-broker | closed | raw2science: add generic filtering capability | apache spark infrastructure services | **Describe the issue**
Currently, the `raw2science` service uses hard-coded filters. The idea would be to make this part as much generic as possible, such that the user can upload filters on-demand. | 1.0 | raw2science: add generic filtering capability - **Describe the issue**
Currently, the `raw2science` service uses hard-coded filters. The idea would be to make this part as much generic as possible, such that the user can upload filters on-demand. | infrastructure | add generic filtering capability describe the issue currently the service uses hard coded filters the idea would be to make this part as much generic as possible such that the user can upload filters on demand | 1 |
11,473 | 9,203,340,210 | IssuesEvent | 2019-03-08 02:01:18 | GSA/datagov-deploy | https://api.github.com/repos/GSA/datagov-deploy | closed | Populate: Copy into structure related files | infrastructure server web-proxy | Task: Copy in appropriate file and edit/or amend as needed.
Repo: datagov-deploy-nginx | 1.0 | Populate: Copy into structure related files - Task: Copy in appropriate file and edit/or amend as needed.
Repo: datagov-deploy-nginx | infrastructure | populate copy into structure related files task copy in appropriate file and edit or amend as needed repo datagov deploy nginx | 1 |
551 | 7,114,413,941 | IssuesEvent | 2018-01-18 00:34:58 | brave/browser-laptop | https://api.github.com/repos/brave/browser-laptop | opened | Automate updates when default search engines change favicons | automation polish | If we ship hardcoded favicons for default search engines (#12689) then we should expect those assets to gradually become dated. We should automatically fetch new favicons for each release so that the icons users see are never older than their browser build.
This does imply that we'll be fetching online resources as part of the build process, so it also presents a challenge for reproducible builds and for offline builds. Additionally, we should think about how we can validate or sanitize those assets to ensure that we're not shipping images which are broken, exploit the renderer, are huge, or otherwise mess things up. | 1.0 | Automate updates when default search engines change favicons - If we ship hardcoded favicons for default search engines (#12689) then we should expect those assets to gradually become dated. We should automatically fetch new favicons for each release so that the icons users see are never older than their browser build.
This does imply that we'll be fetching online resources as part of the build process, so it also presents a challenge for reproducible builds and for offline builds. Additionally, we should think about how we can validate or sanitize those assets to ensure that we're not shipping images which are broken, exploit the renderer, are huge, or otherwise mess things up. | non_infrastructure | automate updates when default search engines change favicons if we ship hardcoded favicons for default search engines then we should expect those assets to gradually become dated we should automatically fetch new favicons for each release so that the icons users see are never older than their browser build this does imply that we ll be fetching online resources as part of the build process so it also presents a challenge for reproducible builds and for offline builds additionally we should think about how we can validate or sanitize those assets to ensure that we re not shipping images which are broken exploit the renderer are huge or otherwise mess things up | 0 |
26,054 | 5,221,653,117 | IssuesEvent | 2017-01-27 02:47:52 | IQSS/dataverse | https://api.github.com/repos/IQSS/dataverse | closed | Guides - Building epub version results in warnings | Component: Documentation Status: QA | As noted by @kcondon in #3181 there are warnings when building the docs, which are a results of the epub version being created. This also relates to the warnings when building the pdf version, which are outline in #3134.
```
writing content.opf file...
WARNING: unknown mimetype for _static/jquery.js, ignoring
WARNING: unknown mimetype for _static/underscore.js, ignoring
WARNING: unknown mimetype for _static/doctools.js, ignoring
WARNING: unknown mimetype for _static/websupport.js, ignoring
WARNING: unknown mimetype for _static/navbar_from_dataverse_org.js, ignoring
WARNING: unknown mimetype for _static/navbarscroll.js, ignoring
WARNING: unknown mimetype for _static/searchtools.js, ignoring
WARNING: unknown mimetype for _static/installation/files/issues/2180/grizzly-patch/glassfish-grizzly-extra-all.jar, ignoring
WARNING: unknown mimetype for _static/installation/files/issues/2180/grizzly-patch/readme.md, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/init.d/solr, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/init.d/glassfish, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/httpd/conf.d/dataverse.example.edu.conf, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/httpd/conf.d/ssl.conf, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/shibboleth/shibboleth2.xml, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/shibboleth/attribute-map.xml, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/shibboleth/dataverse-idp-metadata.xml, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/shibboleth/shibGroupTestShib.json, ignoring
WARNING: unknown mimetype for _static/installation/files/home/rpmbuild/rpmbuild/RPMS/x86_64/rapache-1.2.7-rpm0.x86_64.rpm, ignoring
writing toc.ncx file...
writing Dataverse.epub file...
build succeeded, 18 warnings.
``` | 1.0 | Guides - Building epub version results in warnings - As noted by @kcondon in #3181 there are warnings when building the docs, which are a results of the epub version being created. This also relates to the warnings when building the pdf version, which are outline in #3134.
```
writing content.opf file...
WARNING: unknown mimetype for _static/jquery.js, ignoring
WARNING: unknown mimetype for _static/underscore.js, ignoring
WARNING: unknown mimetype for _static/doctools.js, ignoring
WARNING: unknown mimetype for _static/websupport.js, ignoring
WARNING: unknown mimetype for _static/navbar_from_dataverse_org.js, ignoring
WARNING: unknown mimetype for _static/navbarscroll.js, ignoring
WARNING: unknown mimetype for _static/searchtools.js, ignoring
WARNING: unknown mimetype for _static/installation/files/issues/2180/grizzly-patch/glassfish-grizzly-extra-all.jar, ignoring
WARNING: unknown mimetype for _static/installation/files/issues/2180/grizzly-patch/readme.md, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/init.d/solr, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/init.d/glassfish, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/httpd/conf.d/dataverse.example.edu.conf, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/httpd/conf.d/ssl.conf, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/shibboleth/shibboleth2.xml, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/shibboleth/attribute-map.xml, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/shibboleth/dataverse-idp-metadata.xml, ignoring
WARNING: unknown mimetype for _static/installation/files/etc/shibboleth/shibGroupTestShib.json, ignoring
WARNING: unknown mimetype for _static/installation/files/home/rpmbuild/rpmbuild/RPMS/x86_64/rapache-1.2.7-rpm0.x86_64.rpm, ignoring
writing toc.ncx file...
writing Dataverse.epub file...
build succeeded, 18 warnings.
``` | non_infrastructure | guides building epub version results in warnings as noted by kcondon in there are warnings when building the docs which are a results of the epub version being created this also relates to the warnings when building the pdf version which are outline in writing content opf file warning unknown mimetype for static jquery js ignoring warning unknown mimetype for static underscore js ignoring warning unknown mimetype for static doctools js ignoring warning unknown mimetype for static websupport js ignoring warning unknown mimetype for static navbar from dataverse org js ignoring warning unknown mimetype for static navbarscroll js ignoring warning unknown mimetype for static searchtools js ignoring warning unknown mimetype for static installation files issues grizzly patch glassfish grizzly extra all jar ignoring warning unknown mimetype for static installation files issues grizzly patch readme md ignoring warning unknown mimetype for static installation files etc init d solr ignoring warning unknown mimetype for static installation files etc init d glassfish ignoring warning unknown mimetype for static installation files etc httpd conf d dataverse example edu conf ignoring warning unknown mimetype for static installation files etc httpd conf d ssl conf ignoring warning unknown mimetype for static installation files etc shibboleth xml ignoring warning unknown mimetype for static installation files etc shibboleth attribute map xml ignoring warning unknown mimetype for static installation files etc shibboleth dataverse idp metadata xml ignoring warning unknown mimetype for static installation files etc shibboleth shibgrouptestshib json ignoring warning unknown mimetype for static installation files home rpmbuild rpmbuild rpms rapache rpm ignoring writing toc ncx file writing dataverse epub file build succeeded warnings | 0 |
28,797 | 23,496,506,665 | IssuesEvent | 2022-08-18 02:26:20 | happy-travel/agent-app-project | https://api.github.com/repos/happy-travel/agent-app-project | closed | Need logging prices on the 1st and 2nd search step | infrastructure logging 2022q2 | Need to log a full request and response.
Have to send them to a messaging queue. | 1.0 | Need logging prices on the 1st and 2nd search step - Need to log a full request and response.
Have to send them to a messaging queue. | infrastructure | need logging prices on the and search step need to log a full request and response have to send them to a messaging queue | 1 |
32,714 | 26,931,416,446 | IssuesEvent | 2023-02-07 17:05:22 | scalableminds/webknossos | https://api.github.com/repos/scalableminds/webknossos | closed | Schema mismatch details are no longer sent to slack | bug backend infrastructure | regression introduced in #6803
In wk startup, diff schema is run, the output used to be sent to slack, now that message content is missing. | 1.0 | Schema mismatch details are no longer sent to slack - regression introduced in #6803
In wk startup, diff schema is run, the output used to be sent to slack, now that message content is missing. | infrastructure | schema mismatch details are no longer sent to slack regression introduced in in wk startup diff schema is run the output used to be sent to slack now that message content is missing | 1 |
34,157 | 28,374,170,116 | IssuesEvent | 2023-04-12 19:24:03 | strictdoc-project/strictdoc | https://api.github.com/repos/strictdoc-project/strictdoc | closed | tests/end2end: a timeout of 5s is not enough on slower (Windows) machines | Infrastructure / Developer experience SQA | ```
tests\end2end\project_index\UC55_import_document_from_reqif\UC55_T03_import_tree_of_two_documents\test_UC55_T03_import_tree_of_two_documents.py -> Test 13/95
Failed to get an expected response from the server within 5.05 seconds.
--- Received input ---
WARNING: Current configuration will not reload as not all conditions are met, please refer to documentation.
...
received_lines = "".join(received_input)
print(f"\n--- Received input ---\n\n{received_lines}") # noqa: T201
> raise ReadTimeout(timeout_exception.seconds_passed) from None
E tests.end2end.server.ReadTimeout: 5.05
tests\end2end\server.py:326: ReadTimeout
----------- Latest Logs dir: D:\a\strictdoc\strictdoc\latest_logs\ ------------
=========================== short test summary info ===========================
FAILED tests/end2end/project_index/UC55_import_document_from_reqif/UC55_T03_import_tree_of_two_documents/test_UC55_T03_import_tree_of_two_documents.py::Test_UC55_T03_ImportTreeOfTwoDocuments::test_01 - tests.end2end.server.ReadTimeout: 5.05
!!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!
================== 1 failed, 12 passed in 220.31s (0:03:40) ===================
``` | 1.0 | tests/end2end: a timeout of 5s is not enough on slower (Windows) machines - ```
tests\end2end\project_index\UC55_import_document_from_reqif\UC55_T03_import_tree_of_two_documents\test_UC55_T03_import_tree_of_two_documents.py -> Test 13/95
Failed to get an expected response from the server within 5.05 seconds.
--- Received input ---
WARNING: Current configuration will not reload as not all conditions are met, please refer to documentation.
...
received_lines = "".join(received_input)
print(f"\n--- Received input ---\n\n{received_lines}") # noqa: T201
> raise ReadTimeout(timeout_exception.seconds_passed) from None
E tests.end2end.server.ReadTimeout: 5.05
tests\end2end\server.py:326: ReadTimeout
----------- Latest Logs dir: D:\a\strictdoc\strictdoc\latest_logs\ ------------
=========================== short test summary info ===========================
FAILED tests/end2end/project_index/UC55_import_document_from_reqif/UC55_T03_import_tree_of_two_documents/test_UC55_T03_import_tree_of_two_documents.py::Test_UC55_T03_ImportTreeOfTwoDocuments::test_01 - tests.end2end.server.ReadTimeout: 5.05
!!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!
================== 1 failed, 12 passed in 220.31s (0:03:40) ===================
``` | infrastructure | tests a timeout of is not enough on slower windows machines tests project index import document from reqif import tree of two documents test import tree of two documents py test failed to get an expected response from the server within seconds received input warning current configuration will not reload as not all conditions are met please refer to documentation received lines join received input print f n received input n n received lines noqa raise readtimeout timeout exception seconds passed from none e tests server readtimeout tests server py readtimeout latest logs dir d a strictdoc strictdoc latest logs short test summary info failed tests project index import document from reqif import tree of two documents test import tree of two documents py test importtreeoftwodocuments test tests server readtimeout stopping after failures failed passed in | 1 |
788,222 | 27,747,508,101 | IssuesEvent | 2023-03-15 18:03:47 | svthalia/concrexit | https://api.github.com/repos/svthalia/concrexit | closed | Send promo requests overview to promocie | priority: high board easy and fun chore promotion | <!--
Please add the appropriate label for what change should be made:
docs: changes to the documentation)
refactor: refactoring production code, eg. renaming a variable or rewriting a function
test: adding missing tests, refactoring tests; no production code change
chore: updating poetry etc; no production code change
-->
### Describe the change
Since the splitting of paparazcie and promocie, promo mails should go to promocie, not paparazcie.
### Additional context
There's a default var in settings.py. | 1.0 | Send promo requests overview to promocie - <!--
Please add the appropriate label for what change should be made:
docs: changes to the documentation)
refactor: refactoring production code, eg. renaming a variable or rewriting a function
test: adding missing tests, refactoring tests; no production code change
chore: updating poetry etc; no production code change
-->
### Describe the change
Since the splitting of paparazcie and promocie, promo mails should go to promocie, not paparazcie.
### Additional context
There's a default var in settings.py. | non_infrastructure | send promo requests overview to promocie please add the appropriate label for what change should be made docs changes to the documentation refactor refactoring production code eg renaming a variable or rewriting a function test adding missing tests refactoring tests no production code change chore updating poetry etc no production code change describe the change since the splitting of paparazcie and promocie promo mails should go to promocie not paparazcie additional context there s a default var in settings py | 0 |
91,728 | 26,474,392,397 | IssuesEvent | 2023-01-17 10:01:22 | ocaml/ocaml.org | https://api.github.com/repos/ocaml/ocaml.org | closed | Dockerfile incorrect management of opam dependencies | bug build | Project `Dockerfile` intention is to “freeze time” at some point in `opam-repository` history.
This used to be done by having:
```dockerfile
RUN cd ~/opam-repository && git pull origin master && git reset --hard 6496b2727e && opam update
```
This was changed into:
```dockerfile
RUN cd ~/opam-repository && git pull origin 6496b2727e51cc19c9ebaeb73008ded920eea501 && opam update
```
by PR https://github.com/ocaml/ocaml.org/pull/671, which I authored. My bad, this was faulty.
I wrongly assumed the first line of the `Dockerfile`, which contains this:
```
FROM ocaml/opam:alpine-3.17-ocaml-4.14 as build
```
Always resolves into the same image. This assumption is false; this image is built and renewed periodically.
Building on my faulty assumption, I thought: “It's useless to download commits from 6496b2727e to the current HEAD and discard them right away. Let's just move up to where it's needed”. The tricky thing is, at the time you write something as I did, it works. But some time after, HEAD overtakes the hardcoded commit, pull does nothing, and we're floating with a recent HEAD and corresponding untested set of dependencies.
The even trickier thing is, for a reason I fail to understand yet, I wasn't immediately able to reproduce the issue on my side because referring to `ocaml/opam:alpine-3.17-ocaml-4.14` did not invalidate my local docker cache!
Thanks: @mtelvers for helping me understand this | 1.0 | Dockerfile incorrect management of opam dependencies - Project `Dockerfile` intention is to “freeze time” at some point in `opam-repository` history.
This used to be done by having:
```dockerfile
RUN cd ~/opam-repository && git pull origin master && git reset --hard 6496b2727e && opam update
```
This was changed into:
```dockerfile
RUN cd ~/opam-repository && git pull origin 6496b2727e51cc19c9ebaeb73008ded920eea501 && opam update
```
by PR https://github.com/ocaml/ocaml.org/pull/671, which I authored. My bad, this was faulty.
I wrongly assumed the first line of the `Dockerfile`, which contains this:
```
FROM ocaml/opam:alpine-3.17-ocaml-4.14 as build
```
Always resolves into the same image. This assumption is false; this image is built and renewed periodically.
Building on my faulty assumption, I thought: “It's useless to download commits from 6496b2727e to the current HEAD and discard them right away. Let's just move up to where it's needed”. The tricky thing is, at the time you write something as I did, it works. But some time after, HEAD overtakes the hardcoded commit, pull does nothing, and we're floating with a recent HEAD and corresponding untested set of dependencies.
The even trickier thing is, for a reason I fail to understand yet, I wasn't immediately able to reproduce the issue on my side because referring to `ocaml/opam:alpine-3.17-ocaml-4.14` did not invalidate my local docker cache!
Thanks: @mtelvers for helping me understand this | non_infrastructure | dockerfile incorrect management of opam dependencies project dockerfile intention is to “freeze time” at some point in opam repository history this used to be done by having dockerfile run cd opam repository git pull origin master git reset hard opam update this was changed into dockerfile run cd opam repository git pull origin opam update by pr which i authored my bad this was faulty i wrongly assumed the first line of the dockerfile which contains this from ocaml opam alpine ocaml as build always resolves into the same image this assumption is false this image is built and renewed periodically building on my faulty assumption i thought “it s useless to download commits from to the current head and discard them right away let s just move up to where it s needed” the tricky thing is at the time you write something as i did it works but some time after head overtakes the hardcoded commit pull does nothing and we re floating with a recent head and corresponding untested set of dependencies the even trickier thing is for a reason i fail to understand yet i wasn t immediately able to reproduce the issue on my side because referring to ocaml opam alpine ocaml did not invalidate my local docker cache thanks mtelvers for helping me understand this | 0 |
10,264 | 8,461,572,598 | IssuesEvent | 2018-10-22 22:19:06 | GoogleCloudPlatform/forseti-security | https://api.github.com/repos/GoogleCloudPlatform/forseti-security | closed | Data model is broken | module: infrastructure module: model priority: p1 triaged: yes | One of our customers reported the error, with ~1 million inventory objects
> Traceback (most recent call last):\n File \"/usr/local/lib/python2.7/dist-packages/forseti_security-2.1.0-py2.7.egg/google/cloud/forse
ti/services/model/importer/importer.py\", line 283, in run\n 1000\n File \"/usr/local/lib/python2.7/dist-packages/forseti_security-2.1.0-py2.7.eg
g/google/cloud/forseti/services/model/importer/importer.py\", line 332, in model_action_wrapper\n for idx, inventory_data in enumerate(inventory_i
terable, start=1):\n File \"/usr/local/lib/python2.7/dist-packages/forseti_security-2.1.0-py2.7.egg/google/cloud/forseti/services/inventory/storage.
py\", line 879, in iter\n for row in base_query.yield_per(PER_YIELD):\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/orm/loading.py\", line 98,
in instances\n util.raise_from_cause(err)\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/util/compat.py\", line 265, in raise_from_cause\n
reraise(type(exception), exception, tb=exc_tb, cause=cause)\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/orm/loading.py\", line 71, in instances
\n fetch = cursor.fetchmany(query._yield_per)\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/engine/result.py\", line 1166, in fetchmany\n s
elf.cursor, self.context)\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/engine/base.py\", line 1413, in _handle_dbapi_exception\n exc_info\n
File \"build/bdist.linux-x86_64/egg/sqlalchemy/util/compat.py\", line 265, in raise_from_cause\n reraise(type(exception), exception, tb=exc_tb, ca
use=cause)\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/engine/result.py\", line 1159, in fetchmany\n l = self.process_rows(self._fetchmany_i
mpl(size))\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/engine/result.py\", line 1318, in _fetchmany_impl\n row = self._fetchone_impl()\n Fi
le \"build/bdist.linux-x86_64/egg/sqlalchemy/engine/result.py\", line 1308, in _fetchone_impl\n self.__buffer_rows()\n File \"build/bdist.linux-x
86_64/egg/sqlalchemy/engine/result.py\", line 1295, in __buffer_rows\n self.__rowbuffer = collections.deque(self.cursor.fetchmany(size))\n File \
"build/bdist.linux-x86_64/egg/MySQLdb/cursors.py\", line 441, in fetchmany\n r = self._fetch_row(size or self.arraysize)\n File \"build/bdist.lin
ux-x86_64/egg/MySQLdb/cursors.py\", line 327, in _fetch_row\n return self._result.fetch_row(size, self._fetch_type)\nOperationalError: (_mysql_exc
eptions.OperationalError) (2013, 'Lost connection to MySQL server during query') (Background on this error at: http://sqlalche.me/e/e3q8)\n",
"createdAt": "07 August 2018 - 11:11:11",
"description": "{\"pristine\": true, \"source\": \"inventory\", \"source_root\": \"organization/1234567891234\", \"gsuite_enabled\": false, \"source
_info\": {\"inventory_index_id\": 123456789123456789}}" | 1.0 | Data model is broken - One of our customers reported the error, with ~1 million inventory objects
> Traceback (most recent call last):\n File \"/usr/local/lib/python2.7/dist-packages/forseti_security-2.1.0-py2.7.egg/google/cloud/forse
ti/services/model/importer/importer.py\", line 283, in run\n 1000\n File \"/usr/local/lib/python2.7/dist-packages/forseti_security-2.1.0-py2.7.eg
g/google/cloud/forseti/services/model/importer/importer.py\", line 332, in model_action_wrapper\n for idx, inventory_data in enumerate(inventory_i
terable, start=1):\n File \"/usr/local/lib/python2.7/dist-packages/forseti_security-2.1.0-py2.7.egg/google/cloud/forseti/services/inventory/storage.
py\", line 879, in iter\n for row in base_query.yield_per(PER_YIELD):\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/orm/loading.py\", line 98,
in instances\n util.raise_from_cause(err)\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/util/compat.py\", line 265, in raise_from_cause\n
reraise(type(exception), exception, tb=exc_tb, cause=cause)\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/orm/loading.py\", line 71, in instances
\n fetch = cursor.fetchmany(query._yield_per)\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/engine/result.py\", line 1166, in fetchmany\n s
elf.cursor, self.context)\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/engine/base.py\", line 1413, in _handle_dbapi_exception\n exc_info\n
File \"build/bdist.linux-x86_64/egg/sqlalchemy/util/compat.py\", line 265, in raise_from_cause\n reraise(type(exception), exception, tb=exc_tb, ca
use=cause)\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/engine/result.py\", line 1159, in fetchmany\n l = self.process_rows(self._fetchmany_i
mpl(size))\n File \"build/bdist.linux-x86_64/egg/sqlalchemy/engine/result.py\", line 1318, in _fetchmany_impl\n row = self._fetchone_impl()\n Fi
le \"build/bdist.linux-x86_64/egg/sqlalchemy/engine/result.py\", line 1308, in _fetchone_impl\n self.__buffer_rows()\n File \"build/bdist.linux-x
86_64/egg/sqlalchemy/engine/result.py\", line 1295, in __buffer_rows\n self.__rowbuffer = collections.deque(self.cursor.fetchmany(size))\n File \
"build/bdist.linux-x86_64/egg/MySQLdb/cursors.py\", line 441, in fetchmany\n r = self._fetch_row(size or self.arraysize)\n File \"build/bdist.lin
ux-x86_64/egg/MySQLdb/cursors.py\", line 327, in _fetch_row\n return self._result.fetch_row(size, self._fetch_type)\nOperationalError: (_mysql_exc
eptions.OperationalError) (2013, 'Lost connection to MySQL server during query') (Background on this error at: http://sqlalche.me/e/e3q8)\n",
"createdAt": "07 August 2018 - 11:11:11",
"description": "{\"pristine\": true, \"source\": \"inventory\", \"source_root\": \"organization/1234567891234\", \"gsuite_enabled\": false, \"source
_info\": {\"inventory_index_id\": 123456789123456789}}" | infrastructure | data model is broken one of our customers reported the error with million inventory objects traceback most recent call last n file usr local lib dist packages forseti security egg google cloud forse ti services model importer importer py line in run n n file usr local lib dist packages forseti security eg g google cloud forseti services model importer importer py line in model action wrapper n for idx inventory data in enumerate inventory i terable start n file usr local lib dist packages forseti security egg google cloud forseti services inventory storage py line in iter n for row in base query yield per per yield n file build bdist linux egg sqlalchemy orm loading py line in instances n util raise from cause err n file build bdist linux egg sqlalchemy util compat py line in raise from cause n reraise type exception exception tb exc tb cause cause n file build bdist linux egg sqlalchemy orm loading py line in instances n fetch cursor fetchmany query yield per n file build bdist linux egg sqlalchemy engine result py line in fetchmany n s elf cursor self context n file build bdist linux egg sqlalchemy engine base py line in handle dbapi exception n exc info n file build bdist linux egg sqlalchemy util compat py line in raise from cause n reraise type exception exception tb exc tb ca use cause n file build bdist linux egg sqlalchemy engine result py line in fetchmany n l self process rows self fetchmany i mpl size n file build bdist linux egg sqlalchemy engine result py line in fetchmany impl n row self fetchone impl n fi le build bdist linux egg sqlalchemy engine result py line in fetchone impl n self buffer rows n file build bdist linux x egg sqlalchemy engine result py line in buffer rows n self rowbuffer collections deque self cursor fetchmany size n file build bdist linux egg mysqldb cursors py line in fetchmany n r self fetch row size or self arraysize n file build bdist lin ux egg mysqldb cursors py line in fetch row n return self result fetch row size self fetch type noperationalerror mysql exc eptions operationalerror lost connection to mysql server during query background on this error at createdat august description pristine true source inventory source root organization gsuite enabled false source info inventory index id | 1 |
13,469 | 10,270,381,138 | IssuesEvent | 2019-08-23 11:28:01 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | Send period in metricbeat data | :infrastructure Metricbeat candidate enhancement | For querying metricbeat documents for the Kibana Infra UI app, it would be very helpful to know the interval in which data is sent by the various metricbeat modules.
One idea that came up in a call is to include the configured `period` in the documents themselves.
| 1.0 | Send period in metricbeat data - For querying metricbeat documents for the Kibana Infra UI app, it would be very helpful to know the interval in which data is sent by the various metricbeat modules.
One idea that came up in a call is to include the configured `period` in the documents themselves.
| infrastructure | send period in metricbeat data for querying metricbeat documents for the kibana infra ui app it would be very helpful to know the interval in which data is sent by the various metricbeat modules one idea that came up in a call is to include the configured period in the documents themselves | 1 |
28,981 | 23,646,790,715 | IssuesEvent | 2022-08-25 23:35:03 | ThorstenSauter/NoPlan | https://api.github.com/repos/ThorstenSauter/NoPlan | closed | Bring all cloud resources under management by Pulumi | infrastructure | In the ongoing effort to bring all cloud infrastructure resources under management by Pulumi, a few critical resouces are still missing.
### Missing resouces
- [ ] `Azure Container App` running the `NoPlan.Api` project
- [x] #112 | 1.0 | Bring all cloud resources under management by Pulumi - In the ongoing effort to bring all cloud infrastructure resources under management by Pulumi, a few critical resouces are still missing.
### Missing resouces
- [ ] `Azure Container App` running the `NoPlan.Api` project
- [x] #112 | infrastructure | bring all cloud resources under management by pulumi in the ongoing effort to bring all cloud infrastructure resources under management by pulumi a few critical resouces are still missing missing resouces azure container app running the noplan api project | 1 |
4,366 | 3,358,231,937 | IssuesEvent | 2015-11-19 08:16:44 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Mismatched number of arguments in call to EditorExportPlatformPC::export_project | bug confirmed topic:buildsystem | As seen in the current master branch, on gcc 4.9.2:
```
g++ -o platform/windows/export/export.x11.opt.tools.64.o -c -O2 -ffast-math -DDEBUG_ENABLED -DFREETYPE_ENABLED -DOPENGL_ENABLED -DGLEW_ENABLED -DALSA_ENABLED -DPULSEAUDIO_ENABLED -DX11_ENABLED -DUNIX_ENABLED -DGLES2_ENABLED -DGLES_OVER_GL -DMUSEPACK_ENABLED -DSQUISH_ENABLED -DVORBIS_ENABLED -DOPUS_ENABLED -DTHEORA_ENABLED -DPNG_ENABLED -DDDS_ENABLED -DPVR_ENABLED -DJPG_ENABLED -DWEBP_ENABLED -DSPEEX_ENABLED -DTOOLS_ENABLED -DGDSCRIPT_ENABLED -DMINIZIP_ENABLED -DXML_ENABLED -DETC1_ENABLED -D_REENTRANT -Icore -Icore/math -Itools -Idrivers -I. -Iplatform/x11 -I/usr/include/freetype2 -I/usr/include/libpng16 -Itools/vorbis -Idrivers/opus/silk/float -Idrivers/opus -Idrivers/opus/celt -Idrivers/opus/silk -Idrivers/opus/silk/float platform/windows/export/export.cpp
platform/windows/export/export.cpp: In member function 'Error EditorExportPlatformWindows::export_project(const String&, bool, bool, bool)':
platform/windows/export/export.cpp:202:92: error: no matching function for call to 'EditorExportPlatformWindows::export_project(const String&, bool&, bool&, bool&)'
Error err = EditorExportPlatformPC::export_project(p_path, p_debug, p_dumb, p_remote_debug);
^
platform/windows/export/export.cpp:202:92: note: candidate is:
In file included from platform/windows/export/export.h:1:0,
from platform/windows/export/export.cpp:30:
./tools/editor/editor_import_export.h:200:16: note: virtual Error EditorExportPlatformPC::export_project(const String&, bool, int)
virtual Error export_project(const String& p_path, bool p_debug, int p_flags=0);
^
./tools/editor/editor_import_export.h:200:16: note: candidate expects 3 arguments, 4 provided
``` | 1.0 | Mismatched number of arguments in call to EditorExportPlatformPC::export_project - As seen in the current master branch, on gcc 4.9.2:
```
g++ -o platform/windows/export/export.x11.opt.tools.64.o -c -O2 -ffast-math -DDEBUG_ENABLED -DFREETYPE_ENABLED -DOPENGL_ENABLED -DGLEW_ENABLED -DALSA_ENABLED -DPULSEAUDIO_ENABLED -DX11_ENABLED -DUNIX_ENABLED -DGLES2_ENABLED -DGLES_OVER_GL -DMUSEPACK_ENABLED -DSQUISH_ENABLED -DVORBIS_ENABLED -DOPUS_ENABLED -DTHEORA_ENABLED -DPNG_ENABLED -DDDS_ENABLED -DPVR_ENABLED -DJPG_ENABLED -DWEBP_ENABLED -DSPEEX_ENABLED -DTOOLS_ENABLED -DGDSCRIPT_ENABLED -DMINIZIP_ENABLED -DXML_ENABLED -DETC1_ENABLED -D_REENTRANT -Icore -Icore/math -Itools -Idrivers -I. -Iplatform/x11 -I/usr/include/freetype2 -I/usr/include/libpng16 -Itools/vorbis -Idrivers/opus/silk/float -Idrivers/opus -Idrivers/opus/celt -Idrivers/opus/silk -Idrivers/opus/silk/float platform/windows/export/export.cpp
platform/windows/export/export.cpp: In member function 'Error EditorExportPlatformWindows::export_project(const String&, bool, bool, bool)':
platform/windows/export/export.cpp:202:92: error: no matching function for call to 'EditorExportPlatformWindows::export_project(const String&, bool&, bool&, bool&)'
Error err = EditorExportPlatformPC::export_project(p_path, p_debug, p_dumb, p_remote_debug);
^
platform/windows/export/export.cpp:202:92: note: candidate is:
In file included from platform/windows/export/export.h:1:0,
from platform/windows/export/export.cpp:30:
./tools/editor/editor_import_export.h:200:16: note: virtual Error EditorExportPlatformPC::export_project(const String&, bool, int)
virtual Error export_project(const String& p_path, bool p_debug, int p_flags=0);
^
./tools/editor/editor_import_export.h:200:16: note: candidate expects 3 arguments, 4 provided
``` | non_infrastructure | mismatched number of arguments in call to editorexportplatformpc export project as seen in the current master branch on gcc g o platform windows export export opt tools o c ffast math ddebug enabled dfreetype enabled dopengl enabled dglew enabled dalsa enabled dpulseaudio enabled enabled dunix enabled enabled dgles over gl dmusepack enabled dsquish enabled dvorbis enabled dopus enabled dtheora enabled dpng enabled ddds enabled dpvr enabled djpg enabled dwebp enabled dspeex enabled dtools enabled dgdscript enabled dminizip enabled dxml enabled enabled d reentrant icore icore math itools idrivers i iplatform i usr include i usr include itools vorbis idrivers opus silk float idrivers opus idrivers opus celt idrivers opus silk idrivers opus silk float platform windows export export cpp platform windows export export cpp in member function error editorexportplatformwindows export project const string bool bool bool platform windows export export cpp error no matching function for call to editorexportplatformwindows export project const string bool bool bool error err editorexportplatformpc export project p path p debug p dumb p remote debug platform windows export export cpp note candidate is in file included from platform windows export export h from platform windows export export cpp tools editor editor import export h note virtual error editorexportplatformpc export project const string bool int virtual error export project const string p path bool p debug int p flags tools editor editor import export h note candidate expects arguments provided | 0 |
33,080 | 27,206,687,821 | IssuesEvent | 2023-02-20 13:37:17 | getlago/lago | https://api.github.com/repos/getlago/lago | closed | [FEAT]: Kubernetes helm chart | Infrastructure issue: docker | **Is your feature request related to a problem? Please describe.**
I would like to install Lago on a Kubernetes cluster.
**Describe the solution you'd like**
Docker-compose files are already available, it shouldn't be too complicated to write an Helm chart
**Describe alternatives you've considered**
No real alternatives available for Kubernetes.
**Additional context**
Nothing to add here.
| 1.0 | [FEAT]: Kubernetes helm chart - **Is your feature request related to a problem? Please describe.**
I would like to install Lago on a Kubernetes cluster.
**Describe the solution you'd like**
Docker-compose files are already available, it shouldn't be too complicated to write an Helm chart
**Describe alternatives you've considered**
No real alternatives available for Kubernetes.
**Additional context**
Nothing to add here.
| infrastructure | kubernetes helm chart is your feature request related to a problem please describe i would like to install lago on a kubernetes cluster describe the solution you d like docker compose files are already available it shouldn t be too complicated to write an helm chart describe alternatives you ve considered no real alternatives available for kubernetes additional context nothing to add here | 1 |
346,207 | 24,886,625,626 | IssuesEvent | 2022-10-28 08:20:49 | jeromehjj/ped | https://api.github.com/repos/jeromehjj/ped | opened | Name of product not consistent in the UG | severity.VeryLow type.DocumentationBug | In the product description and title for the UG, it is given as "myStudents", however in the FAQ section, it is named as "myStudent".


<!--session: 1666944177595-663c4fc1-c96d-4939-b420-c48d52103a05-->
<!--Version: Web v3.4.4--> | 1.0 | Name of product not consistent in the UG - In the product description and title for the UG, it is given as "myStudents", however in the FAQ section, it is named as "myStudent".


<!--session: 1666944177595-663c4fc1-c96d-4939-b420-c48d52103a05-->
<!--Version: Web v3.4.4--> | non_infrastructure | name of product not consistent in the ug in the product description and title for the ug it is given as mystudents however in the faq section it is named as mystudent | 0 |
33,076 | 27,206,214,263 | IssuesEvent | 2023-02-20 13:16:46 | Kalipo-BV/kalipo-core | https://api.github.com/repos/Kalipo-BV/kalipo-core | closed | As auditor, I want to be able to create and manage different roles, so that I can assign different levels of access and permissions to users. | CSC Infrastructure | ## User story specification
As auditor, I want to be able to create and manage different roles, so that I can assign different levels of access and permissions to users.
<!--
## Acceptance criteria
### Functionals
**Pre conditions**
- ...
**Process**
- De gebruiker opent ...
- Het systeem geeft ...
- De gebruiker vult ...
**Result**
- ...
**Wireframe:**
...
### Non-functionals
- Veiligheid
- Snelheid
- Beschikbaarheid
- Betrouwbaarheid
**ISO 25010**
[https://docs.google.com/spreadsheets/d/1ge06Lyd5Bhn-QTgtjGqhO41yXi2Z6NaX/edit#gid=1777548652](url)
## Constraints
- C: Commerce, communicatie, cultuur
- O: Organisatie
- P: Members
- A: Organisatie
- F: Financiën
- **_I: Informatie/ Software_**
Blockchain
SDK
- J: Juridische
- **_T: Technologie_**
LiskSDK
Vue
- H: Huisvesting
## DoD
Een user story voldoet aan de DoD wanneer deze voldoet aan:
- De stijl van Kalipo
- Gereviewd door de product owner en voldoet aan acceptatiecriteria
- Er is een spellingscontrole gedaan
- Er zijn peer-reviews geweest
- Getest
- Gedocumenteerd
## Deliverables
- [ ]
--> | 1.0 | As auditor, I want to be able to create and manage different roles, so that I can assign different levels of access and permissions to users. - ## User story specification
As auditor, I want to be able to create and manage different roles, so that I can assign different levels of access and permissions to users.
<!--
## Acceptance criteria
### Functionals
**Pre conditions**
- ...
**Process**
- De gebruiker opent ...
- Het systeem geeft ...
- De gebruiker vult ...
**Result**
- ...
**Wireframe:**
...
### Non-functionals
- Veiligheid
- Snelheid
- Beschikbaarheid
- Betrouwbaarheid
**ISO 25010**
[https://docs.google.com/spreadsheets/d/1ge06Lyd5Bhn-QTgtjGqhO41yXi2Z6NaX/edit#gid=1777548652](url)
## Constraints
- C: Commerce, communicatie, cultuur
- O: Organisatie
- P: Members
- A: Organisatie
- F: Financiën
- **_I: Informatie/ Software_**
Blockchain
SDK
- J: Juridische
- **_T: Technologie_**
LiskSDK
Vue
- H: Huisvesting
## DoD
Een user story voldoet aan de DoD wanneer deze voldoet aan:
- De stijl van Kalipo
- Gereviewd door de product owner en voldoet aan acceptatiecriteria
- Er is een spellingscontrole gedaan
- Er zijn peer-reviews geweest
- Getest
- Gedocumenteerd
## Deliverables
- [ ]
--> | infrastructure | as auditor i want to be able to create and manage different roles so that i can assign different levels of access and permissions to users user story specification as auditor i want to be able to create and manage different roles so that i can assign different levels of access and permissions to users acceptance criteria functionals pre conditions process de gebruiker opent het systeem geeft de gebruiker vult result wireframe non functionals veiligheid snelheid beschikbaarheid betrouwbaarheid iso url constraints c commerce communicatie cultuur o organisatie p members a organisatie f financiën i informatie software blockchain sdk j juridische t technologie lisksdk vue h huisvesting dod een user story voldoet aan de dod wanneer deze voldoet aan de stijl van kalipo gereviewd door de product owner en voldoet aan acceptatiecriteria er is een spellingscontrole gedaan er zijn peer reviews geweest getest gedocumenteerd deliverables | 1 |
88,897 | 3,787,129,718 | IssuesEvent | 2016-03-21 09:09:40 | salesagility/SuiteCRM | https://api.github.com/repos/salesagility/SuiteCRM | closed | Since upgrade from 7.4.3 to 7.5 issues | bug High Priority | Hello; We just installed the update from 7.4.3 to 7.5
The unified search has stopped working
Aswell as the Dashlets, Upgrade Wizard
All of the pages go to a blank white page.
| 1.0 | Since upgrade from 7.4.3 to 7.5 issues - Hello; We just installed the update from 7.4.3 to 7.5
The unified search has stopped working
Aswell as the Dashlets, Upgrade Wizard
All of the pages go to a blank white page.
| non_infrastructure | since upgrade from to issues hello we just installed the update from to the unified search has stopped working aswell as the dashlets upgrade wizard all of the pages go to a blank white page | 0 |
5,257 | 5,541,258,424 | IssuesEvent | 2017-03-22 12:22:21 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Move corefx from OSX 10.10 to 10.12 | area-Infrastructure | For 2.0 we are supporting only a min of OSX 10.12 so we need to move any references to 10.12.
- Update any CI or build machines
- Update any RID stuff in the repo
fyi @gkhanna79 @ericstj | 1.0 | Move corefx from OSX 10.10 to 10.12 - For 2.0 we are supporting only a min of OSX 10.12 so we need to move any references to 10.12.
- Update any CI or build machines
- Update any RID stuff in the repo
fyi @gkhanna79 @ericstj | infrastructure | move corefx from osx to for we are supporting only a min of osx so we need to move any references to update any ci or build machines update any rid stuff in the repo fyi ericstj | 1 |
102,099 | 31,824,205,217 | IssuesEvent | 2023-09-14 06:13:54 | abetlen/llama-cpp-python | https://api.github.com/repos/abetlen/llama-cpp-python | closed | Failed to build 0.2.2 on Ubuntu 22.04 with `--user` | bug build | Ran the following on an intel Ubuntu 22.04 system:
```
$ pip3 install --user llama-cpp-python Collecting llama-cpp-python
Using cached llama_cpp_python-0.2.2.tar.gz (1.5 MB)
Installing build dependencies ... done
Getting requirements to build wheel ... done
Installing backend dependencies ... done
Preparing metadata (pyproject.toml) ... done
Collecting diskcache>=5.6.1
Using cached diskcache-5.6.3-py3-none-any.whl (45 kB)
Requirement already satisfied: typing-extensions>=4.5.0 in /home/lyda@azdsthq.net/.local/lib/python3.10/site-packages (from llama-cpp-python) (4.7.1)
Collecting numpy>=1.20.0
Using cached numpy-1.25.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (18.2 MB)
Building wheels for collected packages: llama-cpp-python
Building wheel for llama-cpp-python (pyproject.toml) ... error
error: subprocess-exited-with-error
× Building wheel for llama-cpp-python (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [71 lines of output]
*** scikit-build-core 0.5.0 using CMake 3.27.4 (wheel)
*** Configuring CMake...
loading initial cache file /tmp/tmp9u4e4cug/build/CMakeInit.txt
-- The C compiler identification is GNU 11.4.0
-- The CXX compiler identification is GNU 11.4.0
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Check for working C compiler: /usr/bin/cc - skipped
-- Detecting C compile features
-- Detecting C compile features - done
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Found Git: /usr/bin/git (found version "2.41.0")
fatal: not a git repository (or any of the parent directories): .git
fatal: not a git repository (or any of the parent directories): .git
CMake Warning at vendor/llama.cpp/CMakeLists.txt:125 (message):
Git repository not found; to enable automatic generation of build info,
make sure Git is installed and the project is a Git repository.
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success
-- Found Threads: TRUE
-- CMAKE_SYSTEM_PROCESSOR: x86_64
-- x86 detected
-- Configuring done (0.3s)
-- Generating done (0.0s)
-- Build files have been written to: /tmp/tmp9u4e4cug/build
*** Building project with Ninja...
Change Dir: '/tmp/tmp9u4e4cug/build'
Run Build Command(s): /tmp/pip-build-env-l3ehx8u0/normal/local/lib/python3.10/dist-packages/ninja/data/bin/ninja -v
[1/10] /usr/bin/cc -DGGML_USE_K_QUANTS -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Wno-unused-function -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/CMakeFiles/ggml.dir/ggml-alloc.c.o -MF vendor/llama.cpp/CMakeFiles/ggml.dir/ggml-alloc.c.o.d -o vendor/llama.cpp/CMakeFiles/ggml.dir/ggml-alloc.c.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/ggml-alloc.c
[2/10] /usr/bin/c++ -DGGML_USE_K_QUANTS -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/common/. -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -Wno-format-truncation -Wno-array-bounds -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/console.cpp.o -MF vendor/llama.cpp/common/CMakeFiles/common.dir/console.cpp.o.d -o vendor/llama.cpp/common/CMakeFiles/common.dir/console.cpp.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/common/console.cpp
[3/10] /usr/bin/c++ -DGGML_USE_K_QUANTS -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/common/. -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -Wno-format-truncation -Wno-array-bounds -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/grammar-parser.cpp.o -MF vendor/llama.cpp/common/CMakeFiles/common.dir/grammar-parser.cpp.o.d -o vendor/llama.cpp/common/CMakeFiles/common.dir/grammar-parser.cpp.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/common/grammar-parser.cpp
[4/10] /usr/bin/cc -DGGML_USE_K_QUANTS -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Wno-unused-function -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/CMakeFiles/ggml.dir/k_quants.c.o -MF vendor/llama.cpp/CMakeFiles/ggml.dir/k_quants.c.o.d -o vendor/llama.cpp/CMakeFiles/ggml.dir/k_quants.c.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/k_quants.c
[5/10] /usr/bin/c++ -DGGML_USE_K_QUANTS -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/common/. -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -Wno-format-truncation -Wno-array-bounds -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/common.cpp.o -MF vendor/llama.cpp/common/CMakeFiles/common.dir/common.cpp.o.d -o vendor/llama.cpp/common/CMakeFiles/common.dir/common.cpp.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/common/common.cpp
[6/10] /usr/bin/cc -DGGML_USE_K_QUANTS -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Wno-unused-function -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/CMakeFiles/ggml.dir/ggml.c.o -MF vendor/llama.cpp/CMakeFiles/ggml.dir/ggml.c.o.d -o vendor/llama.cpp/CMakeFiles/ggml.dir/ggml.c.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/ggml.c
[7/10] : && /usr/bin/cc -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libggml_shared.so -o vendor/llama.cpp/libggml_shared.so vendor/llama.cpp/CMakeFiles/ggml.dir/ggml.c.o vendor/llama.cpp/CMakeFiles/ggml.dir/ggml-alloc.c.o vendor/llama.cpp/CMakeFiles/ggml.dir/k_quants.c.o && :
[8/10] : && /tmp/pip-build-env-l3ehx8u0/normal/local/lib/python3.10/dist-packages/cmake/data/bin/cmake -E rm -f vendor/llama.cpp/libggml_static.a && /usr/bin/ar qc vendor/llama.cpp/libggml_static.a vendor/llama.cpp/CMakeFiles/ggml.dir/ggml.c.o vendor/llama.cpp/CMakeFiles/ggml.dir/ggml-alloc.c.o vendor/llama.cpp/CMakeFiles/ggml.dir/k_quants.c.o && /usr/bin/ranlib vendor/llama.cpp/libggml_static.a && :
[9/10] /usr/bin/c++ -DGGML_USE_K_QUANTS -DLLAMA_BUILD -DLLAMA_SHARED -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -Dllama_EXPORTS -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -Wno-format-truncation -Wno-array-bounds -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/CMakeFiles/llama.dir/llama.cpp.o -MF vendor/llama.cpp/CMakeFiles/llama.dir/llama.cpp.o.d -o vendor/llama.cpp/CMakeFiles/llama.dir/llama.cpp.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/llama.cpp
[10/10] : && /usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libllama.so -o vendor/llama.cpp/libllama.so vendor/llama.cpp/CMakeFiles/ggml.dir/ggml.c.o vendor/llama.cpp/CMakeFiles/ggml.dir/ggml-alloc.c.o vendor/llama.cpp/CMakeFiles/ggml.dir/k_quants.c.o vendor/llama.cpp/CMakeFiles/llama.dir/llama.cpp.o && :
*** Installing project into wheel...
-- Install configuration: "Release"
-- Installing: /tmp/tmp9u4e4cug/wheel/platlib/lib/libggml_shared.so
-- Installing: /tmp/tmp9u4e4cug/wheel/platlib/lib/libllama.so
-- Installing: /tmp/tmp9u4e4cug/wheel/platlib/bin/convert.py
-- Installing: /tmp/tmp9u4e4cug/wheel/platlib/bin/convert-lora-to-ggml.py
-- Installing: /tmp/tmp9u4e4cug/wheel/platlib/llama_cpp/libllama.so
-- Installing: /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/llama_cpp/libllama.so
*** Making wheel...
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/pip/_vendor/pep517/in_process/_in_process.py", line 363, in <module>
main()
File "/usr/lib/python3/dist-packages/pip/_vendor/pep517/in_process/_in_process.py", line 345, in main
json_out['return_val'] = hook(**hook_input['kwargs'])
File "/usr/lib/python3/dist-packages/pip/_vendor/pep517/in_process/_in_process.py", line 261, in build_wheel
return _build_backend().build_wheel(wheel_directory, config_settings,
File "/tmp/pip-build-env-l3ehx8u0/overlay/local/lib/python3.10/dist-packages/scikit_build_core/build/__init__.py", line 32, in build_wheel
return _build_wheel_impl(
File "/tmp/pip-build-env-l3ehx8u0/overlay/local/lib/python3.10/dist-packages/scikit_build_core/build/wheel.py", line 261, in _build_wheel_impl
mapping = packages_to_file_mapping(
File "/tmp/pip-build-env-l3ehx8u0/overlay/local/lib/python3.10/dist-packages/scikit_build_core/build/_pathutil.py", line 43, in packages_to_file_mapping
for filepath in each_unignored_file(
File "/tmp/pip-build-env-l3ehx8u0/overlay/local/lib/python3.10/dist-packages/scikit_build_core/build/_file_processor.py", line 41, in each_unignored_file
exclude_spec = pathspec.GitIgnoreSpec.from_lines(exclude_lines)
AttributeError: module 'pathspec' has no attribute 'GitIgnoreSpec'
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building wheel for llama-cpp-python
Failed to build llama-cpp-python
ERROR: Could not build wheels for llama-cpp-python, which is required to install pyproject.toml-based projects
``` | 1.0 | Failed to build 0.2.2 on Ubuntu 22.04 with `--user` - Ran the following on an intel Ubuntu 22.04 system:
```
$ pip3 install --user llama-cpp-python Collecting llama-cpp-python
Using cached llama_cpp_python-0.2.2.tar.gz (1.5 MB)
Installing build dependencies ... done
Getting requirements to build wheel ... done
Installing backend dependencies ... done
Preparing metadata (pyproject.toml) ... done
Collecting diskcache>=5.6.1
Using cached diskcache-5.6.3-py3-none-any.whl (45 kB)
Requirement already satisfied: typing-extensions>=4.5.0 in /home/lyda@azdsthq.net/.local/lib/python3.10/site-packages (from llama-cpp-python) (4.7.1)
Collecting numpy>=1.20.0
Using cached numpy-1.25.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (18.2 MB)
Building wheels for collected packages: llama-cpp-python
Building wheel for llama-cpp-python (pyproject.toml) ... error
error: subprocess-exited-with-error
× Building wheel for llama-cpp-python (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [71 lines of output]
*** scikit-build-core 0.5.0 using CMake 3.27.4 (wheel)
*** Configuring CMake...
loading initial cache file /tmp/tmp9u4e4cug/build/CMakeInit.txt
-- The C compiler identification is GNU 11.4.0
-- The CXX compiler identification is GNU 11.4.0
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Check for working C compiler: /usr/bin/cc - skipped
-- Detecting C compile features
-- Detecting C compile features - done
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Found Git: /usr/bin/git (found version "2.41.0")
fatal: not a git repository (or any of the parent directories): .git
fatal: not a git repository (or any of the parent directories): .git
CMake Warning at vendor/llama.cpp/CMakeLists.txt:125 (message):
Git repository not found; to enable automatic generation of build info,
make sure Git is installed and the project is a Git repository.
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success
-- Found Threads: TRUE
-- CMAKE_SYSTEM_PROCESSOR: x86_64
-- x86 detected
-- Configuring done (0.3s)
-- Generating done (0.0s)
-- Build files have been written to: /tmp/tmp9u4e4cug/build
*** Building project with Ninja...
Change Dir: '/tmp/tmp9u4e4cug/build'
Run Build Command(s): /tmp/pip-build-env-l3ehx8u0/normal/local/lib/python3.10/dist-packages/ninja/data/bin/ninja -v
[1/10] /usr/bin/cc -DGGML_USE_K_QUANTS -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Wno-unused-function -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/CMakeFiles/ggml.dir/ggml-alloc.c.o -MF vendor/llama.cpp/CMakeFiles/ggml.dir/ggml-alloc.c.o.d -o vendor/llama.cpp/CMakeFiles/ggml.dir/ggml-alloc.c.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/ggml-alloc.c
[2/10] /usr/bin/c++ -DGGML_USE_K_QUANTS -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/common/. -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -Wno-format-truncation -Wno-array-bounds -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/console.cpp.o -MF vendor/llama.cpp/common/CMakeFiles/common.dir/console.cpp.o.d -o vendor/llama.cpp/common/CMakeFiles/common.dir/console.cpp.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/common/console.cpp
[3/10] /usr/bin/c++ -DGGML_USE_K_QUANTS -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/common/. -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -Wno-format-truncation -Wno-array-bounds -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/grammar-parser.cpp.o -MF vendor/llama.cpp/common/CMakeFiles/common.dir/grammar-parser.cpp.o.d -o vendor/llama.cpp/common/CMakeFiles/common.dir/grammar-parser.cpp.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/common/grammar-parser.cpp
[4/10] /usr/bin/cc -DGGML_USE_K_QUANTS -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Wno-unused-function -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/CMakeFiles/ggml.dir/k_quants.c.o -MF vendor/llama.cpp/CMakeFiles/ggml.dir/k_quants.c.o.d -o vendor/llama.cpp/CMakeFiles/ggml.dir/k_quants.c.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/k_quants.c
[5/10] /usr/bin/c++ -DGGML_USE_K_QUANTS -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/common/. -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -Wno-format-truncation -Wno-array-bounds -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/common.cpp.o -MF vendor/llama.cpp/common/CMakeFiles/common.dir/common.cpp.o.d -o vendor/llama.cpp/common/CMakeFiles/common.dir/common.cpp.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/common/common.cpp
[6/10] /usr/bin/cc -DGGML_USE_K_QUANTS -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Wno-unused-function -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/CMakeFiles/ggml.dir/ggml.c.o -MF vendor/llama.cpp/CMakeFiles/ggml.dir/ggml.c.o.d -o vendor/llama.cpp/CMakeFiles/ggml.dir/ggml.c.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/ggml.c
[7/10] : && /usr/bin/cc -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libggml_shared.so -o vendor/llama.cpp/libggml_shared.so vendor/llama.cpp/CMakeFiles/ggml.dir/ggml.c.o vendor/llama.cpp/CMakeFiles/ggml.dir/ggml-alloc.c.o vendor/llama.cpp/CMakeFiles/ggml.dir/k_quants.c.o && :
[8/10] : && /tmp/pip-build-env-l3ehx8u0/normal/local/lib/python3.10/dist-packages/cmake/data/bin/cmake -E rm -f vendor/llama.cpp/libggml_static.a && /usr/bin/ar qc vendor/llama.cpp/libggml_static.a vendor/llama.cpp/CMakeFiles/ggml.dir/ggml.c.o vendor/llama.cpp/CMakeFiles/ggml.dir/ggml-alloc.c.o vendor/llama.cpp/CMakeFiles/ggml.dir/k_quants.c.o && /usr/bin/ranlib vendor/llama.cpp/libggml_static.a && :
[9/10] /usr/bin/c++ -DGGML_USE_K_QUANTS -DLLAMA_BUILD -DLLAMA_SHARED -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -Dllama_EXPORTS -I/tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -Wno-format-truncation -Wno-array-bounds -mf16c -mfma -mavx -mavx2 -MD -MT vendor/llama.cpp/CMakeFiles/llama.dir/llama.cpp.o -MF vendor/llama.cpp/CMakeFiles/llama.dir/llama.cpp.o.d -o vendor/llama.cpp/CMakeFiles/llama.dir/llama.cpp.o -c /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/vendor/llama.cpp/llama.cpp
[10/10] : && /usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libllama.so -o vendor/llama.cpp/libllama.so vendor/llama.cpp/CMakeFiles/ggml.dir/ggml.c.o vendor/llama.cpp/CMakeFiles/ggml.dir/ggml-alloc.c.o vendor/llama.cpp/CMakeFiles/ggml.dir/k_quants.c.o vendor/llama.cpp/CMakeFiles/llama.dir/llama.cpp.o && :
*** Installing project into wheel...
-- Install configuration: "Release"
-- Installing: /tmp/tmp9u4e4cug/wheel/platlib/lib/libggml_shared.so
-- Installing: /tmp/tmp9u4e4cug/wheel/platlib/lib/libllama.so
-- Installing: /tmp/tmp9u4e4cug/wheel/platlib/bin/convert.py
-- Installing: /tmp/tmp9u4e4cug/wheel/platlib/bin/convert-lora-to-ggml.py
-- Installing: /tmp/tmp9u4e4cug/wheel/platlib/llama_cpp/libllama.so
-- Installing: /tmp/pip-install-2zaya40h/llama-cpp-python_6d3abaf7e49f4232888c7f28a96156a1/llama_cpp/libllama.so
*** Making wheel...
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/pip/_vendor/pep517/in_process/_in_process.py", line 363, in <module>
main()
File "/usr/lib/python3/dist-packages/pip/_vendor/pep517/in_process/_in_process.py", line 345, in main
json_out['return_val'] = hook(**hook_input['kwargs'])
File "/usr/lib/python3/dist-packages/pip/_vendor/pep517/in_process/_in_process.py", line 261, in build_wheel
return _build_backend().build_wheel(wheel_directory, config_settings,
File "/tmp/pip-build-env-l3ehx8u0/overlay/local/lib/python3.10/dist-packages/scikit_build_core/build/__init__.py", line 32, in build_wheel
return _build_wheel_impl(
File "/tmp/pip-build-env-l3ehx8u0/overlay/local/lib/python3.10/dist-packages/scikit_build_core/build/wheel.py", line 261, in _build_wheel_impl
mapping = packages_to_file_mapping(
File "/tmp/pip-build-env-l3ehx8u0/overlay/local/lib/python3.10/dist-packages/scikit_build_core/build/_pathutil.py", line 43, in packages_to_file_mapping
for filepath in each_unignored_file(
File "/tmp/pip-build-env-l3ehx8u0/overlay/local/lib/python3.10/dist-packages/scikit_build_core/build/_file_processor.py", line 41, in each_unignored_file
exclude_spec = pathspec.GitIgnoreSpec.from_lines(exclude_lines)
AttributeError: module 'pathspec' has no attribute 'GitIgnoreSpec'
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building wheel for llama-cpp-python
Failed to build llama-cpp-python
ERROR: Could not build wheels for llama-cpp-python, which is required to install pyproject.toml-based projects
``` | non_infrastructure | failed to build on ubuntu with user ran the following on an intel ubuntu system install user llama cpp python collecting llama cpp python using cached llama cpp python tar gz mb installing build dependencies done getting requirements to build wheel done installing backend dependencies done preparing metadata pyproject toml done collecting diskcache using cached diskcache none any whl kb requirement already satisfied typing extensions in home lyda azdsthq net local lib site packages from llama cpp python collecting numpy using cached numpy manylinux whl mb building wheels for collected packages llama cpp python building wheel for llama cpp python pyproject toml error error subprocess exited with error × building wheel for llama cpp python pyproject toml did not run successfully │ exit code ╰─ scikit build core using cmake wheel configuring cmake loading initial cache file tmp build cmakeinit txt the c compiler identification is gnu the cxx compiler identification is gnu detecting c compiler abi info detecting c compiler abi info done check for working c compiler usr bin cc skipped detecting c compile features detecting c compile features done detecting cxx compiler abi info detecting cxx compiler abi info done check for working cxx compiler usr bin c skipped detecting cxx compile features detecting cxx compile features done found git usr bin git found version fatal not a git repository or any of the parent directories git fatal not a git repository or any of the parent directories git cmake warning at vendor llama cpp cmakelists txt message git repository not found to enable automatic generation of build info make sure git is installed and the project is a git repository performing test cmake have libc pthread performing test cmake have libc pthread success found threads true cmake system processor detected configuring done generating done build files have been written to tmp build building project with ninja change dir tmp build run build command s tmp pip build env normal local lib dist packages ninja data bin ninja v usr bin cc dggml use k quants d gnu source d xopen source i tmp pip install llama cpp python vendor llama cpp dndebug std fpic wall wextra wpedantic wcast qual wdouble promotion wshadow wstrict prototypes wpointer arith wmissing prototypes werror implicit int wno unused function mfma mavx md mt vendor llama cpp cmakefiles ggml dir ggml alloc c o mf vendor llama cpp cmakefiles ggml dir ggml alloc c o d o vendor llama cpp cmakefiles ggml dir ggml alloc c o c tmp pip install llama cpp python vendor llama cpp ggml alloc c usr bin c dggml use k quants d gnu source d xopen source i tmp pip install llama cpp python vendor llama cpp common i tmp pip install llama cpp python vendor llama cpp dndebug std gnu fpic wall wextra wpedantic wcast qual wno unused function wno multichar wno format truncation wno array bounds mfma mavx md mt vendor llama cpp common cmakefiles common dir console cpp o mf vendor llama cpp common cmakefiles common dir console cpp o d o vendor llama cpp common cmakefiles common dir console cpp o c tmp pip install llama cpp python vendor llama cpp common console cpp usr bin c dggml use k quants d gnu source d xopen source i tmp pip install llama cpp python vendor llama cpp common i tmp pip install llama cpp python vendor llama cpp dndebug std gnu fpic wall wextra wpedantic wcast qual wno unused function wno multichar wno format truncation wno array bounds mfma mavx md mt vendor llama cpp common cmakefiles common dir grammar parser cpp o mf vendor llama cpp common cmakefiles common dir grammar parser cpp o d o vendor llama cpp common cmakefiles common dir grammar parser cpp o c tmp pip install llama cpp python vendor llama cpp common grammar parser cpp usr bin cc dggml use k quants d gnu source d xopen source i tmp pip install llama cpp python vendor llama cpp dndebug std fpic wall wextra wpedantic wcast qual wdouble promotion wshadow wstrict prototypes wpointer arith wmissing prototypes werror implicit int wno unused function mfma mavx md mt vendor llama cpp cmakefiles ggml dir k quants c o mf vendor llama cpp cmakefiles ggml dir k quants c o d o vendor llama cpp cmakefiles ggml dir k quants c o c tmp pip install llama cpp python vendor llama cpp k quants c usr bin c dggml use k quants d gnu source d xopen source i tmp pip install llama cpp python vendor llama cpp common i tmp pip install llama cpp python vendor llama cpp dndebug std gnu fpic wall wextra wpedantic wcast qual wno unused function wno multichar wno format truncation wno array bounds mfma mavx md mt vendor llama cpp common cmakefiles common dir common cpp o mf vendor llama cpp common cmakefiles common dir common cpp o d o vendor llama cpp common cmakefiles common dir common cpp o c tmp pip install llama cpp python vendor llama cpp common common cpp usr bin cc dggml use k quants d gnu source d xopen source i tmp pip install llama cpp python vendor llama cpp dndebug std fpic wall wextra wpedantic wcast qual wdouble promotion wshadow wstrict prototypes wpointer arith wmissing prototypes werror implicit int wno unused function mfma mavx md mt vendor llama cpp cmakefiles ggml dir ggml c o mf vendor llama cpp cmakefiles ggml dir ggml c o d o vendor llama cpp cmakefiles ggml dir ggml c o c tmp pip install llama cpp python vendor llama cpp ggml c usr bin cc fpic dndebug shared wl soname libggml shared so o vendor llama cpp libggml shared so vendor llama cpp cmakefiles ggml dir ggml c o vendor llama cpp cmakefiles ggml dir ggml alloc c o vendor llama cpp cmakefiles ggml dir k quants c o tmp pip build env normal local lib dist packages cmake data bin cmake e rm f vendor llama cpp libggml static a usr bin ar qc vendor llama cpp libggml static a vendor llama cpp cmakefiles ggml dir ggml c o vendor llama cpp cmakefiles ggml dir ggml alloc c o vendor llama cpp cmakefiles ggml dir k quants c o usr bin ranlib vendor llama cpp libggml static a usr bin c dggml use k quants dllama build dllama shared d gnu source d xopen source dllama exports i tmp pip install llama cpp python vendor llama cpp dndebug std gnu fpic wall wextra wpedantic wcast qual wno unused function wno multichar wno format truncation wno array bounds mfma mavx md mt vendor llama cpp cmakefiles llama dir llama cpp o mf vendor llama cpp cmakefiles llama dir llama cpp o d o vendor llama cpp cmakefiles llama dir llama cpp o c tmp pip install llama cpp python vendor llama cpp llama cpp usr bin c fpic dndebug shared wl soname libllama so o vendor llama cpp libllama so vendor llama cpp cmakefiles ggml dir ggml c o vendor llama cpp cmakefiles ggml dir ggml alloc c o vendor llama cpp cmakefiles ggml dir k quants c o vendor llama cpp cmakefiles llama dir llama cpp o installing project into wheel install configuration release installing tmp wheel platlib lib libggml shared so installing tmp wheel platlib lib libllama so installing tmp wheel platlib bin convert py installing tmp wheel platlib bin convert lora to ggml py installing tmp wheel platlib llama cpp libllama so installing tmp pip install llama cpp python llama cpp libllama so making wheel traceback most recent call last file usr lib dist packages pip vendor in process in process py line in main file usr lib dist packages pip vendor in process in process py line in main json out hook hook input file usr lib dist packages pip vendor in process in process py line in build wheel return build backend build wheel wheel directory config settings file tmp pip build env overlay local lib dist packages scikit build core build init py line in build wheel return build wheel impl file tmp pip build env overlay local lib dist packages scikit build core build wheel py line in build wheel impl mapping packages to file mapping file tmp pip build env overlay local lib dist packages scikit build core build pathutil py line in packages to file mapping for filepath in each unignored file file tmp pip build env overlay local lib dist packages scikit build core build file processor py line in each unignored file exclude spec pathspec gitignorespec from lines exclude lines attributeerror module pathspec has no attribute gitignorespec note this error originates from a subprocess and is likely not a problem with pip error failed building wheel for llama cpp python failed to build llama cpp python error could not build wheels for llama cpp python which is required to install pyproject toml based projects | 0 |
7,634 | 7,032,391,036 | IssuesEvent | 2017-12-27 03:00:09 | casey/just | https://api.github.com/repos/casey/just | closed | Add to macports | help wanted infrastructure | Macports is easy, since I'm already the maintainer of the just package. I'm not sure if just meets the notability clause to be accepted to homebrew.
Edit: @Ezku added just to homebrew | 1.0 | Add to macports - Macports is easy, since I'm already the maintainer of the just package. I'm not sure if just meets the notability clause to be accepted to homebrew.
Edit: @Ezku added just to homebrew | infrastructure | add to macports macports is easy since i m already the maintainer of the just package i m not sure if just meets the notability clause to be accepted to homebrew edit ezku added just to homebrew | 1 |
276,778 | 21,000,091,553 | IssuesEvent | 2022-03-29 16:35:49 | jrenjq/PainAssessment | https://api.github.com/repos/jrenjq/PainAssessment | closed | D2: Apply SOLID Principles | documentation enhancement | Evidence and Rationale of SOLID Principles
- [x] Apply at least 2 SOLID Principles | 1.0 | D2: Apply SOLID Principles - Evidence and Rationale of SOLID Principles
- [x] Apply at least 2 SOLID Principles | non_infrastructure | apply solid principles evidence and rationale of solid principles apply at least solid principles | 0 |
11,640 | 9,335,894,786 | IssuesEvent | 2019-03-28 19:46:44 | servo/servo | https://api.github.com/repos/servo/servo | closed | Failed to open sync WPT upstream PR due to using wrong branch | A-infrastructure | I just found sometimes we'll miss to open sync WPT upstream PRs.
After checking the logs on buildbot, I found it's caused by checking out to wrong branch.
For example, in [build 1010](https://build.servo.org/builders/mac-nightly/builds/1010/),
we created and switched to branch `wpt_update_16-03-2019` in [step fetch-and-update-expectations](https://build.servo.org/builders/mac-nightly/builds/1010/steps/shell__4/logs/stdio)
```
Switched to a new branch 'wpt_update_16-03-2019'
[wpt_update_16-03-2019 df6dbf1b30] Update web-platform-tests to revision 78a0ccb785b63ee5124bd86adf438248cc4922f5
Date: Sat Mar 16 21:43:39 2019 -0400
```
however, we checked out to branch `wpt_update_17-03-2019` which is an unknown branch on that day in [step open-pr](https://build.servo.org/builders/mac-nightly/builds/1010/steps/shell__5/logs/stdio).
```
error: pathspec 'wpt_update_17-03-2019' did not match any file(s) known to git.
program finished with exit code 0
```
So, we didn't open the PR for `16-03-2019`.
(By the way, there's another [case](https://build.servo.org/builders/mac-nightly/builds/1007) for `13-03-2019` with same issue)
I'm wondering, if it's possible that this issue is caused by long running time in `step fetch-and-update-expectations`, then we run the next step in a new day.
https://github.com/servo/servo/blob/db7bb2a5101ea6042654b59b3b81725e2da65891/etc/ci/update-wpt-checkout#L13-L14
Maybe we can make the `CURRENT_DATE` as a environment variable so that we can always use same variable between steps in a same build ? | 1.0 | Failed to open sync WPT upstream PR due to using wrong branch - I just found sometimes we'll miss to open sync WPT upstream PRs.
After checking the logs on buildbot, I found it's caused by checking out to wrong branch.
For example, in [build 1010](https://build.servo.org/builders/mac-nightly/builds/1010/),
we created and switched to branch `wpt_update_16-03-2019` in [step fetch-and-update-expectations](https://build.servo.org/builders/mac-nightly/builds/1010/steps/shell__4/logs/stdio)
```
Switched to a new branch 'wpt_update_16-03-2019'
[wpt_update_16-03-2019 df6dbf1b30] Update web-platform-tests to revision 78a0ccb785b63ee5124bd86adf438248cc4922f5
Date: Sat Mar 16 21:43:39 2019 -0400
```
however, we checked out to branch `wpt_update_17-03-2019` which is an unknown branch on that day in [step open-pr](https://build.servo.org/builders/mac-nightly/builds/1010/steps/shell__5/logs/stdio).
```
error: pathspec 'wpt_update_17-03-2019' did not match any file(s) known to git.
program finished with exit code 0
```
So, we didn't open the PR for `16-03-2019`.
(By the way, there's another [case](https://build.servo.org/builders/mac-nightly/builds/1007) for `13-03-2019` with same issue)
I'm wondering, if it's possible that this issue is caused by long running time in `step fetch-and-update-expectations`, then we run the next step in a new day.
https://github.com/servo/servo/blob/db7bb2a5101ea6042654b59b3b81725e2da65891/etc/ci/update-wpt-checkout#L13-L14
Maybe we can make the `CURRENT_DATE` as a environment variable so that we can always use same variable between steps in a same build ? | infrastructure | failed to open sync wpt upstream pr due to using wrong branch i just found sometimes we ll miss to open sync wpt upstream prs after checking the logs on buildbot i found it s caused by checking out to wrong branch for example in we created and switched to branch wpt update in switched to a new branch wpt update update web platform tests to revision date sat mar however we checked out to branch wpt update which is an unknown branch on that day in error pathspec wpt update did not match any file s known to git program finished with exit code so we didn t open the pr for by the way there s another for with same issue i m wondering if it s possible that this issue is caused by long running time in step fetch and update expectations then we run the next step in a new day maybe we can make the current date as a environment variable so that we can always use same variable between steps in a same build | 1 |
162,128 | 20,164,397,322 | IssuesEvent | 2022-02-10 01:48:39 | kapseliboi/ILIAS-Pegasus | https://api.github.com/repos/kapseliboi/ILIAS-Pegasus | opened | CVE-2020-7608 (Medium) detected in multiple libraries | security vulnerability | ## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-8.1.0.tgz</b>, <b>yargs-parser-11.1.1.tgz</b>, <b>yargs-parser-7.0.0.tgz</b>, <b>yargs-parser-13.1.1.tgz</b></p></summary>
<p>
<details><summary><b>yargs-parser-8.1.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-8.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-8.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/cli-highlight/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- typeorm-0.1.21.tgz (Root Library)
- cli-highlight-1.2.3.tgz
- yargs-10.1.2.tgz
- :x: **yargs-parser-8.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-11.1.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/protractor/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- protractor-5.4.4.tgz (Root Library)
- yargs-12.0.5.tgz
- :x: **yargs-parser-11.1.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-7.0.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- typeorm-0.1.21.tgz (Root Library)
- yargs-9.0.1.tgz
- :x: **yargs-parser-7.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-13.1.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@angular/compiler-cli/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- compiler-cli-8.2.14.tgz (Root Library)
- yargs-13.1.0.tgz
- :x: **yargs-parser-13.1.1.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2">https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2</a></p>
<p>Release Date: 2020-03-16</p>
<p>Fix Resolution (yargs-parser): 13.1.2</p>
<p>Direct dependency fix Resolution (typeorm): 0.2.15-rc.1</p><p>Fix Resolution (yargs-parser): 13.1.2</p>
<p>Direct dependency fix Resolution (protractor): 7.0.0</p><p>Fix Resolution (yargs-parser): 13.1.2</p>
<p>Direct dependency fix Resolution (typeorm): 0.2.15-rc.1</p><p>Fix Resolution (yargs-parser): 13.1.2</p>
<p>Direct dependency fix Resolution (@angular/compiler-cli): 9.0.0-next.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7608 (Medium) detected in multiple libraries - ## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-8.1.0.tgz</b>, <b>yargs-parser-11.1.1.tgz</b>, <b>yargs-parser-7.0.0.tgz</b>, <b>yargs-parser-13.1.1.tgz</b></p></summary>
<p>
<details><summary><b>yargs-parser-8.1.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-8.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-8.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/cli-highlight/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- typeorm-0.1.21.tgz (Root Library)
- cli-highlight-1.2.3.tgz
- yargs-10.1.2.tgz
- :x: **yargs-parser-8.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-11.1.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/protractor/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- protractor-5.4.4.tgz (Root Library)
- yargs-12.0.5.tgz
- :x: **yargs-parser-11.1.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-7.0.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-7.0.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- typeorm-0.1.21.tgz (Root Library)
- yargs-9.0.1.tgz
- :x: **yargs-parser-7.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-13.1.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/@angular/compiler-cli/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- compiler-cli-8.2.14.tgz (Root Library)
- yargs-13.1.0.tgz
- :x: **yargs-parser-13.1.1.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2">https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2</a></p>
<p>Release Date: 2020-03-16</p>
<p>Fix Resolution (yargs-parser): 13.1.2</p>
<p>Direct dependency fix Resolution (typeorm): 0.2.15-rc.1</p><p>Fix Resolution (yargs-parser): 13.1.2</p>
<p>Direct dependency fix Resolution (protractor): 7.0.0</p><p>Fix Resolution (yargs-parser): 13.1.2</p>
<p>Direct dependency fix Resolution (typeorm): 0.2.15-rc.1</p><p>Fix Resolution (yargs-parser): 13.1.2</p>
<p>Direct dependency fix Resolution (@angular/compiler-cli): 9.0.0-next.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries yargs parser tgz yargs parser tgz yargs parser tgz yargs parser tgz yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file package json path to vulnerable library node modules cli highlight node modules yargs parser package json dependency hierarchy typeorm tgz root library cli highlight tgz yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file package json path to vulnerable library node modules protractor node modules yargs parser package json dependency hierarchy protractor tgz root library yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file package json path to vulnerable library node modules yargs parser package json dependency hierarchy typeorm tgz root library yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file package json path to vulnerable library node modules angular compiler cli node modules yargs parser package json dependency hierarchy compiler cli tgz root library yargs tgz x yargs parser tgz vulnerable library found in base branch master vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution yargs parser direct dependency fix resolution typeorm rc fix resolution yargs parser direct dependency fix resolution protractor fix resolution yargs parser direct dependency fix resolution typeorm rc fix resolution yargs parser direct dependency fix resolution angular compiler cli next step up your open source security game with whitesource | 0 |
165,781 | 20,620,527,422 | IssuesEvent | 2022-03-07 16:59:23 | Baneeishaque/Gym_Class_Accounter | https://api.github.com/repos/Baneeishaque/Gym_Class_Accounter | opened | CVE-2018-1000180 (High) detected in bcprov-jdk15on-1.56.jar | security vulnerability | ## CVE-2018-1000180 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.56.jar</b></p></summary>
<p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p>
<p>Library home page: <a href="http://www.bouncycastle.org/java.html">http://www.bouncycastle.org/java.html</a></p>
<p>Path to dependency file: /app/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.bouncycastle/bcprov-jdk15on/1.56/a153c6f9744a3e9dd6feab5e210e1c9861362ec7/bcprov-jdk15on-1.56.jar</p>
<p>
Dependency Hierarchy:
- lint-gradle-26.6.4.jar (Root Library)
- builder-3.6.4.jar
- :x: **bcprov-jdk15on-1.56.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Baneeishaque/Gym_Class_Accounter/commit/d0b484ab35339c58b68f75b5f03b3958d80ce307">d0b484ab35339c58b68f75b5f03b3958d80ce307</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Bouncy Castle BC 1.54 - 1.59, BC-FJA 1.0.0, BC-FJA 1.0.1 and earlier have a flaw in the Low-level interface to RSA key pair generator, specifically RSA Key Pairs generated in low-level API with added certainty may have less M-R tests than expected. This appears to be fixed in versions BC 1.60 beta 4 and later, BC-FJA 1.0.2 and later.
<p>Publish Date: 2018-06-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000180>CVE-2018-1000180</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000180">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000180</a></p>
<p>Release Date: 2018-06-05</p>
<p>Fix Resolution: org.bouncycastle:bc-fips:1.0.2;org.bouncycastle:bcprov-jdk15on:1.60;org.bouncycastle:bcprov-jdk14:1.60;org.bouncycastle:bcprov-ext-jdk14:1.60;org.bouncycastle:bcprov-ext-jdk15on:1.60;org.bouncycastle:bcprov-debug-jdk14:1.60;org.bouncycastle:bcprov-debug-jdk15on:1.60</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-1000180 (High) detected in bcprov-jdk15on-1.56.jar - ## CVE-2018-1000180 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.56.jar</b></p></summary>
<p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p>
<p>Library home page: <a href="http://www.bouncycastle.org/java.html">http://www.bouncycastle.org/java.html</a></p>
<p>Path to dependency file: /app/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.bouncycastle/bcprov-jdk15on/1.56/a153c6f9744a3e9dd6feab5e210e1c9861362ec7/bcprov-jdk15on-1.56.jar</p>
<p>
Dependency Hierarchy:
- lint-gradle-26.6.4.jar (Root Library)
- builder-3.6.4.jar
- :x: **bcprov-jdk15on-1.56.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Baneeishaque/Gym_Class_Accounter/commit/d0b484ab35339c58b68f75b5f03b3958d80ce307">d0b484ab35339c58b68f75b5f03b3958d80ce307</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Bouncy Castle BC 1.54 - 1.59, BC-FJA 1.0.0, BC-FJA 1.0.1 and earlier have a flaw in the Low-level interface to RSA key pair generator, specifically RSA Key Pairs generated in low-level API with added certainty may have less M-R tests than expected. This appears to be fixed in versions BC 1.60 beta 4 and later, BC-FJA 1.0.2 and later.
<p>Publish Date: 2018-06-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000180>CVE-2018-1000180</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000180">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000180</a></p>
<p>Release Date: 2018-06-05</p>
<p>Fix Resolution: org.bouncycastle:bc-fips:1.0.2;org.bouncycastle:bcprov-jdk15on:1.60;org.bouncycastle:bcprov-jdk14:1.60;org.bouncycastle:bcprov-ext-jdk14:1.60;org.bouncycastle:bcprov-ext-jdk15on:1.60;org.bouncycastle:bcprov-debug-jdk14:1.60;org.bouncycastle:bcprov-debug-jdk15on:1.60</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in bcprov jar cve high severity vulnerability vulnerable library bcprov jar the bouncy castle crypto package is a java implementation of cryptographic algorithms this jar contains jce provider and lightweight api for the bouncy castle cryptography apis for jdk to jdk library home page a href path to dependency file app build gradle path to vulnerable library home wss scanner gradle caches modules files org bouncycastle bcprov bcprov jar dependency hierarchy lint gradle jar root library builder jar x bcprov jar vulnerable library found in head commit a href vulnerability details bouncy castle bc bc fja bc fja and earlier have a flaw in the low level interface to rsa key pair generator specifically rsa key pairs generated in low level api with added certainty may have less m r tests than expected this appears to be fixed in versions bc beta and later bc fja and later publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org bouncycastle bc fips org bouncycastle bcprov org bouncycastle bcprov org bouncycastle bcprov ext org bouncycastle bcprov ext org bouncycastle bcprov debug org bouncycastle bcprov debug step up your open source security game with whitesource | 0 |
706,691 | 24,281,458,248 | IssuesEvent | 2022-09-28 17:47:51 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | opened | [CloudPortal][Multi Region GEO Partitioning] After region is removed from existing cluster data pinned to that region is not deleted | kind/bug priority/medium jira-originated | Jira Link: [DB-3693](https://yugabyte.atlassian.net/browse/DB-3693)
| 1.0 | [CloudPortal][Multi Region GEO Partitioning] After region is removed from existing cluster data pinned to that region is not deleted - Jira Link: [DB-3693](https://yugabyte.atlassian.net/browse/DB-3693)
| non_infrastructure | after region is removed from existing cluster data pinned to that region is not deleted jira link | 0 |
4,386 | 5,030,487,210 | IssuesEvent | 2016-12-16 01:02:34 | zulip/zulip | https://api.github.com/repos/zulip/zulip | closed | Replace deprecated `self.assertEquals` with `self.assertEqual` | area: testing-infrastructure bite size bug | If you run `test-backend`, with Django 1.10, we now get tons of these warnings.
/home/tabbott/zulip/zerver/tests/test_upload.py:582: DeprecationWarning: Please use assertEqual instead.
self.assertEquals(sanitize_name(u'tarball.tar.gz'), u'tarball.tar.gz')
We should be able to clean them up pretty quickly with `git grep` and `sed -i`. | 1.0 | Replace deprecated `self.assertEquals` with `self.assertEqual` - If you run `test-backend`, with Django 1.10, we now get tons of these warnings.
/home/tabbott/zulip/zerver/tests/test_upload.py:582: DeprecationWarning: Please use assertEqual instead.
self.assertEquals(sanitize_name(u'tarball.tar.gz'), u'tarball.tar.gz')
We should be able to clean them up pretty quickly with `git grep` and `sed -i`. | infrastructure | replace deprecated self assertequals with self assertequal if you run test backend with django we now get tons of these warnings home tabbott zulip zerver tests test upload py deprecationwarning please use assertequal instead self assertequals sanitize name u tarball tar gz u tarball tar gz we should be able to clean them up pretty quickly with git grep and sed i | 1 |
7,002 | 3,933,932,686 | IssuesEvent | 2016-04-25 20:49:29 | jens-maus/yam | https://api.github.com/repos/jens-maus/yam | closed | Warn about incompatible filetype when importing mails | #trivial @normal enhancement fixed GUI nightly build | **Originally by _mario@cattaneo.de_ on 2011-01-22 19:47:12 +0100**
___
Try to import eg. S:Startup-Sequence (I know, you have to change the filter to do so, but this is just an example ;)).
YAM will silently ignore that file, ie. do nothing. Would be nice, if there was a requester informing the user that something went wrong (incompatiible/corrupted...) with the selected file. | 1.0 | Warn about incompatible filetype when importing mails - **Originally by _mario@cattaneo.de_ on 2011-01-22 19:47:12 +0100**
___
Try to import eg. S:Startup-Sequence (I know, you have to change the filter to do so, but this is just an example ;)).
YAM will silently ignore that file, ie. do nothing. Would be nice, if there was a requester informing the user that something went wrong (incompatiible/corrupted...) with the selected file. | non_infrastructure | warn about incompatible filetype when importing mails originally by mario cattaneo de on try to import eg s startup sequence i know you have to change the filter to do so but this is just an example yam will silently ignore that file ie do nothing would be nice if there was a requester informing the user that something went wrong incompatiible corrupted with the selected file | 0 |
31,340 | 25,571,670,069 | IssuesEvent | 2022-11-30 18:12:26 | CDCgov/data-exchange-hl7 | https://api.github.com/repos/CDCgov/data-exchange-hl7 | closed | SPIKE: Event Grid / Eventhub access review | bug infrastructure waiting | We ran into a problem last week regarding Eventhub access via various integrations (EDAV Databricks, function apps, eventgrid access via storage account changes, developer access to it), especially with the setup of the SystemAssigned vs UserAssigned roles.
## Goal
The goal is to define a suitable method for the architecture within our DEV environment Eventhub so that access to it can be secured on a network and access based level. Our test currently involves dropping a file or files into a storage account container, ensure that change events are fired/pushed to eventgrid, and eventgrid pushes the event over to Eventhub to be processed by our function application.
## Technical Background
We've worked with an OHMS cloud engineer on Oct 19 and 20 on an encountered problem where events from a [blobcreate changes](https://learn.microsoft.com/en-us/azure/event-grid/system-topics) flowing to Event Grid system topic were not publishing to our Eventhub.
The flow can be understood as such. Both of these event pushes are push mechanism.
Storage Account -> blobcreate event ---> Event Grid system topic ---> Eventhub.
### Network Problems
We focused debugging on Event Grid system topic access, and found that this traffic cannot use a private endpoint to Eventhub because these system topic changes likely come from public IPs within the Azure IP range. They're outside of our VNET boundary based on [this article](https://learn.microsoft.com/en-us/azure/event-grid/consume-private-endpoints). We made good use of the [Event Grid's diagnostic log](https://learn.microsoft.com/en-us/azure/event-grid/enable-diagnostic-logs-topic) to find out that IP filtering from the Eventhub firewall was happening and dropping Event Grid event pushes.

We changed this setting to allow for public network access via selected networks with the setting "Allow trusted Microsoft services to bypass this firewall: Yes", thinking that Event Grid will be part of the "trusted Microsoft services" to resolve the network routing / firewall problems.
### Access Problems
In addition to network problems, we noticed access problems when creating a Event Grid system topic with a Managed Identity of "None".
We wanted to use a "SystemAssigned" identity but did not have the ability to grant this identity to our event grid system topic.
We created an user identity under Managed Identity, and with an OHMS engineer's help, they provided roles to the user identity. The Event Grid system topic is now using this user identity, but after reviewing the debug logs for Eventhub access - events are still being dropped via "Unauthorized" access.
This is still being actively looked into. | 1.0 | SPIKE: Event Grid / Eventhub access review - We ran into a problem last week regarding Eventhub access via various integrations (EDAV Databricks, function apps, eventgrid access via storage account changes, developer access to it), especially with the setup of the SystemAssigned vs UserAssigned roles.
## Goal
The goal is to define a suitable method for the architecture within our DEV environment Eventhub so that access to it can be secured on a network and access based level. Our test currently involves dropping a file or files into a storage account container, ensure that change events are fired/pushed to eventgrid, and eventgrid pushes the event over to Eventhub to be processed by our function application.
## Technical Background
We've worked with an OHMS cloud engineer on Oct 19 and 20 on an encountered problem where events from a [blobcreate changes](https://learn.microsoft.com/en-us/azure/event-grid/system-topics) flowing to Event Grid system topic were not publishing to our Eventhub.
The flow can be understood as such. Both of these event pushes are push mechanism.
Storage Account -> blobcreate event ---> Event Grid system topic ---> Eventhub.
### Network Problems
We focused debugging on Event Grid system topic access, and found that this traffic cannot use a private endpoint to Eventhub because these system topic changes likely come from public IPs within the Azure IP range. They're outside of our VNET boundary based on [this article](https://learn.microsoft.com/en-us/azure/event-grid/consume-private-endpoints). We made good use of the [Event Grid's diagnostic log](https://learn.microsoft.com/en-us/azure/event-grid/enable-diagnostic-logs-topic) to find out that IP filtering from the Eventhub firewall was happening and dropping Event Grid event pushes.

We changed this setting to allow for public network access via selected networks with the setting "Allow trusted Microsoft services to bypass this firewall: Yes", thinking that Event Grid will be part of the "trusted Microsoft services" to resolve the network routing / firewall problems.
### Access Problems
In addition to network problems, we noticed access problems when creating a Event Grid system topic with a Managed Identity of "None".
We wanted to use a "SystemAssigned" identity but did not have the ability to grant this identity to our event grid system topic.
We created an user identity under Managed Identity, and with an OHMS engineer's help, they provided roles to the user identity. The Event Grid system topic is now using this user identity, but after reviewing the debug logs for Eventhub access - events are still being dropped via "Unauthorized" access.
This is still being actively looked into. | infrastructure | spike event grid eventhub access review we ran into a problem last week regarding eventhub access via various integrations edav databricks function apps eventgrid access via storage account changes developer access to it especially with the setup of the systemassigned vs userassigned roles goal the goal is to define a suitable method for the architecture within our dev environment eventhub so that access to it can be secured on a network and access based level our test currently involves dropping a file or files into a storage account container ensure that change events are fired pushed to eventgrid and eventgrid pushes the event over to eventhub to be processed by our function application technical background we ve worked with an ohms cloud engineer on oct and on an encountered problem where events from a flowing to event grid system topic were not publishing to our eventhub the flow can be understood as such both of these event pushes are push mechanism storage account blobcreate event event grid system topic eventhub network problems we focused debugging on event grid system topic access and found that this traffic cannot use a private endpoint to eventhub because these system topic changes likely come from public ips within the azure ip range they re outside of our vnet boundary based on we made good use of the to find out that ip filtering from the eventhub firewall was happening and dropping event grid event pushes we changed this setting to allow for public network access via selected networks with the setting allow trusted microsoft services to bypass this firewall yes thinking that event grid will be part of the trusted microsoft services to resolve the network routing firewall problems access problems in addition to network problems we noticed access problems when creating a event grid system topic with a managed identity of none we wanted to use a systemassigned identity but did not have the ability to grant this identity to our event grid system topic we created an user identity under managed identity and with an ohms engineer s help they provided roles to the user identity the event grid system topic is now using this user identity but after reviewing the debug logs for eventhub access events are still being dropped via unauthorized access this is still being actively looked into | 1 |
257,566 | 8,139,053,898 | IssuesEvent | 2018-08-20 16:26:32 | processing/p5.js-web-editor | https://api.github.com/repos/processing/p5.js-web-editor | closed | Forum button links to old forum | good first issue help wanted priority:medium type:bug | #### Nature of issue?
- [ ] Found a bug
- [x] Existing feature enhancement
- [ ] New feature request
(Though this might count as a bug?)
#### Details:
Clicking on "Forum", after clicking on "about" in the "Help & Feedback" drop-down, still sends you to the [old forum ](https://forum.processing.org/two/) rather than the [new one](https://discourse.processing.org/). | 1.0 | Forum button links to old forum - #### Nature of issue?
- [ ] Found a bug
- [x] Existing feature enhancement
- [ ] New feature request
(Though this might count as a bug?)
#### Details:
Clicking on "Forum", after clicking on "about" in the "Help & Feedback" drop-down, still sends you to the [old forum ](https://forum.processing.org/two/) rather than the [new one](https://discourse.processing.org/). | non_infrastructure | forum button links to old forum nature of issue found a bug existing feature enhancement new feature request though this might count as a bug details clicking on forum after clicking on about in the help feedback drop down still sends you to the rather than the | 0 |
13,492 | 10,291,103,135 | IssuesEvent | 2019-08-27 13:42:20 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | Automatically bisect failing builds on the Flutter 3xH builder | P1-High area-infrastructure | We now have two 3xH Flutter builders:
- flutter-engine-linux (builds every single commit, is often far behind despite concurrent builds).
- flutter-engine-linux-fast (merges builds, but does not pinpoint the failure to a particular commit).
If we add automatic bisection of failures to the -fast builder, we can get the best of both worlds and can free up ~4 linux bots continuously building flutter-engine-linux. There will still be a lot of latency because the builder takes 2 hours to run successfully, but with the bisection we should be able to get the results within hours, rather than days.
The proposal is to:
- If a build fails and the previous build was a success, start bisection.
- Trigger a build in the middle of the blamelist.
- If the build fails, trigger a build in the middle of the older builds of the blamelist.
- If the build succeeds, trigger a build in the middle of the newer builds of the blamelist.
- Continue until the blamelist becomes a single commit.
Additionally, because there may be multiple failures caused by different commits, we can "fan out" when we encounter a new failure and start to bisect that separately (in parallel). | 1.0 | Automatically bisect failing builds on the Flutter 3xH builder - We now have two 3xH Flutter builders:
- flutter-engine-linux (builds every single commit, is often far behind despite concurrent builds).
- flutter-engine-linux-fast (merges builds, but does not pinpoint the failure to a particular commit).
If we add automatic bisection of failures to the -fast builder, we can get the best of both worlds and can free up ~4 linux bots continuously building flutter-engine-linux. There will still be a lot of latency because the builder takes 2 hours to run successfully, but with the bisection we should be able to get the results within hours, rather than days.
The proposal is to:
- If a build fails and the previous build was a success, start bisection.
- Trigger a build in the middle of the blamelist.
- If the build fails, trigger a build in the middle of the older builds of the blamelist.
- If the build succeeds, trigger a build in the middle of the newer builds of the blamelist.
- Continue until the blamelist becomes a single commit.
Additionally, because there may be multiple failures caused by different commits, we can "fan out" when we encounter a new failure and start to bisect that separately (in parallel). | infrastructure | automatically bisect failing builds on the flutter builder we now have two flutter builders flutter engine linux builds every single commit is often far behind despite concurrent builds flutter engine linux fast merges builds but does not pinpoint the failure to a particular commit if we add automatic bisection of failures to the fast builder we can get the best of both worlds and can free up linux bots continuously building flutter engine linux there will still be a lot of latency because the builder takes hours to run successfully but with the bisection we should be able to get the results within hours rather than days the proposal is to if a build fails and the previous build was a success start bisection trigger a build in the middle of the blamelist if the build fails trigger a build in the middle of the older builds of the blamelist if the build succeeds trigger a build in the middle of the newer builds of the blamelist continue until the blamelist becomes a single commit additionally because there may be multiple failures caused by different commits we can fan out when we encounter a new failure and start to bisect that separately in parallel | 1 |
41,750 | 2,869,079,039 | IssuesEvent | 2015-06-05 23:09:52 | dart-lang/polymer-dart | https://api.github.com/repos/dart-lang/polymer-dart | closed | unresolved attribute doesn't work without platform.js | bug invalid Priority-High | <a href="https://github.com/jakemac53"><img src="https://avatars.githubusercontent.com/u/984921?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [jakemac53](https://github.com/jakemac53)**
_Originally opened as dart-lang/sdk#20934_
----
It looks like the unresolved attribute no longer works if you don't have platform.js included :( | 1.0 | unresolved attribute doesn't work without platform.js - <a href="https://github.com/jakemac53"><img src="https://avatars.githubusercontent.com/u/984921?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [jakemac53](https://github.com/jakemac53)**
_Originally opened as dart-lang/sdk#20934_
----
It looks like the unresolved attribute no longer works if you don't have platform.js included :( | non_infrastructure | unresolved attribute doesn t work without platform js issue by originally opened as dart lang sdk it looks like the unresolved attribute no longer works if you don t have platform js included | 0 |
29,644 | 24,135,201,381 | IssuesEvent | 2022-09-21 10:41:00 | openforis/fra-platform | https://api.github.com/repos/openforis/fra-platform | closed | Data Export: Data comparison | infrastructure | - [x] write initial script to compare data and produce differences
- [x] verify diffs and note issues #1533
| 1.0 | Data Export: Data comparison - - [x] write initial script to compare data and produce differences
- [x] verify diffs and note issues #1533
| infrastructure | data export data comparison write initial script to compare data and produce differences verify diffs and note issues | 1 |
24,290 | 12,060,304,271 | IssuesEvent | 2020-04-15 20:56:25 | ory/kratos | https://api.github.com/repos/ory/kratos | closed | Rebrand hooks to workflow(s) | module:2fa module:docs module:selfservice rfc | Hooks (`selfservice.login.(before|after)`) determine how hive should behave when a login or registration _was completed without an error state_. They do not modify the actual data of the identity and/or session. Therefore, they should be called "workflow" and "workflow step/handler/runner/part/item" from now on.
## Example: Password Registration
In this example, we want the user to be immediately signed in and redirected to our home page after registering a new account using email/password.
Configuration:
```yaml
selfservice:
registration:
after:
password:
- run: session
- run: redirect
config:
default_redirect_url: http://example.com/welcome
allow_user_defined_redirect: true # Allows "return_to" feature
```
[](https://mermaidjs.github.io/mermaid-live-editor/#/edit/eyJjb2RlIjoiZ3JhcGggVERcbiAgICBJQ1tJZGVudGl0eSBDcmVhdGVkXSAtLT4gVltTZXNzaW9uIENvb2tpZSBpc3N1ZWRdXG4gICAgVi0tPldbUmVkaXJlY3QgdG8gd2VsY29tZSBwYWdlXSIsIm1lcm1haWQiOnsidGhlbWUiOiJkZWZhdWx0In19)
## Example: Password Registration with (Email/Phone) Verification before login
In this example, we want the user to activate his/her account (by verifying the email or phone) before being able to sign in to our platform using email/password.
Configuration:
```yaml
selfservice:
registration:
after:
password:
- run: verify
- run: redirect
config:
default_redirect_url: http://example.com/please-active-account.html
allow_user_defined_redirect: false
login:
after:
password:
- run: verify # enforce that at least one email or phone is verified
- run: session
- run: redirect
config:
default_redirect_url: http://example.com/welcome
```
Flow:
[](https://mermaidjs.github.io/mermaid-live-editor/#/edit/eyJjb2RlIjoiZ3JhcGggVERcbiAgICBJQ1tJZGVudGl0eSBDcmVhdGVkXSAtLT4gVltSZXF1ZXN0IGFjY291bnQgYWN0aXZhdGlvbiAvIGVtYWlsIHZlcmlmaWNhdGlvbl1cbiAgICBWLS0-V1tSZWRpcmVjdCB0byBhIHBhZ2UgYXNraW5nIHlvdSB0byBjaGVjayB5b3VyIGVtYWlsIGluYm94XSIsIm1lcm1haWQiOnsidGhlbWUiOiJkZWZhdWx0In19)
## Example: Password Registration with (Email/Phone) Verification and immediate Login
Alternatively, we could want our new users to be signed in immediately after registration, but still require email verification. The welcome page would then, for example, show a red notification bar with "please verify your email"
Configuration:
```yaml
selfservice:
registration:
after:
- run: verify
- run: session
- run: redirect
config:
default_redirect_url: http://example.com/welcome.html
allow_user_defined_redirect: true
```
## Example: Password Registration with JSON Response
In this example, we would expect a `application/json` response. This could be used for native apps for example.
Configuration:
```yaml
selfservice:
registration:
after:
password:
- run: json
```
## Example: Disallow Registration during Weekdays
We might want to disallow registration during weekdays:
```yaml
selfservice:
registration:
before:
password:
- run: json-rpc
url: http://api.example.com/workflows/before-registration
```
*before-registration.js*
```js
const route = (r, w) => {
if (isWeekDay) {
w.send(403)
return
}
}
```
## Example: Disallow Registration based on external service
Assuming we want to forbid registration because some upstream system (e.g. CRM) says "no" to the identity data
```yaml
selfservice:
registration:
after:
password:
- run: json-rpc
url: http://api.example.com/workflows/after-registration
```
*after-registration.js*
```js
const route = (r, w) => {
if (checkIfAllowedByCRM(r.body.identity)) {
w.send(403) // this will tell hive to delete the identity
return
}
}
```
## Example: Create user at Stripe after Login / Registration
```yaml
selfservice:
login:
after:
password:
- run: json-rpc
url: http://api.example.com/workflows/stripe
registration:
after:
password:
- run: json-rpc
url: http://api.example.com/workflows/stripe
```
*stripe.js*
```js
const route = (r, w) => {
const user = r.body.identity.id
const stripeData = createUserAtStripe(user)
w.json({ traits: { stripe: stripeData } })
}
```
## Example: enforced 2FA
Here we're using the built-in 2fa module:
```yaml
selfservice:
login:
after:
password:
- run: 2fa
registration:
after:
password:
- run: 2fa
```
*hive/2fa.go*
```go
if hasExecuted2FA(ctx.Request) {
session.Enhance("I am now 2fa yay")
return nil
} else if has2faEnabled(ctx.Request) {
return ErrorRedirect("https://example.org/2fa-login")
} else !has2faEnabled(ctx.Request) {
return ErrorRedirect("https://example.org/2fa-setup")
}
```
## Example: contextual 2FA (e.g. based on source ip)
Assuming there is some context for our 2fa like a non-internal network ip
```yaml
selfservice:
login:
after:
password:
- run: json-rpc
url: http://api.example.com/workflows/2fa
registration:
after:
password:
- run: json-rpc
url: http://api.example.com/workflows/2fa
```
*2fa.js*
```js
if (ipRange !== "192.168.0.0/24") {
if hive.is2faEnabled(r.query.request)
send({ redirect_to: "https://example.org/2fa-setup" })
else if hive.is2faEnabled(r.query.request)
send({ redirect_to: "https://example.org/2fa-login" })
else
send(200)
}
``` | 1.0 | Rebrand hooks to workflow(s) - Hooks (`selfservice.login.(before|after)`) determine how hive should behave when a login or registration _was completed without an error state_. They do not modify the actual data of the identity and/or session. Therefore, they should be called "workflow" and "workflow step/handler/runner/part/item" from now on.
## Example: Password Registration
In this example, we want the user to be immediately signed in and redirected to our home page after registering a new account using email/password.
Configuration:
```yaml
selfservice:
registration:
after:
password:
- run: session
- run: redirect
config:
default_redirect_url: http://example.com/welcome
allow_user_defined_redirect: true # Allows "return_to" feature
```
[](https://mermaidjs.github.io/mermaid-live-editor/#/edit/eyJjb2RlIjoiZ3JhcGggVERcbiAgICBJQ1tJZGVudGl0eSBDcmVhdGVkXSAtLT4gVltTZXNzaW9uIENvb2tpZSBpc3N1ZWRdXG4gICAgVi0tPldbUmVkaXJlY3QgdG8gd2VsY29tZSBwYWdlXSIsIm1lcm1haWQiOnsidGhlbWUiOiJkZWZhdWx0In19)
## Example: Password Registration with (Email/Phone) Verification before login
In this example, we want the user to activate his/her account (by verifying the email or phone) before being able to sign in to our platform using email/password.
Configuration:
```yaml
selfservice:
registration:
after:
password:
- run: verify
- run: redirect
config:
default_redirect_url: http://example.com/please-active-account.html
allow_user_defined_redirect: false
login:
after:
password:
- run: verify # enforce that at least one email or phone is verified
- run: session
- run: redirect
config:
default_redirect_url: http://example.com/welcome
```
Flow:
[](https://mermaidjs.github.io/mermaid-live-editor/#/edit/eyJjb2RlIjoiZ3JhcGggVERcbiAgICBJQ1tJZGVudGl0eSBDcmVhdGVkXSAtLT4gVltSZXF1ZXN0IGFjY291bnQgYWN0aXZhdGlvbiAvIGVtYWlsIHZlcmlmaWNhdGlvbl1cbiAgICBWLS0-V1tSZWRpcmVjdCB0byBhIHBhZ2UgYXNraW5nIHlvdSB0byBjaGVjayB5b3VyIGVtYWlsIGluYm94XSIsIm1lcm1haWQiOnsidGhlbWUiOiJkZWZhdWx0In19)
## Example: Password Registration with (Email/Phone) Verification and immediate Login
Alternatively, we could want our new users to be signed in immediately after registration, but still require email verification. The welcome page would then, for example, show a red notification bar with "please verify your email"
Configuration:
```yaml
selfservice:
registration:
after:
- run: verify
- run: session
- run: redirect
config:
default_redirect_url: http://example.com/welcome.html
allow_user_defined_redirect: true
```
## Example: Password Registration with JSON Response
In this example, we would expect a `application/json` response. This could be used for native apps for example.
Configuration:
```yaml
selfservice:
registration:
after:
password:
- run: json
```
## Example: Disallow Registration during Weekdays
We might want to disallow registration during weekdays:
```yaml
selfservice:
registration:
before:
password:
- run: json-rpc
url: http://api.example.com/workflows/before-registration
```
*before-registration.js*
```js
const route = (r, w) => {
if (isWeekDay) {
w.send(403)
return
}
}
```
## Example: Disallow Registration based on external service
Assuming we want to forbid registration because some upstream system (e.g. CRM) says "no" to the identity data
```yaml
selfservice:
registration:
after:
password:
- run: json-rpc
url: http://api.example.com/workflows/after-registration
```
*after-registration.js*
```js
const route = (r, w) => {
if (checkIfAllowedByCRM(r.body.identity)) {
w.send(403) // this will tell hive to delete the identity
return
}
}
```
## Example: Create user at Stripe after Login / Registration
```yaml
selfservice:
login:
after:
password:
- run: json-rpc
url: http://api.example.com/workflows/stripe
registration:
after:
password:
- run: json-rpc
url: http://api.example.com/workflows/stripe
```
*stripe.js*
```js
const route = (r, w) => {
const user = r.body.identity.id
const stripeData = createUserAtStripe(user)
w.json({ traits: { stripe: stripeData } })
}
```
## Example: enforced 2FA
Here we're using the built-in 2fa module:
```yaml
selfservice:
login:
after:
password:
- run: 2fa
registration:
after:
password:
- run: 2fa
```
*hive/2fa.go*
```go
if hasExecuted2FA(ctx.Request) {
session.Enhance("I am now 2fa yay")
return nil
} else if has2faEnabled(ctx.Request) {
return ErrorRedirect("https://example.org/2fa-login")
} else !has2faEnabled(ctx.Request) {
return ErrorRedirect("https://example.org/2fa-setup")
}
```
## Example: contextual 2FA (e.g. based on source ip)
Assuming there is some context for our 2fa like a non-internal network ip
```yaml
selfservice:
login:
after:
password:
- run: json-rpc
url: http://api.example.com/workflows/2fa
registration:
after:
password:
- run: json-rpc
url: http://api.example.com/workflows/2fa
```
*2fa.js*
```js
if (ipRange !== "192.168.0.0/24") {
if hive.is2faEnabled(r.query.request)
send({ redirect_to: "https://example.org/2fa-setup" })
else if hive.is2faEnabled(r.query.request)
send({ redirect_to: "https://example.org/2fa-login" })
else
send(200)
}
``` | non_infrastructure | rebrand hooks to workflow s hooks selfservice login before after determine how hive should behave when a login or registration was completed without an error state they do not modify the actual data of the identity and or session therefore they should be called workflow and workflow step handler runner part item from now on example password registration in this example we want the user to be immediately signed in and redirected to our home page after registering a new account using email password configuration yaml selfservice registration after password run session run redirect config default redirect url allow user defined redirect true allows return to feature example password registration with email phone verification before login in this example we want the user to activate his her account by verifying the email or phone before being able to sign in to our platform using email password configuration yaml selfservice registration after password run verify run redirect config default redirect url allow user defined redirect false login after password run verify enforce that at least one email or phone is verified run session run redirect config default redirect url flow example password registration with email phone verification and immediate login alternatively we could want our new users to be signed in immediately after registration but still require email verification the welcome page would then for example show a red notification bar with please verify your email configuration yaml selfservice registration after run verify run session run redirect config default redirect url allow user defined redirect true example password registration with json response in this example we would expect a application json response this could be used for native apps for example configuration yaml selfservice registration after password run json example disallow registration during weekdays we might want to disallow registration during weekdays yaml selfservice registration before password run json rpc url before registration js js const route r w if isweekday w send return example disallow registration based on external service assuming we want to forbid registration because some upstream system e g crm says no to the identity data yaml selfservice registration after password run json rpc url after registration js js const route r w if checkifallowedbycrm r body identity w send this will tell hive to delete the identity return example create user at stripe after login registration yaml selfservice login after password run json rpc url registration after password run json rpc url stripe js js const route r w const user r body identity id const stripedata createuseratstripe user w json traits stripe stripedata example enforced here we re using the built in module yaml selfservice login after password run registration after password run hive go go if ctx request session enhance i am now yay return nil else if ctx request return errorredirect else ctx request return errorredirect example contextual e g based on source ip assuming there is some context for our like a non internal network ip yaml selfservice login after password run json rpc url registration after password run json rpc url js js if iprange if hive r query request send redirect to else if hive r query request send redirect to else send | 0 |
102,377 | 8,826,275,188 | IssuesEvent | 2019-01-03 00:56:34 | decentralized-identity/sidetree-core | https://api.github.com/repos/decentralized-identity/sidetree-core | closed | Fix test code coverage execution | bug good first issue test | Currently getting the following error when running code coverage:
Error: Cannot find module './json/didDocumentTemplate.json' | 1.0 | Fix test code coverage execution - Currently getting the following error when running code coverage:
Error: Cannot find module './json/didDocumentTemplate.json' | non_infrastructure | fix test code coverage execution currently getting the following error when running code coverage error cannot find module json diddocumenttemplate json | 0 |
5,220 | 5,530,088,061 | IssuesEvent | 2017-03-21 00:59:49 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | System.Data facade has a lot of bogus global stuff that should be removed | area-Infrastructure bug | GenFacades isn't clearing out the global module references from the assembly when it rewrites it. We should remove those to eliminate bogus type and member references. | 1.0 | System.Data facade has a lot of bogus global stuff that should be removed - GenFacades isn't clearing out the global module references from the assembly when it rewrites it. We should remove those to eliminate bogus type and member references. | infrastructure | system data facade has a lot of bogus global stuff that should be removed genfacades isn t clearing out the global module references from the assembly when it rewrites it we should remove those to eliminate bogus type and member references | 1 |
29,957 | 24,426,677,751 | IssuesEvent | 2022-10-06 03:46:21 | C7-Game/Prototype | https://api.github.com/repos/C7-Game/Prototype | closed | Migrate to Godot 3.5 | infrastructure godot | Minor version updates are recommended for all users but may require regression testing and some minor changes. Once released, target Godot 3.5 and make whatever file updates that entails | 1.0 | Migrate to Godot 3.5 - Minor version updates are recommended for all users but may require regression testing and some minor changes. Once released, target Godot 3.5 and make whatever file updates that entails | infrastructure | migrate to godot minor version updates are recommended for all users but may require regression testing and some minor changes once released target godot and make whatever file updates that entails | 1 |
217,330 | 24,329,575,690 | IssuesEvent | 2022-09-30 18:02:07 | Nexmo/station | https://api.github.com/repos/Nexmo/station | opened | postcss-smart-import-0.7.6.tgz: 1 vulnerabilities (highest severity is: 7.5) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postcss-smart-import-0.7.6.tgz</b></p></summary>
<p></p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/Nexmo/station/commit/1e5781423c543a0c9bfedb4c5a57ca049920974b">1e5781423c543a0c9bfedb4c5a57ca049920974b</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-23343](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | path-parse-1.0.6.tgz | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23343</summary>
### Vulnerable Library - <b>path-parse-1.0.6.tgz</b></p>
<p>Node.js path.parse() ponyfill</p>
<p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz</a></p>
<p>
Dependency Hierarchy:
- postcss-smart-import-0.7.6.tgz (Root Library)
- resolve-1.20.0.tgz
- :x: **path-parse-1.0.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Nexmo/station/commit/1e5781423c543a0c9bfedb4c5a57ca049920974b">1e5781423c543a0c9bfedb4c5a57ca049920974b</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution: path-parse - 1.0.7</p>
</p>
<p></p>
</details> | True | postcss-smart-import-0.7.6.tgz: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postcss-smart-import-0.7.6.tgz</b></p></summary>
<p></p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/Nexmo/station/commit/1e5781423c543a0c9bfedb4c5a57ca049920974b">1e5781423c543a0c9bfedb4c5a57ca049920974b</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-23343](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | path-parse-1.0.6.tgz | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23343</summary>
### Vulnerable Library - <b>path-parse-1.0.6.tgz</b></p>
<p>Node.js path.parse() ponyfill</p>
<p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz</a></p>
<p>
Dependency Hierarchy:
- postcss-smart-import-0.7.6.tgz (Root Library)
- resolve-1.20.0.tgz
- :x: **path-parse-1.0.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Nexmo/station/commit/1e5781423c543a0c9bfedb4c5a57ca049920974b">1e5781423c543a0c9bfedb4c5a57ca049920974b</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution: path-parse - 1.0.7</p>
</p>
<p></p>
</details> | non_infrastructure | postcss smart import tgz vulnerabilities highest severity is vulnerable library postcss smart import tgz found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high path parse tgz transitive n a details cve vulnerable library path parse tgz node js path parse ponyfill library home page a href dependency hierarchy postcss smart import tgz root library resolve tgz x path parse tgz vulnerable library found in head commit a href found in base branch main vulnerability details all versions of package path parse are vulnerable to regular expression denial of service redos via splitdevicere splittailre and splitpathre regular expressions redos exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution path parse | 0 |
28,283 | 23,132,480,554 | IssuesEvent | 2022-07-28 11:41:17 | spring-projects-experimental/spring-modulith | https://api.github.com/repos/spring-projects-experimental/spring-modulith | opened | Investigate support for build time optimization | in: infrastructure in test support | ## Context
A common challenge for modulithic applications is that a change in the version control system triggers a build of the overall system. That build usually consists of steps that are executed for the entire system, which means: for all modules, whether the change in question actually affects a module directly or transitively.
The structural information we have about the system could be combined with knowledge about the change set at hand to identify which modules are directly affected by the change itself. We could then find out about all modules directly or transitively depending on the "dirty" modules and limit the test execution to filter test cases contained in those modules only. Similarly to limiting the parts of the application that are actually bootstrapped during module-specific test via `@ModuleTest`.
## Ideas
JUnit 5 has dedicated support to hook into the [test discovery process](https://junit.org/junit5/docs/current/user-guide/#launcher-api-discovery), i.e. it should be possible to place a JAR in the classpath of e.g. the Maven Surefire execution that filters the test cases to actually be executed.
We also need to find a mechanism to identify, what the set of changes is, we want to start from. During local development, this is likely to be the files currently considered changed by Git. During a CI run (the more important aspect) it's likely the last change set or the set of change sets in between the previous build and the current one.
Original ticket:
* moduliths/moduliths#174 | 1.0 | Investigate support for build time optimization - ## Context
A common challenge for modulithic applications is that a change in the version control system triggers a build of the overall system. That build usually consists of steps that are executed for the entire system, which means: for all modules, whether the change in question actually affects a module directly or transitively.
The structural information we have about the system could be combined with knowledge about the change set at hand to identify which modules are directly affected by the change itself. We could then find out about all modules directly or transitively depending on the "dirty" modules and limit the test execution to filter test cases contained in those modules only. Similarly to limiting the parts of the application that are actually bootstrapped during module-specific test via `@ModuleTest`.
## Ideas
JUnit 5 has dedicated support to hook into the [test discovery process](https://junit.org/junit5/docs/current/user-guide/#launcher-api-discovery), i.e. it should be possible to place a JAR in the classpath of e.g. the Maven Surefire execution that filters the test cases to actually be executed.
We also need to find a mechanism to identify, what the set of changes is, we want to start from. During local development, this is likely to be the files currently considered changed by Git. During a CI run (the more important aspect) it's likely the last change set or the set of change sets in between the previous build and the current one.
Original ticket:
* moduliths/moduliths#174 | infrastructure | investigate support for build time optimization context a common challenge for modulithic applications is that a change in the version control system triggers a build of the overall system that build usually consists of steps that are executed for the entire system which means for all modules whether the change in question actually affects a module directly or transitively the structural information we have about the system could be combined with knowledge about the change set at hand to identify which modules are directly affected by the change itself we could then find out about all modules directly or transitively depending on the dirty modules and limit the test execution to filter test cases contained in those modules only similarly to limiting the parts of the application that are actually bootstrapped during module specific test via moduletest ideas junit has dedicated support to hook into the i e it should be possible to place a jar in the classpath of e g the maven surefire execution that filters the test cases to actually be executed we also need to find a mechanism to identify what the set of changes is we want to start from during local development this is likely to be the files currently considered changed by git during a ci run the more important aspect it s likely the last change set or the set of change sets in between the previous build and the current one original ticket moduliths moduliths | 1 |
58,985 | 11,924,598,821 | IssuesEvent | 2020-04-01 09:48:13 | home-assistant/brands | https://api.github.com/repos/home-assistant/brands | closed | Tesla is missing brand images | has-codeowner has-config-flow |
## The problem
The Tesla integration has missing brand images.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/tesla/logo.png`
- `src/tesla/icon@2x.png`
- `src/tesla/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @zabuldon, @alandtse! Mind taking a look at this issue as it is with an integration (tesla) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/tesla/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
| 1.0 | Tesla is missing brand images -
## The problem
The Tesla integration has missing brand images.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/tesla/logo.png`
- `src/tesla/icon@2x.png`
- `src/tesla/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @zabuldon, @alandtse! Mind taking a look at this issue as it is with an integration (tesla) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/tesla/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
| non_infrastructure | tesla is missing brand images the problem the tesla integration has missing brand images we recently started this brands repository to create a centralized storage of all brand related images these images are used on our website and the home assistant frontend the following images are missing and would ideally be added src tesla logo png src tesla icon png src tesla logo png for image specifications and requirements please see additional information for more information about this repository read the file of this repository it contains information on how this repository works and image specification and requirements codeowner mention hi there zabuldon alandtse mind taking a look at this issue as it is with an integration tesla you are listed as a for thanks resolving this issue is not limited to codeowners if you want to help us out feel free to resolve this issue thanks already | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.